Chapter 5: Process Scheduling Outline Basic Concepts CPU-I/O Burst Cycle CPU Scheduler Preemptive Scheduling Dispatcher Scheduling Criteria Scheduling Algorithms First-Come, First-Serve Scheduling Shortest-Job-First Scheduling Priority Scheduling Round-Robin Scheduling Multilevel Queue Scheduling Multilevel Feedback-Queue Scheduling Basic Concepts CPU: One of the primary computer resource Maximum CPU utilization obtained with multiprogramming Several processes are kept in memory CPU–I/O Burst Cycle – Process execution consists of a cycle of CPU execution and I/O wait. CPU burst distribution determines how to schedule the processes. Alternating Sequence of CPU And I/O Bursts Histogram of CPU-burst Times Normally a process starts with frequent short CPU bursts and then it stabilizes and shifts to long but infrequent CPU bursts CPU-bound program I/O bound processes usually have many small CPU bursts CPU bound programs usually have few but long CPU bursts CPU Scheduler Selects from among the processes in memory that are ready to execute, and allocates the CPU to one of them. Is it the same as Short-term scheduler? … Yes CPU scheduling decisions may take place when a process: 1. Switches from running to waiting state. 2. Switches from running to ready state. 3. Switches from waiting to ready. 4. Terminates.(Own will) Scheduling Scheme given at 1 and 4 is nonpreemptive (Cooperative). All other scheduling is preemptive. Preemptive vs Non Preemptive Scheduling Non Preemptive scheduling If a process is being executed then it keeps the CPU until one or four conditions of previous slide occur. MS Windows 3.x which was sort of multitasking system used this method Not at all efficient. Preemptive A process in running state can be brought back to ready state without any of the conditions of previous slide being met. Can you guess how it is done? Problems with preemptive Although better than non-preemptive still suffers from synchronization problems E.g. Process P and Q share some data (can be memory location or a file) P does some calculations changes some of the data but before it finishes its time expires Q reads the data but is the data valid? Dispatcher Dispatcher module gives control of the CPU to the process selected by the short-term scheduler; this involves: switching context switching to user mode jumping to the proper location in the user program to restart that program Dispatch latency – time it takes for the dispatcher to stop one process and start another running. Should be minimum. Scheduling Criteria Following criteria is kept under consideration while designing a scheduler CPU utilization – Keep the CPU as busy as possible Throughput – # of processes that complete their execution per unit time (Given Time). Turnaround time – Amount of time to execute a particular process (Sum of ready, waiting, running, loading times,I/O) Waiting time – Amount of time a process has been waiting in the ready queue Response time – Amount of time it takes from when a request was submitted until the first response is produced, not output (for interactive environment) Time it takes to start responding Not the time it takes to output the response Optimization Criteria So the best CPU Scheduling algorithm is which … Maximizes CPU utilization Maximizes throughput Minimizes turnaround time Minimizes waiting time Minimizes response time Outline Basic Concepts CPU-I/O Burst Cycle CPU Scheduler Preemptive Scheduling Dispatcher Scheduling Criteria Scheduling Algorithms First-Come, First-Serve Scheduling Shortest-Job-First Scheduling Priority Scheduling Round-Robin Scheduling Multilevel Queue Scheduling Multilevel Feedback-Queue Scheduling First-Come, First-Served (FCFS) Scheduling Process Burst Time P1 24 P2 3 P3 3 Suppose that the processes arrive in the order: P1 , P2 , P3 The Gantt Chart for the schedule is: P1 0 P2 24 P3 27 Waiting time for P1 = 0; P2 = 24; P3 = 27 Average waiting time: (0 + 24 + 27)/3 = 17 30 FCFS Scheduling (Cont.) Suppose that the processes arrive in the order P2 , P3 , P1 . The Gantt chart for the schedule is: P2 0 P3 3 P1 6 30 Waiting time for P1 = 6; P2 = 0; P3 = 3 Average waiting time: (6 + 0 + 3)/3 = 3 Much better than previous case. CPU bound processes will get and hold the CPU Convoy effect short process behind long process (Bad CPU utilization) Shortest-Job-First (SJR) Scheduling Associate with each process the length of its next CPU burst. Use these lengths to schedule the process with the shortest time. Two schemes: nonpreemptive – once CPU given to the process it cannot be preempted until it completes its CPU burst. preemptive – if a new process arrives with CPU burst length less than remaining time of current executing process, preempt. This scheme is know as the Shortest-Remaining-Time-First (SRTF). SJF is optimal – gives minimum average waiting time for a given set of processes. Shortest-next-cpu-burst algorithm Example of Non-Preemptive SJF Process P1 P2 P3 P4 Burst Time 6 8 7 3 SJF 8 P1 P4 0 3 P3 9 P2 16 Average waiting time = (3+16+9+0)/4=7 24 FCFS=10.25 Shortest-Job-First (SJR) Scheduling Good for long-term Scheduler in batch systems User specifies process time during job submission Lower value may mean faster response Too low a value will cause a time-limit-exceeded error Difficult to implement at the level of short-term CPU scheduler Determining Length of Next CPU Burst Can only predict/ estimate the length of next CPU burst. Can be done by using the length of previous CPU bursts, using exponential averaging. Examples of Exponential Averaging =0 n+1 = n Recent history does not count. =1 n+1 =αtn +(1- α) n tn: length of the nth CPU burst n+1 : Our predicted Value α: weight of recent & past history 0<= α<=1 n+1 = tn Only the actual last CPU burst counts. Since both and (1 - ) are less than or equal to 1, each successive term has less weight than its predecessor. Prediction of the Length of the Next CPU Burst Initial guess for T(1) = 10 ms Actual 6 Determination of 2 Cycle based on actual (6) and last guess (10) T(2) = 0.5 x 6 + 0.5 x 10 = 8 ms Determination of 3 cycle based on actual(4) and last guess (8) T(3) = 0.5 x 4 + 0.5 x 8 = 6 ms … Now you can do it For example: Suppose a process p is given a default expected burst length of 5 time units. When it is run, the actually burst lengths are 10,10,10,1,1,1 (although this information is not known in advance to any algorithm). The prediction of burst times for this process works as follows. Let e(1) = 5, as a default value. When process p runs, its first burst actually runs 10 time units, so, a(1) = 10. e(2) =0.5 * e(1) * 0.5 + a(1) = 0.5 * 5 + 0.5 * 10 = 7.5 This is the prediction for the second cpu burst The actual second cpu burst is 10. So the prediction for the third cpu burst is: e(3) = 0.5 * e(2) * 0.5 + a(2) = 0.5 * 7.5 + 0.5 * 10 = 8.75 e(4) = 0.5 * e(3) * 0.5 + a(3) = 0.5 * 8.75 + 0.5 * 10 = 9.38, So, we predict that the next burst will be close to 10 (9.38) because the recent bursts have been of length 10. At this point, it happens that the process starts having shorter bursts, with a(4) = 1, so the algorithm gradually adjusts its estimated cpu burst (prediction) e(5) = 0.5 * e(4) * 0.5 + a(4) = 0.5 * 9.38 + 0.5 * 1 = 5.19 e(6) = 0.5 * e(5) * 0.5 + a(5) = 0.5 * 5.19 + 0.5 * 1 = 3.10 e(7) = 0.5 * e(6) * 0.5 + a(6) = 0.5 * 3.10 + 0.5 * 1 = 2.05 Example of Preemptive SJF (Shortest Remaining Time first) Process Arrival Time P1 0 P2 1 P3 2 P4 3 SJF (preemptive) P1 0 1 P2 P1 P4 5 Burst Time 8 4 9 5 10 P3 17 Average waiting time = 6.5 Non-Preemptive Average waiting time = 6.5 26 Outline Basic Concepts CPU-I/O Burst Cycle CPU Scheduler Preemptive Scheduling Dispatcher Scheduling Criteria Scheduling Algorithms First-Come, First-Serve Scheduling Shortest-Job-First Scheduling Priority Scheduling Round-Robin Scheduling Multilevel Queue Scheduling Multilevel Feedback-Queue Scheduling Priority Scheduling A priority number (integer) is associated with each process The CPU is allocated to the process with the highest priority (smallest integer highest priority). Preemptive Non-preemptive SJF is a priority scheduling where priority is the predicted next CPU burst time. Problem Starvation – low priority processes may never execute. Solution Aging – as time progresses increase the priority of the process. Round Robin (RR) Each process gets a small unit of CPU time (time quantum), usually 10-100 milliseconds. After this time has elapsed, the process is preempted and added to the end of the ready queue. If there are n processes in the ready queue and the time quantum is q, then each process gets 1/n of the CPU time in chunks of at most q time units at once. No process waits more than (n-1)q time units. Performance q large FIFO q small q must be large with respect to context switch, otherwise overhead is too high. Example of RR with Time Quantum = 20 Process P1 P2 P3 P4 The Gantt chart is: P1 0 P2 20 37 P3 P4 57 P1 77 Burst Time 53 17 68 24 P3 97 117 P4 P1 P3 P3 121 134 154 162 Typically, higher average turnaround than SJF, but better response. Time Quantum and Context Switch Time If context switching time is approx 10 percent of the time quantum, then about 10 percent of the CPU time will be spent in context switching Most modern systems have time quanta ranging from 10 to 100 milliseconds Context switch typically takes less than 10 microseconds Remember: Context switch takes time Turnaround Time Varies With The Time Quantum 80% of CPU bursts should be shorter than the time quantum Average Turnaround time doesn’t necessarily improve as the time-quantum size increases) Outline Basic Concepts CPU-I/O Burst Cycle CPU Scheduler Preemptive Scheduling Dispatcher Scheduling Criteria Scheduling Algorithms First-Come, First-Serve Scheduling Shortest-Job-First Scheduling Priority Scheduling Round-Robin Scheduling Multilevel Queue Scheduling Multilevel Feedback-Queue Scheduling Multilevel Queue Ready queue is partitioned into separate queues: foreground (interactive) background (batch) Each queue has its own scheduling algorithm, foreground – RR background – FCFS Scheduling must be done between the queues. Fixed priority scheduling; (i.e., serve all from foreground then from background). Possibility of starvation. Time slice – each queue gets a certain amount of CPU time which it can schedule amongst its processes; i.e., 80% to foreground in RR 20% to background in FCFS Multilevel Queue Scheduling Multilevel Feedback Queue A process can move between the various queues; aging can be implemented this way. Multilevel-feedback-queue scheduler defined by the following parameters: number of queues scheduling algorithms for each queue method used to determine when to upgrade a process method used to determine when to demote a process method used to determine which queue a process will enter when that process needs service Example of Multilevel Feedback Queue Consider a case if we have three queues: Q0 – time quantum 8 milliseconds Q1 – time quantum 16 milliseconds Q2 – FCFS Scheduling A new job enters queue Q0 which is served FCFS. When it gains CPU, job receives 8 milliseconds. If it does not finish in 8 milliseconds, job is moved to tail of queue Q1. At Q1 job is again served FCFS and receives 16 additional milliseconds. If it still does not complete, it is preempted and moved to tail of queue Q2. Multilevel Feedback Queues Outline Basic Concepts CPU-I/O Burst Cycle CPU Scheduler Preemptive Scheduling Dispatcher Scheduling Criteria Scheduling Algorithms First-Come, First-Serve Scheduling Shortest-Job-First Scheduling Priority Scheduling Round-Robin Scheduling Multilevel Queue Scheduling Multilevel Feedback-Queue Scheduling Multi-Processor Scheduling Multiple-Processor Scheduling CPU scheduling more complex when multiple CPUs are available. Consider Homogeneous processors within a multiprocessor system. Load sharing Asymmetric multiprocessing – only one processor accesses the system data structures, alleviating the need for data sharing. SMP: Each processor is self scheduling May have separate ready queues Multiple-Processor Scheduling –Process Affinity Cache memory? The Data most recently accessed populates it Process migration to another process Invalidation & Repopulation Some OS attempt to keep a process running on the same processor –Affinity Soft Affinity: Attempt to keep a process running on the same processor but not guaranteeing Hard Affinity: Process cannot migrate to another processor Multiple-Processor Scheduling –Load Balancing Utilize the full benefits of multi-processors Load balancing: Distribute the workload evenly across all processors. Necessary where each processors has its own private queue of eligible processes Not necessary where processors have common run queue Modern OS --private queue or common? Push Migration: move processes from overloaded to idle processors Pull Migration: Idle processor pulls a waiting process from a busy processor Any link with Process Affinity? Multiple-Processor Scheduling Symmetric Multithreading Multiple logical rather than physical processors to run several threads concurrently (HPT) Create logical processors on the same physical processor—a view of many Processors to OS Each has its own machine-state registers Own interrupt handling Logical CPU Logical CPU Logical CPU Logical CPU Physical CPU Physical CPU System Bus Multiple-Processor Scheduling Symmetric Multithreading SMT: a feature provided by hardware Os needn’t be designed differently Performance gains –If OS is aware. But why? **Scheduler** Outline Basic Concepts CPU-I/O Burst Cycle CPU Scheduler Preemptive Scheduling Dispatcher Scheduling Criteria Scheduling Algorithms First-Come, First-Serve Scheduling Shortest-Job-First Scheduling Priority Scheduling Round-Robin Scheduling Multilevel Queue Scheduling Multilevel Feedback-Queue Scheduling Multi-Processor Scheduling OS Examples –Self Study Algorithm Evaluation Algorithm Evaluation How do we select the best algorithm? Defining a criteria CPU Utilization, response time, throughput Criteria may include several measures e.g. Maximize CPU utilization |and| Response time is 1 e.g. Maximize Throughput |and| turnaround time Analytic Evaluation Method The algorithm and some system workload are used to produce a formula or number which gives the performance of the algorithm for that workload. Deterministic modeling Queuing models Simulations Implementation Deterministic Modeling Predetermined workload –defines performance of each algorithm for that workload. Use of Gantt charts. Simple and fast Exact numbers for comparison Answers apply only for the cases considered. Performance figures may not be true in general Suitable: If we are running the same program over and over again. We can easily select an algorithm. Over a set of examples certain trends can be analyzed, e.g. If all the processes arrive at time 0, the SJF policy will always result in min waiting time If the Arriving time is different the SJF may not always result in min average waiting time. Deterministic Modeling Process P1 P2 P3 P4 Gantt chart Arrival Time 0.0 2.0 4.0 5.0 P1 0 P2 2 P3 4 Burst Time 7 4 1 4 P2 5 P4 7 P1 11 Average waiting time = (9 + 1 + 0 +2)/4 = 3 16 Queuing Modeling What if the processes that are run, vary from day to day? -Is deterministic method useful? The distributions of CPU and I/O bursts can be determined/ approximated/ estimated. Mathematical formula: probability of a particular CPU burst Arrival times distribution can also be estimated. Computer system viewed as a network of queues and servers: ready queue, I/O queue, event queues, CPUs, I/O device controllers, etc. e.g. CPU: ready queue, I/O system :device queue Input: Arrival and service rates Output: CPU utilization, average queue length, average waiting time, … Queuing Modeling Little’s Formula: n = λ* W where n = average queue length λ = average arrival rate W = average waiting time in a queue Queuing Modeling Let the average job arrival rate be 0.5 Algorithm Average Queue Length(n) FCFS Average Wait Time W=tw 4.6 SJF 3.6 1.8 SRTF 3.2 1.6 RR (q=1) 7.0 3.5 RR (q=4) 6.0 3.0 2.3 Queuing Modeling Complicated mathematics Distributions (uniform, exponential, etc) for the arrival and departure rates can be difficult to work with Assumptions may not be accurate Approximation of the real system Simulation To get more accurate evaluation Workload generated by assuming some distribution and a random number generator, or by collecting data from the actual system. The distributions can be defined mathematically or empirically. Simulation Simulation Characteristics Expensive: hours of programming and execution time May be erroneous because of the assumptions about distributions Implementation Even simulation is of limited accuracy Best way: Code and implement the scheduling algorithm and see High cost –coding the algorithm Modify the OS to support it Reaction of the users to changing OS Environment changes: if short processes are given high priority then user may break their larger processes into many short ones For example: Design of an automatic Interactive/ non-interactive process classifier Algorithm