Download as pdf or txt
Download as pdf or txt
You are on page 1of 19

Process

A process is basically a program in execution. The execution of a process must progress


in a sequential fashion.
A process is defined as an entity which represents the basic unit of work to be implemented
in the system.

To put it in simple terms, we write our computer programs in a text file and when we
execute this program, it becomes a process which performs all the tasks mentioned in
the program.
When a program is loaded into the memory and it becomes a process, it can be divided
into four sections ─ stack, heap, text and data. The following image shows a simplified
layout of a process inside main memory −

S.N. Component & Description

1
Stack
The process Stack contains the temporary data such as method/function
parameters, return address and local variables.

2
Heap
This is dynamically allocated memory to a process during its run time.
3
Text
This includes the current activity represented by the value of Program Counter and
the contents of the processor's registers.

4
Data
This section contains the global and static variables.

Program
A program is a piece of code which may be a single line or millions of lines. A computer
program is usually written by a computer programmer in a programming language. For
example, here is a simple program written in C programming language −
#include <stdio.h>

int main() {
printf("Hello, World! \n");
return 0;
}
A computer program is a collection of instructions that performs a specific task when
executed by a computer. When we compare a program with a process, we can conclude
that a process is a dynamic instance of a computer program.
A part of a computer program that performs a well-defined task is known as an algorithm.
A collection of computer programs, libraries and related data are referred to as
a software.

Process Life Cycle


When a process executes, it passes through different states. These stages may differ in
different operating systems, and the names of these states are also not standardized.
In general, a process can have one of the following five states at a time.

S.N. State & Description

1
Start
This is the initial state when a process is first started/created.

2
Ready
The process is waiting to be assigned to a processor. Ready processes are waiting
to have the processor allocated to them by the operating system so that they can
run. Process may come into this state after Start state or while running it by but
interrupted by the scheduler to assign CPU to some other process.

3
Running
Once the process has been assigned to a processor by the OS scheduler, the
process state is set to running and the processor executes its instructions.

4
Waiting
Process moves into the waiting state if it needs to wait for a resource, such as
waiting for user input, or waiting for a file to become available.

5
Terminated or Exit
Once the process finishes its execution, or it is terminated by the operating system,
it is moved to the terminated state where it waits to be removed from main memory.

Process Control Block (PCB)


A Process Control Block is a data structure maintained by the Operating System for every
process. The PCB is identified by an integer process ID (PID). A PCB keeps all the
information needed to keep track of a process as listed below in the table −

S.N. Information & Description

1
Process State
The current state of the process i.e., whether it is ready, running, waiting, or
whatever.

2
Process privileges
This is required to allow/disallow access to system resources.

3
Process ID
Unique identification for each of the process in the operating system.

4
Pointer
A pointer to parent process.

5
Program Counter
Program Counter is a pointer to the address of the next instruction to be executed
for this process.

6
CPU registers
Various CPU registers where process need to be stored for execution for running
state.

7
CPU Scheduling Information
Process priority and other scheduling information which is required to schedule the
process.

8
Memory management information
This includes the information of page table, memory limits, Segment table
depending on memory used by the operating system.

9
Accounting information
This includes the amount of CPU used for process execution, time limits, execution
ID etc.

10
IO status information
This includes a list of I/O devices allocated to the process.

The architecture of a PCB is completely dependent on Operating System and may contain
different information in different operating systems. Here is a simplified diagram of a PCB

The PCB is maintained for a process throughout its lifetime, and is deleted once the
process terminates.

Definition
The process scheduling is the activity of the process manager that handles the removal
of the running process from the CPU and the selection of another process on the basis
of a particular strategy.
Process scheduling is an essential part of a Multiprogramming operating systems. Such
operating systems allow more than one process to be loaded into the executable memory
at a time and the loaded process shares the CPU using time multiplexing.

Categories of Scheduling
There are two categories of scheduling:

1. Non-preemptive: Here the resource can’t be taken from a process until the process
completes execution. The switching of resources occurs when the running process
terminates and moves to a waiting state.
2. Preemptive: Here the OS allocates the resources to a process for a fixed amount of
time. During resource allocation, the process switches from running state to ready
state or from waiting state to ready state. This switching occurs as the CPU may give
priority to other processes and replace the process with higher priority with the
running process.
Process Scheduling Queues
The OS maintains all Process Control Blocks (PCBs) in Process Scheduling Queues. The
OS maintains a separate queue for each of the process states and PCBs of all processes
in the same execution state are placed in the same queue. When the state of a process
is changed, its PCB is unlinked from its current queue and moved to its new state queue.
The Operating System maintains the following important process scheduling queues −
• Job queue − This queue keeps all the processes in the system.
• Ready queue − This queue keeps a set of all processes residing in main memory,
ready and waiting to execute. A new process is always put in this queue.
• Device queues − The processes which are blocked due to unavailability of an I/O
device constitute this queue.

The OS can use different policies to manage each queue (FIFO, Round Robin, Priority,
etc.). The OS scheduler determines how to move processes between the ready and run
queues which can only have one entry per processor core on the system; in the above
diagram, it has been merged with the CPU.

Two-State Process Model


Two-state process model refers to running and non-running states which are described
below −

S.N. State & Description

1
Running
When a new process is created, it enters into the system as in the running state.
2
Not Running
Processes that are not running are kept in queue, waiting for their turn to execute.
Each entry in the queue is a pointer to a particular process. Queue is implemented
by using linked list. Use of dispatcher is as follows. When a process is interrupted,
that process is transferred in the waiting queue. If the process has completed or
aborted, the process is discarded. In either case, the dispatcher then selects a
process from the queue to execute.

Types of Schedulers
Schedulers are special system software which handle process scheduling in various
ways. Their main task is to select the jobs to be submitted into the system and to decide
which process to run. Schedulers are of three types −

• Long-Term Scheduler
• Short-Term Scheduler
• Medium-Term Scheduler

Long Term Scheduler


It is also called a job scheduler. A long-term scheduler determines which programs are
admitted to the system for processing. It selects processes from the queue and loads
them into memory for execution. Process loads into the memory for CPU scheduling.
The primary objective of the job scheduler is to provide a balanced mix of jobs, such as
I/O bound and processor bound. It also controls the degree of multiprogramming. If the
degree of multiprogramming is stable, then the average rate of process creation must be
equal to the average departure rate of processes leaving the system.
On some systems, the long-term scheduler may not be available or minimal. Time-sharing
operating systems have no long term scheduler. When a process changes the state from
new to ready, then there is use of long-term scheduler.

Short Term Scheduler


It is also called as CPU scheduler. Its main objective is to increase system performance
in accordance with the chosen set of criteria. It is the change of ready state to running
state of the process. CPU scheduler selects a process among the processes that are
ready to execute and allocates CPU to one of them.
Short-term schedulers, also known as dispatchers, make the decision of which process
to execute next. Short-term schedulers are faster than long-term schedulers.

Medium Term Scheduler


Medium-term scheduling is a part of swapping. It removes the processes from the
memory. It reduces the degree of multiprogramming. The medium-term scheduler is in-
charge of handling the swapped out-processes.
A running process may become suspended if it makes an I/O request. A suspended
processes cannot make any progress towards completion. In this condition, to remove
the process from memory and make space for other processes, the suspended process
is moved to the secondary storage. This process is called swapping, and the process is
said to be swapped out or rolled out. Swapping may be necessary to improve the process
mix.

Comparison among Scheduler


S.N. Long-Term Scheduler Short-Term Scheduler Medium-Term Scheduler

1 It is a job scheduler It is a CPU scheduler It is a process swapping


scheduler.

2 Speed is lesser than short Speed is fastest among Speed is in between both
term scheduler other two short and long term
scheduler.

3 It controls the degree of It provides lesser control It reduces the degree of


multiprogramming over degree of multiprogramming.
multiprogramming

4 It is almost absent or It is also minimal in time It is a part of Time sharing


minimal in time sharing sharing system systems.
system

5 It selects processes from It selects those It can re-introduce the


pool and loads them into processes which are process into memory and
memory for execution ready to execute execution can be continued.

Context Switching
A context switching is the mechanism to store and restore the state or context of a CPU
in Process Control block so that a process execution can be resumed from the same point
at a later time. Using this technique, a context switcher enables multiple processes to
share a single CPU. Context switching is an essential part of a multitasking operating
system features.
When the scheduler switches the CPU from executing one process to execute another,
the state from the current running process is stored into the process control block. After
this, the state for the process to run next is loaded from its own PCB and used to set the
PC, registers, etc. At that point, the second process can start executing.
Context switches are computationally intensive since register and memory state must be
saved and restored. To avoid the amount of context switching time, some hardware
systems employ two or more sets of processor registers. When the process is switched,
the following information is stored for later use.

• Program Counter
• Scheduling information
• Base and limit register value
• Currently used register
• Changed State
• I/O State information
• Accounting information
Context Switching in OS
The process of context switching involves the storage of the context/state of a given process in a
way that it can be reloaded whenever required, and its execution can be then resumed from the
very same point as earlier. It is basically a feature of the multitasking OS, and it allows the
sharing of just a single CPU by multiple processes.

In this article, we will look more into the Context Switching in OS according to the GATE
Syllabus for (Computer Science Engineering) CSE. We will read ahead to find out more about it.

Table of Contents

• What is Context Switching in OS?


• Why Do We Need Context Switching?
• Examples of Context Switching
Context Switching Triggers
Steps of Context Switching

What is Context Switching in OS?


Context switching refers to a technique/method used by the OS to switch processes from a given
state to another one for the execution of its function using the CPUs present in the system. When
switching is performed in the system, the old running process’s status is stored as registers, and
the CPU is assigned to a new process for the execution of its tasks. While new processes are
running in a system, the previous ones must wait in the ready queue. The old process’s execution
begins at that particular point at which another process happened to stop it. It describes the
features of a multitasking OS where multiple processes share a similar CPU to perform various
tasks without the requirement for further processors in the given system.

Why Do We Need Context Switching?


Context switching helps in sharing a single CPU among all processes so as to complete the
execution and store the status of the system’s tasks. Whenever the process reloads in a system, its
execution begins at the very point in which there is conflict.

Here are some of the reasons why an OS would need context switching:

1. The switching of a given process to another one isn’t directly in the system. Context switching
helps an OS switch between multiple processes to use the resources of the CPU for
accomplishing its tasks and storing its context. The service of a process can be resumed at the
same point later on. In case we don’t store the data or context of the currently running process,
this stored info may be lost when switching between the given processes.
2. In case a high-priority process is falling in a ready queue, the process running currently would be
shut down/stopped with the help of a high-priority process for completing its tasks in a system.
3. In case a running process needs various I/O resources in a system, another process will switch
the current process if it wants to use the CPU. And when it meets the I/O requirements, the
previous process would go into a ready state so that it can wait for the CPU execution. Context
switching helps in storing the process’s state to resume the tasks in an OS. Else, the process has
to restart the execution from the very initial levels.
4. In case an interrupt occurs when a process runs in an OS, the status of the process is saved as
the registers using context switching. After the interrupts are resolved, the process would switch
from a wait to a ready state so as to resume its execution later at the very same point at which
the OS interrupt occurs.
5. Using context switching, a single CPU can simultaneously handle various process requests
without requiring any additional processors.

Examples of Context Switching


Suppose that numerous processes get stored in a PCB (Process Control Block), and a process is
in its running state for the execution of its task using the CPUs. As this process runs, another one
arrives in the ready queue, which has a comparatively higher priority of completing its assigned
task using the system’s CPU. In this case, we used context switching such that it switches the
currently running process with the new one that requires the CPU to finish its assigned tasks.
When a process is switching, the context switch saves the old process’s status in registers.
Whenever any process reloads into a CPU, it initiates the process execution in which the new
process intercepts the old process. In case we don’t save the process’s state, we have to begin its
execution at its initial level. This way, context switching helps an OS to switch between the
given processes and reload or store the process.

Context Switching Triggers


Here are the triggers that lead to context switches in a system:

Interrupts: The CPU requests the data to be read from a disk. In case there are interrupts, the
context switching would automatically switch a part of the hardware that needs less time to
handle the interrupts.
Multitasking: Context switching is the characteristic of multitasking. They allow a process to
switch from the CPU to allow another process to run. When switching a given process, the old
state gets saved so as to resume the execution of the process at the very same point in a system.

Kernel/User Switch: It’s used in the OS when it is switching between the kernel mode and the
user mode.

What is the PCB?


PCB refers to a data structure that is used in the OS to store all the data-related info to the
process. For instance, whenever a process is formed in the OS, the info of the process is updated,
information about the process switching, and the process terminated in the PCB.

Steps of Context Switching


Several steps are involved in the context switching of a process. The diagram given below
represents context switching between two processes, P1 and P2, in case of an interrupt, I/O need,
or the occurrence of a priority-based process in the PCB’s ready queue.
As you can see in the illustration above, the process P1 is initially running on the CPU for the
execution of its task. At the very same time, P2, another process, is in its ready state. If an
interruption or error has occurred or if the process needs I/O, the P1 process would switch the
state from running to waiting.

Before the change of the state of the P1 process, context switching helps in saving the context of
the P1 process as registers along with the program counter (to PCB1). Then it loads the P2
process state from its ready state (of PCB2) to its running state.

Here are the steps are taken to switch the P1 to P2:

1. The context switching must save the P1’s state as the program counter and register to PCB that
is in its running state.
2. Now it updates the PCB1 to the process P1 and then moves the process to its appropriate
queue, like the ready queue, waiting queue and I/O queue.
3. Then, another process enters the running state. We can also select a new process instead of
from the ready state that needs to be executed or when a process has a higher priority of
executing its task.
4. Thus, now we need to update the PCB for the P2 selected process. It involves switching a given
process state from its running state or from any other state, such as exit, blocked, or suspended.
5. In case the CPU already performs the execution of the P2 process, then we must get the P2
process’s status so as to resume the execution of it at the very same time at the same point at
which there’s a system interrupt.

In a similar manner, the P2 process is switched off from the system’s CPU to let the process P1
resume its execution. The process P1 is reloaded to the running state from PCB1 to resume its
assigned task at the very same point. Else, the data is lost, so when the process is again executed,
it starts the execution at its initial level.

Process Scheduler schedules


A Process Scheduler schedules different processes to be assigned to the CPU based on
particular scheduling algorithms. There are six popular process scheduling algorithms
which we are going to discuss in this chapter −

• First-Come, First-Served (FCFS) Scheduling


• Shortest-Job-Next (SJN) Scheduling
• Priority Scheduling
• Shortest Remaining Time
• Round Robin(RR) Scheduling
• Multiple-Level Queues Scheduling
These algorithms are either non-preemptive or preemptive. Non-preemptive algorithms
are designed so that once a process enters the running state, it cannot be preempted
until it completes its allotted time, whereas the preemptive scheduling is based on priority
where a scheduler may preempt a low priority running process anytime when a high
priority process enters into a ready state.

First Come First Serve (FCFS)


• Jobs are executed on first come, first serve basis.
• It is a non-preemptive, pre-emptive scheduling algorithm.
• Easy to understand and implement.
• Its implementation is based on FIFO queue.
• Poor in performance as average wait time is high.

Wait time of each process is as follows −

Process Wait Time : Service Time - Arrival Time

P0 0-0=0

P1 5-1=4

P2 8-2=6

P3 16 - 3 = 13

Average Wait Time: (0+4+6+13) / 4 = 5.75

Shortest Job Next (SJN)


• This is also known as shortest job first, or SJF
• This is a non-preemptive, pre-emptive scheduling algorithm.
• Best approach to minimize waiting time.
• Easy to implement in Batch systems where required CPU time is known in
advance.
• Impossible to implement in interactive systems where required CPU time is not
known.
• The processer should know in advance how much time process will take.
Given: Table of processes, and their Arrival time, Execution time

Process Arrival Time Execution Time Service Time

P0 0 5 0

P1 1 3 5

P2 2 8 14

P3 3 6 8

Waiting time of each process is as follows −

Process Waiting Time

P0 0-0=0

P1 5-1=4

P2 14 - 2 = 12

P3 8-3=5

Average Wait Time: (0 + 4 + 12 + 5)/4 = 21 / 4 = 5.25


Priority Based Scheduling
• Priority scheduling is a non-preemptive algorithm and one of the most common
scheduling algorithms in batch systems.
• Each process is assigned a priority. Process with highest priority is to be executed
first and so on.
• Processes with same priority are executed on first come first served basis.
• Priority can be decided based on memory requirements, time requirements or any
other resource requirement.
Given: Table of processes, and their Arrival time, Execution time, and priority. Here we
are considering 1 is the lowest priority.

Process Arrival Time Execution Time Priority Service Time

P0 0 5 1 0

P1 1 3 2 11

P2 2 8 1 14

P3 3 6 3 5

Waiting time of each process is as follows −

Process Waiting Time

P0 0-0=0

P1 11 - 1 = 10

P2 14 - 2 = 12
P3 5-3=2

Average Wait Time: (0 + 10 + 12 + 2)/4 = 24 / 4 = 6

Shortest Remaining Time


• Shortest remaining time (SRT) is the preemptive version of the SJN algorithm.
• The processor is allocated to the job closest to completion but it can be preempted
by a newer ready job with shorter time to completion.
• Impossible to implement in interactive systems where required CPU time is not
known.
• It is often used in batch environments where short jobs need to give preference.
Round Robin Scheduling
• Round Robin is the preemptive process scheduling algorithm.
• Each process is provided a fix time to execute, it is called a quantum.
• Once a process is executed for a given time period, it is preempted and other
process executes for a given time period.
• Context switching is used to save states of preempted processes.

Wait time of each process is as follows −

Process Wait Time : Service Time - Arrival Time

P0 (0 - 0) + (12 - 3) = 9

P1 (3 - 1) = 2

P2 (6 - 2) + (14 - 9) + (20 - 17) = 12

P3 (9 - 3) + (17 - 12) = 11

Average Wait Time: (9+2+12+11) / 4 = 8.5

Multiple-Level Queues Scheduling


Multiple-level queues are not an independent scheduling algorithm. They make use of
other existing algorithms to group and schedule jobs with common characteristics.
• Multiple queues are maintained for processes with common characteristics.
• Each queue can have its own scheduling algorithms.
• Priorities are assigned to each queue.
For example, CPU-bound jobs can be scheduled in one queue and all I/O-bound jobs in
another queue. The Process Scheduler then alternately selects jobs from each queue
and assigns them to the CPU based on the algorithm assigned to the queue.

CPU Scheduling Criteria


Different CPU scheduling algorithms have different properties and the choice of a
particular algorithm depends on various factors. Many criteria have been suggested for
comparing CPU scheduling algorithms.
The criteria include the following:
1. CPU utilization: The main objective of any CPU scheduling algorithm is to keep the
CPU as busy as possible. Theoretically, CPU utilization can range from 0 to 100 but
in a real-time system, it varies from 40 to 90 percent depending on the load upon the
system.
2. Throughput: A measure of the work done by the CPU is the number of processes
being executed and completed per unit of time. This is called throughput. The
throughput may vary depending on the length or duration of the processes.
3. Turnaround time: For a particular process, an important criterion is how long it
takes to execute that process. The time elapsed from the time of submission of a
process to the time of completion is known as the turnaround time. Turn-around time
is the sum of times spent waiting to get into memory, waiting in the ready queue,
executing in CPU, and waiting for I/O. The formula to calculate Turn Around Time =
Completion Time – Arrival Time.
4. Waiting time: A scheduling algorithm does not affect the time required to complete
the process once it starts execution. It only affects the waiting time of a process i.e.
time spent by a process waiting in the ready queue. The formula for calculating
Waiting Time = Turnaround Time – Burst Time.
5. Response time: In an interactive system, turn-around time is not the best criterion. A
process may produce some output fairly early and continue computing new results
while previous results are being output to the user. Thus another criterion is the time
taken from submission of the process of the request until the first response is
produced. This measure is called response time. The formula to calculate Response
Time = CPU Allocation Time(when the CPU was allocated for the first) – Arrival
Time
6. Completion time: The completion time is the time when the process stops
executing, which means that the process has completed its burst time and is
completely executed.
7. Priority: If the operating system assigns priorities to processes, the scheduling
mechanism should favor the higher-priority processes.
8. Predictability: A given process always should run in about the same amount of time
under a similar system load.

You might also like