Unit 2 Processes: Structure Page Nos
Unit 2 Processes: Structure Page Nos
UNIT 2 PROCESSES
Structure Page Nos.
2.0 Introduction 25
2.1 Objectives 26
2.2 The Concept of Process 26
2.2.1 Implicit and Explicit Tasking
2.2.2 Processes Relationship
2.2.3 Process States
2.2.4 Implementation of Processes
2.2.5 Process Hierarchy
2.2.6 Threads
2.2.7 Levels of Threads
2.3 System Calls for Process Management 34
2.4 Process Scheduling 35
2.4.1 Scheduling Objectives
2.4.2 Types of Schedulers
2.4.3 Scheduling Criteria
2.5 Scheduling Algorithms 38
2.5.1 First Come First Serve (FCFS)
2.5.2 Shortest Job First (SJF)
2.5.3 Round Robin (RR)
2.5.4 Shortest Remaining Time Next (SRTN)
2.5.5 Priority Based Scheduling or Event Driven (ED) Scheduling
2.6 Performance Evaluation of the Scheduling Algorithms 43
2.7 Summary 45
2.8 Solutions/Answers 45
2.9 Further Readings 48
2.0 INTRODUCTION
In the earlier unit we have studied the overview and the functions of an operating
system. In this unit we will have detailed discussion on the processes and their
management by the operating system. The other resource management features of
operating systems will be discussed in the subsequent units.
The CPU executes a large number of programs. While its main concern is the
execution of user programs, the CPU is also needed for other system activities. These
activities are called processes. A process is a program in execution. Typically, a batch
job is a process. A time-shared user program is a process. A system task, such as
spooling, is also a process. For now, a process may be considered as a job or a
time-shared program, but the concept is actually more general.
In general, a process will need certain resources such as CPU time, memory, files, I/O
devices, etc., to accomplish its task. These resources are given to the process when it
is created. In addition to the various physical and logical resources that a process
obtains when it is created, some initialisation data (input) may be passed along. For
example, a process whose function is to display the status of a file, say F1, on the
screen, will get the name of the file F1 as an input and execute the appropriate
program to obtain the desired information.
We emphasize that a program by itself is not a process; a program is a passive entity,
while a process is an active entity. It is known that two processes may be associated
with the same program; they are nevertheless considered two separate execution
sequences.
A process is the unit of work in a system. Such a system consists of a collection of
processes, some of which are operating system processes, those that execute system
25
Introduction to Operating code, and the rest being user processes, those that execute user code. All of those
Systems and Process
Management
processes can potentially execute concurrently.
The operating system is responsible for the following activities in connection with
processes managed.
• The creation and deletion of both user and system processes;
• The suspension is resumption of processes;
• The provision of mechanisms for process synchronization, and
• The provision of mechanisms for deadlock handling.
We will learn the operating system view of the processes, types of schedulers,
different types of scheduling algorithms, in the subsequent sections of this unit.
2.1 OBJECTIVES
After going through this unit, you should be able to:
• understand the concepts of process, various states in the process and their
scheduling;
• define a process control block;
• classify different types of schedulers;
• understand various types of scheduling algorithms, and
• compare the performance evaluation of the scheduling algorithms.
Explicit tasking means that programs explicitly define each process and some of its
attributes. To improve the performance, a single logical application is divided into
various related processes. Explicit tasking is used in situations where high
performance in desired system programs such as parts of the operating system and real
time applications are common examples of programs defined processes. After
dividing process into several independent processes, a system programs defines the
confines of each individual process. A parent process is then commonly added to
create the environment for and to control execution of individual processes.
27
Introduction to Operating Common reasons for and uses of explicit tasking include:
Systems and Process
Management • Speedup: Explicit tasking can result in faster execution of applications.
• Driving I/O devices that have latency: While one task is waiting for I/O to
complete, another portion of the application can make progress towards
completion if it contains other tasks that can do useful work in the interim.
• User convenience: By creating tasks to handle individual actions, a graphical
interface can allow users to launch several operations concurrently by clicking
on action icons before completion of previous commands.
• Multiprocessing and multicomputing: A program coded as a collection of
tasks can be relatively easily posted to a multiprocessor system, where
individual tasks may be executed on different processors in parallel.
Distributed computing network server can handle multiple concurrent client sessions
by dedicating an individual task to each active client session.
2.2.2 Processes Relationship
In the concurrent environment basically processes have two relationships, competition
and cooperation. In the concurrent environment, processes compete with each other
for allocation of system resources to execute their instructions. In addition, a
collection of related processes that collectively represent a single logical application
cooperate with each other. There should be a proper operating system to support these
relations. In the competition, there should be proper resource allocation and protection
in address generation.
We distinguish between independent process and cooperating process. A process is
independent if it cannot affect or be affected by other processes executing in the
system.
Independent process: These type of processes have following features:
• Their state is not shared in any way by any other process.
• Their execution is deterministic, i.e., the results of execution depend only on the
input values.
• Their execution is reproducible, i.e., the results of execution will always be the
same for the same input.
• Their execution can be stopped and restarted without any negative effect.
Cooperating process: In contrast to independent processes, cooperating processes can
affect or be affected by other processes executing the system. They are characterised
by:
The transition of the process states are shown in Figure 1.and their corresponding
transition are described below:
Context Switch
A context switch (also sometimes referred to as a process switch or a task switch) is
the switching of the CPU (central processing unit) from one process or to another.
A context is the contents of a CPU's registers and program counter at any point in
time.
30
A context switch is sometimes described as the kernel suspending execution of one Processes
process on the CPU and resuming execution of some other process that had previously
been suspended.
Context Switch: Steps
In a context switch, the state of the first process must be saved somehow, so that,
when the scheduler gets back to the execution of the first process, it can restore this
state and continue normally.
The state of the process includes all the registers that the process may be using,
especially the program counter, plus any other operating system specific data that may
be necessary. Often, all the data that is necessary for state is stored in one data
structure, called a process control block (PCB). Now, in order to switch processes, the
PCB for the first process must be created and saved. The PCBs are sometimes stored
upon a per-process stack in the kernel memory, or there may be some specific
operating system defined data structure for this information.
Let us understand with the help of an example. Suppose if two processes A and B are
in ready queue. If CPU is executing Process A and Process B is in wait state. If an
interrupt occurs for Process A, the operating system suspends the execution of the first
process, and stores the current information of Process A in its PCB and context to the
second process namely Process B. In doing so, the program counter from the PCB of
Process B is loaded, and thus execution can continue with the new process. The
switching between two processes, Process A and Process B is illustrated in the
Figure 3 given below:
31
Introduction to Operating A parent process can terminate the execution of one of its children for one of these
Systems and Process
Management
reasons:
• The child process has exceeded its usage of the resources it has been
allocated. In order to do this, a mechanism must be available to allow the
parent process to inspect the state of its children processes.
Old or primitive operating system like MS-DOS are not multiprogrammed so when
one process starts another, the first process is automatically blocked and waits until
the second is finished.
2.2.6 Threads
Threads, sometimes called lightweight processes (LWPs) are independently scheduled
parts of a single program. We say that a task is multithreaded if it is composed of
several independent sub-processes which do work on common data, and if each of
those pieces could (at least in principle) run in parallel.
If we write a program which uses threads – there is only one program, one executable
file, one task in the normal sense. Threads simply enable us to split up that program
into logically separate pieces, and have the pieces run independently of one another,
until they need to communicate. In a sense, threads are a further level of object
orientation for multitasking systems. They allow certain functions to be executed in
parallel with others.
On a truly parallel computer (several CPUs) we might imagine parts of a program
(different subroutines) running on quite different processors, until they need to
communicate. When one part of the program needs to send data to the other part, the
two independent pieces must be synchronized, or be made to wait for one another. But
what is the point of this? We can always run independent procedures in a program as
separate programs, using the process mechanisms we have already introduced. They
could communicate using normal interprocesses communication. Why introduce
another new concept? Why do we need threads?
The point is that threads are cheaper than normal processes, and that they can be
scheduled for execution in a user-dependent way, with less overhead. Threads are
32
cheaper than a whole process because they do not have a full set of resources each. Processes
Whereas the process control block for a heavyweight process is large and costly to
context switch, the PCBs for threads are much smaller, since each thread has only a
stack and some registers to manage. It has no open file lists or resource lists, no
accounting structures to update. All of these resources are shared by all threads within
the process. Threads can be assigned priorities – a higher priority thread will get put to
the front of the queue. Let’s define heavy and lightweight processes with the help of a
table.
Object Resources
Thread (Light Weight Processes) Stack, set of CPU registers, CPU time
Task (Heavy Weight Processes) 1 thread, PCB, Program code, memory
segment etc.
Multithreaded task n- threads, PCB, Program code, memory
segment etc.
Why use threads?
The sharing of resources can be made more effective if the scheduler knows known
exactly what each program was going to do in advance. Of course, the scheduling
algorithm can never know this – but the programmer who wrote the program does
know. Using threads it is possible to organise the execution of a program in such a
way that something is always being done, whenever the scheduler gives the
heavyweight process CPU time.
33
Introduction to Operating • Solaris 2 (user and system level)
Systems and Process
Management • OS/2 (system level only)
• NT threads (user and system level)
• IRIX threads
• POSIX standardized user threads interface.
1) Explain the difference between a process and a thread with some examples.
……………………………………………………………………………..
…………..…………………………………………………………………
……………………………………………………………………………..
2) Identify the different states a live process may occupy and show how a process
moves between these states.
……………………………………………………………………………………
…..………………………………………………………………………………
…………………………………………………………………………………..
3) Define what is meant by a context switch. Explain the reason many systems use
two levels of scheduling.
……………………………………………………………………………………
…..………………………………………………………………………………..
……………………………………………………………………………………
34
Create: In response to the create call the operating system creates a new process with Processes
the specified or default attributes and identifier. Some of the parameters definable at
the process creation time include:
• level of privilege, such as system or user
• priority
• size and memory requirements
• maximum data area and/or stack size
• memory protection information and access rights
• other system dependent data.
Delete: The delete service is also called destroy, terminate or exit. Its execution
causes the operating system to destroy the designated process and remove it from the
system.
Abort: It is used to terminate the process forcibly. Although a process could
conceivably abort itself, the most frequent use of this call is for involuntary
terminations, such as removal of malfunctioning process from the system.
Fork/Join: Another method of process creation and termination is by means of
FORK/JOIN pair, originally introduced as primitives for multiprocessor system. The
FORK operations are used to split a sequence of instruction into two concurrently
executable sequences. JOIN is used to merge the two sequences of code divided by
the FORK and it is available to a parent process for synchronization with a child.
Suspend: The suspend system call is also called BLOCK in some systems. The
designated process is suspended indefinitely and placed in the suspend state. A
process may be suspended itself or another process when authorised to do so.
Resume: The resume system call is also called WAKEUP in some systems. This call
resumes the target process, which is presumably suspended. Obviously a suspended
process can not resume itself because a process must be running to have its operating
system call processed. So a suspended process depends on a partner process to issue
the resume.
Delay: The system call delay is also known as SLEEP. The target process is
suspended for the duration of the specified time period. The time may be expressed in
terms of system clock ticks that are system dependent and not portable or in standard
time units such as seconds and minutes. A process may delay itself or optionally,
delay some other process.
Get_Attributes: It is an enquiry to which the operating system responds by providing
the current values of the process attributes, or their specified subset, from the PCB.
Change Priority: It is an instance of a more general SET-PROCESS-ATTRIBUTES
system call. Obviously, this call is not implemented in systems where process priority
is static.
Short term scheduler: The short-term scheduler selects the process for the processor
among the processes which are already in queue (in memory). The scheduler will
execute quite frequently (mostly at least once every 10 milliseconds). It has to be very
fast in order to achieve a better processor utilisation. The short term scheduler, like all
other OS programs, has to execute on the processor. If it takes 1 millisecond to choose
a process that means ( 1 / ( 10 + 1 )) = 9% of the processor time is being used for short
time scheduling and only 91% may be used by processes for execution.
36
Long term scheduler: The long-term scheduler selects processes from the process Processes
pool and loads selected processes into memory for execution. The long-term scheduler
executes much less frequently when compared with the short term scheduler. It
controls the degree of multiprogramming (no. of processes in memory at a time). If
the degree of multiprogramming is to be kept stable (say 10 processes at a time), the
long-term scheduler may only need to be invoked till the process finishes execution.
The long-term scheduler must select a good process mix of I/O-bound and processor
bound processes. If most of the processes selected are I/O-bound, then the ready
queue will almost be empty, while the device queue(s) will be very crowded. If most
of the processes are processor-bound, then the device queue(s) will almost be empty
while the ready queue is very crowded and that will cause the short-term scheduler to
be invoked very frequently. Time-sharing systems (mostly) have no long-term
scheduler. The stability of these systems either depends upon a physical limitation
(no. of available terminals) or the self-adjusting nature of users (if you can’t get
response, you quit). It can sometimes be good to reduce the degree of
multiprogramming by removing processes from memory and storing them on disk.
These processes can then be reintroduced into memory by the medium-term
scheduler. This operation is also known as swapping. Swapping may be necessary to
improve the process mix or to free memory.
37
Introduction to Operating Waiting time = Turnaround Time - Processing Time
Systems and Process
Management But the scheduling algorithm affects or considers the amount of time that a process
spends waiting in a ready queue. Thus rather than looking at turnaround time waiting
time is usually the waiting time for each process.
Response time: It is most frequently considered in time sharing and real time
operating system. However, its characteristics differ in the two systems. In time
sharing system it may be defined as interval from the time the last character of a
command line of a program or transaction is entered to the time the last result appears
on the terminal. In real time system it may be defined as interval from the time an
internal or external event is signalled to the time the first instruction of the respective
service routine is executed.
Response time = t(first response) – t(submission of request)
One of the problems in designing schedulers and selecting a set of its performance
criteria is that they often conflict with each other. For example, the fastest response
time in time sharing and real time system may result in low CPU utilisation.
Throughput and CPU utilization may be increased by executing the large number of
processes, but then response time may suffer. Therefore, the design of a scheduler
usually requires balance of all the different requirements and constraints. In the next
section we will discuss various scheduling algorithms.
38
Non–Preemptive scheduling: A scheduling discipline is non-preemptive if once a Processes
process has been allotted to the CPU, the CPU cannot be taken away from the process.
A non-preemptible discipline always processes a scheduled request to its completion.
In non-preemptive systems, jobs are made to wait by longer jobs, but the treatment of
all processes is fairer.
First come First Served (FCFS) and Shortest Job First (SJF), are considered to be the
non-preemptive scheduling algorithms.
The decision whether to schedule preemptive or not depends on the environment and
the type of application most likely to be supported by a given operating system.
P1 P2 P3 P4 P5
0 3 6 7 11 13
Using SJF scheduling because the shortest length of process will first get execution,
the Gantt chart will be:
P4 P1 P3 P2
0 3 9 16 24
Because the shortest processing time is of the process P4, then process P1 and then P3
and Process P2. The waiting time for process P1 is 3 ms, for process P2 is 16 ms, for
process P3 is 9 ms and for the process P4 is 0 ms as –
40
Each process is allocated a small time-slice called quantum. No process can run for Processes
more than one quantum while others are waiting in the ready queue. If a process needs
more CPU time to complete after exhausting one quantum, it goes to the end of ready
queue to await the next allocation. To implement the RR scheduling, Queue data
structure is used to maintain the Queue of Ready processes. A new process is added at
the tail of that Queue. The CPU schedular picks the first process from the ready
Queue, Allocate processor for a specified time Quantum. After that time the CPU
schedular will select the next process is the ready Queue.
Consider the following set of process with the processing time given in milliseconds.
P1 P2 P3 P1 P1 P1 P1 P1
0 4 7 10 14 18 22 26 30
Consider the set of four processes arrived as per timings described in the table:
At time 0, only process P1 has entered the system, so it is the process that executes. At
time 1, process P2 arrives. At that time, process P1 has 4 time units left to execute At
41
Introduction to Operating this juncture process 2’s processing time is less compared to the P1 left out time (4
Systems and Process
Management
units). So P2 starts executing at time 1. At time 2, process P3 enters the system with
the processing time 5 units. Process P2 continues executing as it has the minimum
number of time units when compared with P1 and P3. At time 3, process P2
terminates and process P4 enters the system. Of the processes P1, P3 and P4, P4 has
the smallest remaining execution time so it starts executing. When process P1
terminates at time 10, process P3 executes. The Gantt chart is shown below:
P1 P2 P4 P1 P3
0 1 3 6 10 15
Turnaround time for each process can be computed by subtracting the time it
terminated from the arrival time.
Turn around Time = t(Process Completed)– t(Process Submitted)
The turnaround time for each of the processes is:
P1: 10 – 0 = 10
P2: 3–1= 2
P3: 15 – 2 = 13
P4: 6–3= 3
The average turnaround time is (10+2+13+3) / 4 = 7
The waiting time can be computed by subtracting processing time from turnaround
time, yielding the following 4 results for the processes as
P1: 10 – 5 = 5
P2: 2–2=0
P3: 13 – 5 = 8
P4: 3–3=0
The average waiting time = (5+0+8+0) / 4 = 3.25milliseconds
Four jobs executed in 15 time units, so throughput is 15 / 4 = 3.75 time units/job.
2.5.5 Priority Based Scheduling or Event-Driven (ED) Scheduling
A priority is associated with each process and the scheduler always picks up the
highest priority process for execution from the ready queue. Equal priority processes
are scheduled FCFS. The level of priority may be determined on the basis of resource
requirements, processes characteristics and its run time behaviour.
A major problem with a priority based scheduling is indefinite blocking of a lost
priority process by a high priority process. In general, completion of a process within
finite time cannot be guaranteed with this scheduling algorithm. A solution to the
problem of indefinite blockage of low priority process is provided by aging priority.
Aging priority is a technique of gradually increasing the priority of processes (of low
priority) that wait in the system for a long time. Eventually, the older processes attain
high priority and are ensured of completion in a finite time.
As an example, consider the following set of five processes, assumed to have arrived
at the same time with the length of processor timing in milliseconds: –
P2 P5 P1 P3 P4
0 1 6 16 18 19
P1 P2 P3 P4 P5
0 10 39 42 49 61
P3 P4 P1 P5 P2
0 3 10 20 32 61
P1 P2 P3 P4 P5 P2 P5 P2
0 10 20 23 30 40 49 52 61
……………………………………………………………………………
……………………………………………………………………………
……………………………………………………………………………
4) For the given five processes arriving at time 0, in the order with the length of
CPU time in milliseconds:
2.7 SUMMARY
A process is an instance of a program in execution. A process can be defined by the
system or process is an important concept in modem operating system. Processes
provide a suitable means for informing the operating system about independent
activities that may be scheduled for concurrent execution. Each process is represented
by a process control block (PCB). Several PCB’s can be linked together to form a
queue of waiting processes. The selection and allocation of processes is done by a
scheduler. There are several scheduling algorithms. We have discussed FCFS, SJF,
RR, SJRT and priority algorithms along with the performance evaluation.
2.8 SOLUTIONS/ANSWERS
P1 P2 P3
0 13 21 104
P1 P2 P3 P4 P5
0 10 39 42 49 51
P3 P4 P1 P5 P2
0 3 10 20 32 61
From the above calculations of average waiting time we found that SJF policy results
in less than one half of the average waiting time obtained from FCFS, while Round
Robin gives intermediate result.
48