Download as pptx, pdf, or txt
Download as pptx, pdf, or txt
You are on page 1of 70

Unit - III

Memory management and Virtual


memory
Contents :
• Swapping
• Demand paging
• Hybrid System with swapping and demand paging
• Memory management requirements
• Memory partitioning
• Paging
• Segmentation
• Security Issues
• Hardware and control structures
• Operating system software
• Linux & Windows memory management
Process State Transition :
User Running

system call,
1
interrupt, return
interrupt to user
interrupt return return
Preempted
Kernel
2 Running
7
9 preempt
reschedule
Zombie process
sleep
Ready to
4 wakeup
3 Run in Memory
Asleep enough mem
in Created
fork
Memory swap
out swap
out
swap
in
8
not enough mem
6 wakeup 5 (swapping system only)

Sleep, Swapped Ready to Run, Swapped


Swapping :

Fig. Swapping of processes


Swap Space :
• Swap device -> block device configuration

• Allocation temporary not like files.

• I/O faster for group of block

• Map data structure

• Follows first fit algorithm of continuous blocks


Allocating Swap Space :

Address Unit
Allocate 100 unit
1 10000 101 9900
Map

Allocate 50 unit

Allocate 100 unit


251 9750 151 9850
Allocate Swap Space :
• malloc( address_of_map, number_of_unit)
• for (every map entry)
• if (current map entry can fit requested units)
• if (requested units == number of units in entry)
• Delete entry from map
• else
• Adjust start address of entry
• return original address of entry
• return -1
Freeing Swap Space :
Address Unit
101 50
50 unit free at 101
251 9750
251 9750
Map

Case 1: Free resources fill a hole,


but not contiguous to any resources in the map
Freeing Swap Space :
Address Unit 101 50
50 unit free at 101
251 9750 251 9750
Map
100 unit free at 1

1 150

251 9750

Case 2: Free resources fill a hole,


and immediately precedes an entry in the map
Freeing Swap Space :
Address Unit 101 50
50 unit free at 101
251 9750 251 9750
Map
100 unit free at 1

1 150 Allocate 200 unit 1 150

451 9550 251 9750

300 unit free at 151 Case 3: Free resources fill a


hole, and completely fills the
1 10000 gap between entries in the
map
Swapping Process Out
• Memory  Swap device
• Kernel swap out when it needs memory
1. When fork() called for allocate child process
2. When called for increase the size of process
3. When process become larger by growth of its stack
4. Previously swapped out process want to swap in but not
enough memory
Swapping Process Out
Virtual Addresses Physical Addresses Swap device
684
Text 0 278k
1k 432k
:
Data 65k 573k
66k 595k
688
:
Stack 128k 401k
:

Fig. Mapping process onto the swap device


Swapping Process In
Virtual Addresses Physical Addresses Swap device
684
Text 0 278k
1k 432k
:
Data 65k 573k
66k 595k
688
:
Stack 128k 401k
:

Fig. Swapping a process into memory


Fork Swap
• There may not be enough memory when fork() called

• Child process swap out and “ready-to-run”

• Swap in when kernel schedule it


Expansion Swap
• Reserves enough space on the swap device
• Adjust the address translation mapping of the process (virtual
address)
• Swaps out on newly allocated space in swapping device
• When the process swaps the process into memory, it will
allocate physical memory according to new address
translation map
Swapping Operation :
Demand paging :
• Not all page of process resides in memory

• Principle of locality

• Working set, page fault, page hit.

• Page replacement strategies (LRU/OPR)

• The kernel suspends the execution of the process until it reads


the page into memory and makes it accessible to the process
Data Structure for Demand Paging :
• Page table entry

• Disk block descriptors

• Page frame data table (pfdata)

• Swap use table


Page Table Entry and Disk Block
Descriptor
Region Page Table

Page Table Entry Disk Block Descriptor

Page Table Entry

Page address age Cp/wrt mod ref val prot

Disk Block Descriptor

Swap device Block num Type(swap,file,demand


fill 0/1)
Page Table Entry
• Contains the physical address of page and the
following bits:
• Valid: whether the page content legal
• Reference: whether the page is referenced recently
• Modify: whether the page content is modified
• copy on write: kernel must create a new copy when a
process modifies its content (required for fork)
• Age: Age of the page
• Protection: Read/ write permission
pftable entries :
• Page state- on swap device or executable file
• Reference count-no of processes referencing current page
• Logical device and block no
• Pointers to other pftable entries
Memory Management
Requirements
• Memory management is intended to satisfy the
following requirements:
• Relocation
• Protection
• Sharing
• Logical organization
• Physical organization
Relocation

• Relocation is the process of adjusting program


addresses to match the actual physical addresses
where the program resides when it executes
• Why is relocation needed?
• Programmer/translator don’t know which other
programs will be memory resident when
the program executes
Relocation
• Why is relocation needed? (continued)
• Active processes need to be able to be swapped in and out
of main memory in order to maximize processor utilization
• Specifying that a process must be placed in the same
memory region when it is swapped back
in would be limiting
• Consequently it must be possible to
adjust addresses whenever a program
is loaded.
Protection
• Processes need to acquire permission to reference memory
locations for reading or writing purposes
• Location of a program in main memory is unpredictable
• Memory references generated by a process must be checked at run
time
• Mechanisms that support relocation also support protection
Sharing
• Advantageous to allow each process access to the same copy of the
program rather than have their own separate copy
• Memory management must allow controlled access to shared areas
of memory without compromising protection
• Mechanisms used to support relocation support sharing capabilities
Logical Organization

• Main memory is organized as a linear (1-D) address space


consisting of a sequence of bytes or words.
• Programs aren’t necessarily organized this way
• Paging versus segmentation

Programs are written in modules


• modules can be written and compiled independently
• different degrees of protection given to modules (read-
only, execute-only)
• sharing on a module level corresponds to the user’s way of
viewing the problem
Physical Organization

• Two-level memory for program storage:


• Disk (slow and cheap) & RAM (fast and more expensive)
• Main memory is volatile, disk isn’t
• User should not have to be responsible for
organizing movement of code/data between the
two levels.
Physical Organization

Cannot leave the


Memory available for a Programmer does not
programmer with the
program plus its data know how much space
responsibility to manage
may be insufficient will be available
memory

overlaying allows various


modules to be assigned
the same region of
memory but is time
consuming to program
Memory Partitioning
• Virtual memory management brings processes into main memory
for execution by the processor
 involves virtual memory
 based on segmentation and paging
• Partitioned memory management
 used in several variations in some now-obsolete operating
systems
 does not involve virtual memory
Fixed Partitioning
• Equal-size partitions
• The operating system can swap
out a process if all partitions are
full and no process is in the
Ready or Running state
• A program may be too big to fit in a partition
• program needs to be designed with the use of overlays
• Main memory utilization is inefficient
• any program, regardless of size, occupies an entire partition
• internal fragmentation
• wasted space due to the block of data loaded being
smaller than the partition
Unequal Size Partitions
• Using unequal size partitions helps lessen
the problems
• programs up to 16M can be accommodated
without overlays
• partitions smaller than 8M allow smaller
programs to be accommodated with less
internal fragmentation
Memory Assignment
Disadvantages :

• The number of partitions specified at system


generation time limits the number of active
processes in the system
• Small jobs will not utilize partition space efficiently
Dynamic Partitioning :

• Partitions are of variable length and number


• Process is allocated exactly as much memory as it
requires
• This technique was used by IBM’s mainframe operating
system.
Dynamic Partitioning :

External Fragmentation
• memory becomes more and more fragmented
• memory utilization declines

Compaction
• technique for overcoming external fragmentation
• OS shifts processes so that they are contiguous
• free memory is together in one block
• time consuming and wastes CPU time
Placement Algorithms

Best-fit First-fit Next-fit

• chooses the • begins to scan • begins to scan


block that is memory from memory from
closest in size the beginning the location of
to the request and chooses the last
the first placement and
available block chooses the
that is large next available
enough block that is
large enough
Buddy System
• Comprised of fixed and dynamic partitioning
schemes
• Space available for allocation is treated as a
single block
• Memory blocks are available of size 2K words,
L ≤ K ≤ U, where
• 2L = smallest size block that is allocated
• 2U = largest size block that is allocated; (generally 2U is the size of the
entire memory available for allocation)
Buddy System Example
Addresses
Logical

• reference to a memory location independent of the current


assignment of data to memory

Relative

• address is expressed as a location relative to some known


point

Physical or Absolute

• actual location in main memory


Paging :

• Partition memory into equal fixed-size chunks that are relatively


small
• Process is also divided into small fixed-size chunks of the same size

Pages Frames

• chunks of a • available
process chunks of
memory
Page Table
• Maintained by operating system for each process
• Contains the frame location for each page in the process
• Processor must know how to access the page table for the current
process
• Used by processor to produce a physical address
Logical Addresses
Logical-to-Physical Address
Translation - Paging
Segmentation

• A program can be subdivided into segments


 may vary in length
 there is a maximum length
• Addressing consists of two parts:
 segment number
 an offset
• Similar to dynamic partitioning
• Eliminates internal fragmentation
Logical-to-Physical Address
Translation - Segmentation
Security Issues
If a process has not
declared a portion of its
memory to be sharable,
then no other process
should have access to the
contents of that portion
of memory

If a process declares that a


portion of memory may be
shared by other designated
processes then the security
service of the OS must
ensure that only the
designated processes have
access
Buffer Overflow Attacks
• Security threat related to memory management
• Also known as a buffer overrun
• Can occur when a process attempts to store data beyond the limits
of a fixed-sized buffer
• One of the most prevalent and dangerous types of security attacks
Defending Against
Buffer Overflows
• Prevention
• Detecting and aborting
• Countermeasure categories:

Compile-time Defenses
• aim to harden programs to resist attacks in new
programs
Run-time Defenses
• aim to detect and abort attacks in existing programs
Translation Lookaside Buffer
• Each virtual memory reference can cause two physical
memory accesses
• One to fetch the page table
• One to fetch the data
• To overcome this problem a high-speed cache is set up for
page table entries
• Called a Translation Lookaside Buffer (TLB)
• Contains page table entries that have been most recently used
Translation Lookaside Buffer
Translation Lookaside Buffer
• Given a virtual address, processor examines the TLB
• If page table entry is present (TLB hit), the frame number is
retrieved and the real address is formed
• If page table entry is not found in the TLB (TLB miss), the page
number is used to index the process page table.

• First checks if page is already in main memory


• If not in main memory a page fault is issued
• The TLB is updated to include the new page entry
Translation Lookaside Buffer
Translation Lookaside Buffer
Translation Lookaside Buffer
Page Size
• Smaller page size, more pages required per process
• More pages per process means larger page tables
• Larger page tables means large portion of page tables in
virtual memory
• Small page size, large number of pages will be found in main
memory
• As time goes on during execution, the pages in memory will all
contain portions of the process near recent references. Page
faults low.
• Increased page size causes pages to contain locations further
from any recent reference. Page faults rise.
Basic Replacement Algorithms
• FIFO
• LRU
• OPR
• Clock Policy
• Additional bit called a use bit
• When a page is first loaded in memory, the use bit is set to 1
• When the page is referenced, the use bit is set to 1
• When it is time to replace a page, the first frame encountered
with the use bit set to 0 is replaced.
• During the search for replacement, each use bit set to 1 is
changed to 0
Clock Policy
Clock Policy
Resident Set Size
• Fixed-allocation
• Gives a process a fixed number of frames within which to execute
• When a page fault occurs, one of the pages of that process must
be replaced
• Variable-allocation
• Number of pages allocated to a process varies over the lifetime of
the process
Fixed Allocation, Local Replacement
Scope
• Decide ahead of time the amount of allocation to give a
process
• If allocation is too small, there will be a high page fault rate
• If allocation is too large there will be too few programs in main
memory
• Processor idle time
• Swapping
Variable Allocation, Local Replacement
Scope
• When new process added, allocate number of page frames
based on application type, program request, or other criteria
• When page fault occurs, select page from among the resident
set of the process that suffers the fault
• Reevaluate allocation from time to time
Variable Allocation, Global Replacement
Scope
• Easiest to implement
• Adopted by many operating systems
• Operating system keeps list of free frames
• Free frame is added to resident set of process when a page
fault occurs
• If no free frame, replaces one from another process
Cleaning Policy
• Demand cleaning
• A page is written out only when it has been selected for
replacement
• Precleaning
• Modified pages are written before their frame is needed
• Pages are written out in batches
Cleaning Policy
• Best approach uses page buffering
• Replaced pages are placed in two lists
• Modified and unmodified
• Pages in the modified list are periodically written out in batches
• Pages in the unmodified list are either reclaimed if referenced
again or lost when its frame is assigned to another page

You might also like