Other Issues – Page Size
Download
Report
Transcript Other Issues – Page Size
9.5 Allocation of Frames
Each process needs minimum number of pages
Example: machine with all memory reference: at least two
memory accesses per instruction. If indirect addressing, then
paging requires at least 3 frames per process
Example: IBM 370 – 6 pages to handle MVC (multiple
move) instruction:
instruction is 6 bytes, might span 2 pages
source block might straddle 2 pages
destination block might straddle 2 pages
Worst case: multiple level indirection
Two major allocation schemes
equal allocation
priority allocation
Operating System Principles
9.1
Silberschatz, Galvin and Gagne ©2005
Equal Allocation
Equal allocation – For example, if there are 100
frames and 5 processes, give each process 20 frames.
Proportional allocation – Allocate according to the size
of process
si size of process pi
S si
m total number of frames
s
ai allocation for pi i m
S
Ex ample :
m 64
si 10
s 2 127
10
64 5
137
127
64 59
137
a1
a2
Priority Allocation
Use a proportional allocation scheme using priorities
rather than size
Operating System Principles
9.2
Silberschatz, Galvin and Gagne ©2005
Global vs. Local Allocation
Global replacement – process selects a
replacement frame from the set of all frames; one
process can take a frame from another
Allows a high-priority process to increase its frame
allocation at the expense of a low-priority process
Low-priority process cannot control its own page-fault rate
Local replacement – each process selects from
only its own set of allocated frames
Global replacement generally results in greater
system throughput
Operating System Principles
9.3
Silberschatz, Galvin and Gagne ©2005
9.6 Thrashing (痛擊)
If a process does not have “enough” pages, the page-
fault rate is very high. This leads to:
low CPU utilization
operating system thinks that it needs to increase the degree of
multiprogramming
another process added to the system
Thrashing: a process is busy swapping pages in and
out. It spends more time in paging than executing.
Operating System Principles
9.4
Silberschatz, Galvin and Gagne ©2005
Operating System Principles
9.5
Silberschatz, Galvin and Gagne ©2005
Demand Paging and Thrashing
Why does demand paging work? Answer: Locality model
Process migrates from one locality to another
Localities may overlap
A process page faults when it changes locality
If we allocate fewer frames than size of the current locality, the process
will thrash
For a system, why does thrashing occur?
sum of size of locality > total memory size
We can limit the effect of thrashing by using a local replacement
algorithm
If processes are thrashing, the average service time for a page fault
will increase because of the longer queue for the paging device
Operating System Principles
9.6
Silberschatz, Galvin and Gagne ©2005
locality in a
memoryreference
pattern
Operating System Principles
9.7
Silberschatz, Galvin and Gagne ©2005
Working-Set Model
working-set window a fixed number of page references
Example: 10,000 instruction
WSSi (working set size of Process Pi) = total number of pages
referenced in the most recent (varies in time)
if too small will not encompass entire locality
if too large will encompass several localities
if = will encompass entire program
D = WSSi total demand frames of all processes
if D > m (total number of available frames) Thrashing
Policy: if D > m, then suspend one of the processes
Operating System Principles
9.8
Silberschatz, Galvin and Gagne ©2005
Keeping Track of the Working Set
Approximate with a fixed-interval timer interrupt + a
reference bit
Example: = 10,000 and Timer interrupts after every 5000 time units
Keep in memory 2 bits for each page
Whenever a timer interrupts copy and sets all values of current
reference bit to 0
If a page fault occurs, we can examine the current reference bit
and two in-memory reference bits to determine whether a page
was used during the last 10000 to 15000 references
If one of the bits in memory = 1 page in working set
Why is this not completely accurate?
Improvement = 10 bits and interrupt every 1000 time units
Operating System Principles
9.9
Silberschatz, Galvin and Gagne ©2005
Page-Fault Frequency Scheme
Establish “acceptable” page-fault rate
If actual rate too low, process loses frame
If actual rate too high, process gains frame
跳過 9.7
Operating System Principles
9.10
Silberschatz, Galvin and Gagne ©2005
9.8 Allocating Kernel Memory
Treated differently from user mode memory
Kernel requests memory for structures of varying sizes
Some of them are less than a page in size
Some kernel memory needs to be contiguous.
Ex: Hardware devices interact directly with physical memory
Often allocated from a free-memory pool different
from the list used to satisfy ordinary user-mode
processes
Operating System Principles
9.11
Silberschatz, Galvin and Gagne ©2005
Buddy System
Allocates memory from fixed-size segment consisting of
physically-contiguous pages
Memory allocated using power-of-2 allocator
Satisfies requests in units sized as power of 2
Request rounded up to next highest power of 2
When smaller allocation needed than current available, current
chunk split into two buddies of next-lower power of 2
Continue until appropriate sized chunk available
Advantage: adjacent buddies can be combined quickly
to form larger segment using coalescing (聯合)
Drawback: very likely to cause fragmentation within
allocated segments
Operating System Principles
9.12
Silberschatz, Galvin and Gagne ©2005
Buddy System Allocator
Operating System Principles
9.13
Silberschatz, Galvin and Gagne ©2005
Alternate strategy: Slab Allocation
Slab is one or more physically contiguous pages
Cache consists of one or more slabs
Single cache for each unique kernel data structure
Each cache filled with objects – instantiations of the data
structure
When cache created, filled with objects marked as free
When structures stored, objects marked as used
If slab is full of used objects, next object allocated from
empty slab
If no empty slabs, new slab allocated
Benefits include no fragmentation, fast memory request
satisfaction
Operating System Principles
9.14
Silberschatz, Galvin and Gagne ©2005
Slab Allocation
Operating System Principles
9.15
Silberschatz, Galvin and Gagne ©2005
9.9 Other Issues
Prepaging
To reduce the large number of page faults that occurs at process
startup
In a system with working-set model, we keep with each process a
list of pages in its working set. Before suspending a process, its
working set is saved.
Prepage all or some of the pages a process will need, before they
are referenced
But if prepaged pages are unused, I/O and memory was wasted
Assume s pages are prepaged and α of the pages is used
Is cost of the s * α saved pages faults greater or less than the cost of
prepaging s * (1- α) unnecessary pages?
α near zero prepaging loses
α near one prepaging wins
Operating System Principles
9.16
Silberschatz, Galvin and Gagne ©2005
Other Issues – Page Size
Page size selection must take the following into
consideration:
Size of the page table
Internal fragmentation
Need a larger page size
Locality
Need a small page size
Time to read or write a page
A large page size is preferred
Smaller page size to match program locality more accurately
Number of page faults
Need a larger page size to reduce number of page faults
Operating System Principles
9.17
Silberschatz, Galvin and Gagne ©2005
Other Issues – TLB Reach
TLB: expensive and power hungry
TLB Reach -The amount of memory accessible from TLB
TLB Reach = (TLB Size) X (Page Size)
Ideally, the working set of each process is stored in TLB
Otherwise there is a high degree of page faults
Increase the Page Size to increase TLB reach
This may lead to an increase in fragmentation as not all
applications require a large page size
Provide Multiple Page Sizes
This allows applications that require larger page sizes the
opportunity to use them without an increase in fragmentation
Recent trends is to move toward software-managed TLB
Operating System Principles
9.18
Silberschatz, Galvin and Gagne ©2005
Other Issues – Program Structure
Program structure
int[128,128] data;
Each row is stored in one page
Program 1
for (j = 0; j <128; j++)
for (i = 0; i < 128; i++)
data[i,j] = 0;
128 x 128 = 16,384 page faults
Program 2
for (i = 0; i < 128; i++)
for (j = 0; j < 128; j++)
data[i,j] = 0;
Operating System Principles
128 page faults
9.19
Silberschatz, Galvin and Gagne ©2005
Other Issues – Program Structure
Stack has good locality, hash table has bad locality
In addition to locality, other factors:
search speed, total number of memory references, total number of
pages touched
Compiler and loader
Code pages are always read-only
Loader can avoid placing routines across page boundaries
Routines that call each other can be packed into one page
Language
The use of pointers in C and C++ tend to randomize access to
memory, thereby potentially diminishing a process’s locality
OO programs tend to have a poor locality of reference
Operating System Principles
9.20
Silberschatz, Galvin and Gagne ©2005
Other Issues – I/O interlock
I/O Interlock – Pages must sometimes be locked into
memory.
Consider I/O - Pages that are used for copying a file
from a device must be locked from being selected for
eviction by a page replacement algorithm
Solutions
Never execute I/O to user memory. Use system memory
instead. Extra copying between user mamery and system
memory.
Allow pages to be locked into memory with a lock bit.
Lock-bit can be used in preventing replacement of a
newly brought-in page until it can be used at least once.
Useful for low-priority process,.
Operating System Principles
9.21
Silberschatz, Galvin and Gagne ©2005
Reason Why Frames Used For I/O Must Be In Memory
跳過 9.10
Operating System Principles
9.22
Silberschatz, Galvin and Gagne ©2005