Midterm Review

Download Report

Transcript Midterm Review

Midterm Review
CMSC 421
Fall 05
Chapter 1
Introduction
What is an Operating System?


A program that acts as an intermediary
between a user of a computer and the
computer hardware.
Operating system goals:
 Execute
user programs and make solving
user problems easier.
 Make the computer system convenient to
use.

Use the computer hardware in an
efficient manner.
Computer System Structure

Computer system can be divided into four
components

Hardware – provides basic computing resources


Operating system


Controls and coordinates use of hardware among various
applications and users
Application programs – define the ways in which the
system resources are used to solve the computing
problems of the users


CPU, memory, I/O devices
Word processors, compilers, web browsers, database
systems, video games
Users

People, machines, other computers
Operating System Definition

OS is a resource allocator
 Manages
all resources
 Decides between conflicting requests for
efficient and fair resource use

OS is a control program
 Controls
execution of programs to prevent
errors and improper use of the computer
Operating System Structure

Multiprogramming needed for efficiency






Single user cannot keep CPU and I/O devices busy at all times
Multiprogramming organizes jobs (code and data) so CPU
always has one to execute
A subset of total jobs in system is kept in memory
One job selected and run via job scheduling
When it has to wait (for I/O for example), OS switches to another
job
Timesharing (multitasking) is logical extension in
which CPU switches jobs so frequently that users can
interact with each job while it is running, creating
interactive computing

Virtual memory allows execution of processes not completely in
memory
Process Management

Process needs resources to accomplish its task

CPU, memory, I/O, files
 Initialization data


Process termination requires reclaim of any reusable resources
The operating system is responsible for the following activities in
connection with process management:
 Creating and deleting both user and system processes
 Suspending and resuming processes
 Providing mechanisms for process synchronization
 Providing mechanisms for process communication
 Providing mechanisms for deadlock handling
Memory Management



All data in memory before and after processing
All instructions in memory in order to execute
Memory management determines what is in
memory when
 Optimizing
CPU utilization and computer response to
users

Memory management activities
 Keeping
track of which parts of memory are currently
being used and by whom
 Deciding which processes (or parts thereof) and data
to move into and out of memory
 Allocating and deallocating memory space as needed
Storage Management

OS provides uniform, logical view of information storage


Abstracts physical properties to logical storage unit - file
Each medium is controlled by device (i.e., disk drive, tape drive)


Varying properties include access speed, capacity, data-transfer
rate, access method (sequential or random)
File-System management



Files usually organized into directories
Access control on most systems to determine who can access
what
OS activities include




Creating and deleting files and directories
Primitives to manipulate files and dirs
Mapping files onto secondary storage
Backup files onto stable (non-volatile) storage media
I/O Subsystem


One purpose of OS is to hide peculiarities of
hardware devices from the user
I/O subsystem responsible for
 Memory
management of I/O including buffering
(storing data temporarily while it is being transferred),
caching (storing parts of data in faster storage for
performance), spooling (the overlapping of output of
one job with input of other jobs)
 General device-driver interface
 Drivers for specific hardware devices
Chapter 2
Operating-System
Structures
System Calls





Programming interface to the services provided by the
OS
Typically written in a high-level language (C or C++)
Mostly accessed by programs via a high-level
Application Program Interface (API) rather than direct
system call use
Three most common APIs are Win32 API for Windows,
POSIX API for POSIX-based systems (including virtually
all versions of UNIX, Linux, and Mac OS X), and Java
API for the Java virtual machine (JVM)
Why use APIs rather than system calls?
System Call Implementation

Typically, a number associated with each system call



System-call interface maintains a table indexed according to these
numbers
The system call interface invokes intended system call in OS kernel
and returns status of the system call and any return values
The caller need know nothing about how the system call is
implemented


Just needs to obey API and understand what OS will do as a result call
Most details of OS interface hidden from programmer by API

Managed by run-time support library (set of functions built into libraries
included with compiler)
API–System Call–OS Relationship
System Programs

System programs provide a convenient environment for
program development and execution. The can be
divided into:








File manipulation
Status information
File modification
Programming language support
Program loading and execution
Communications
Application programs
Most users’ view of the operation system is defined by
system programs, not the actual system calls
Layered Approach
The operating system is divided into a number of layers (levels), each
built on top of lower layers. The bottom layer (layer 0), is the
hardware; the highest (layer N) is the user interface.
With modularity, layers are selected such that each uses functions
(operations) and services of only lower-level layers
Microkernel System Structure



Moves as much from the kernel into “user” space
Communication takes place between user modules
using message passing
Benefits:





Easier to extend a microkernel
Easier to port the operating system to new architectures
More reliable (less code is running in kernel mode)
More secure
Detriments:

Performance overhead of user space to kernel space
communication
Modules

Most modern operating systems
implement kernel modules
 Uses
object-oriented approach
 Each core component is separate
 Each talks to the others over known interfaces
 Each is loadable as needed within the kernel

Overall, similar to layers but with more
flexible
Virtual Machines



A virtual machine takes the layered approach to
its logical conclusion. It treats hardware and the
operating system kernel as though they were all
hardware
A virtual machine provides an interface identical
to the underlying bare hardware
The operating system creates the illusion of
multiple processes, each executing on its own
processor with its own (virtual) memory
Chapter 3
Processes
Process Concept

An operating system executes a variety of programs:





Batch system – jobs
Time-shared systems – user programs or tasks
Textbook uses the terms job and process almost
interchangeably
Process – a program in execution; process execution
must progress in sequential fashion
A process includes:



program counter
stack
data section
Process State

As a process executes, it changes state
Process Control Block (PCB)








Information associated with each process
Process state
Program counter
CPU registers
CPU scheduling information
Memory-management information
Accounting information
I/O status information
Schedulers






Long-term scheduler (or job scheduler) – selects which processes
should be brought into the ready queue
Short-term scheduler (or CPU scheduler) – selects which process
should be executed next and allocates CPU
Short-term scheduler is invoked very frequently (milliseconds) 
(must be fast)
Long-term scheduler is invoked very infrequently (seconds, minutes)
 (may be slow)
The long-term scheduler controls the degree of multiprogramming
Processes can be described as either:
I/O-bound process – spends more time doing I/O than computations,
many short CPU bursts
 CPU-bound process – spends more time doing computations; few very
long CPU bursts

Context Switch
When CPU switches to another process,
the system must save the state of the old
process and load the saved state for the
new process
 Context-switch time is overhead; the
system does no useful work while
switching

Interprocess Communication (IPC)



Mechanism for processes to communicate and to synchronize their
actions
Message system – processes communicate with each other without
resorting to shared variables
IPC facility provides two operations:
send(message) – message size fixed or variable
 receive(message)


If P and Q wish to communicate, they need to:



establish a communication link between them
exchange messages via send/receive
Implementation of communication link


physical (e.g., shared memory, hardware bus)
logical (e.g., logical properties)
IPC (cont.)
Direct Communication
 Indirect Communication
 Synchronization

 Blocking
(Synchronous)
 Non-Blocking (Asynchronous)

Buffering
Client-Server Communication

Sockets



Endpoint for communication
Communication consists between a pair of
sockets
Remote Procedure Call (RPC)




Abstracts procedure calls between processes on networked
systems.
Stubs – client-side proxy for the actual procedure on the server.
The client-side stub locates the server and marshalls the
parameters.
The server-side stub receives this message, unpacks the
marshalled parameters, and peforms the procedure on the
server.
Chapter 4: Threads
Threads
Benefits: Responsiveness, Resource
Sharing, Economy, Utilization of MP
Architectures
 User Threads: POSIX Pthreads, Win32
threads, Java threads
 Kernel Threads

Multithreading Models

Many-to-One

One-to-One

Many-to-Many
Threading Issues


Semantics of fork() and exec() system
calls
Thread cancellation
 Asynchronous
cancellation
 Deferred cancellation

Thread pools
 Create
a number of threads in a pool where
they await work


Thread specific data
Scheduler activations
 upcalls
Signal Handling


Signals are used in UNIX systems to notify a process
that a particular event has occurred
A signal handler is used to process signals




Signal is generated by particular event
Signal is delivered to a process
Signal is handled
Options:




Deliver the signal to the thread to which the signal applies
Deliver the signal to every thread in the process
Deliver the signal to certain threads in the process
Assign a specific thread to receive all signals for the process
Windows XP Threads


Implements the one-to-one mapping
Each thread contains

A thread id
 Register set
 Separate user and kernel stacks
 Private data storage area


The register set, stacks, and private storage area are known as the
context of the threads
The primary data structures of a thread include:



ETHREAD (executive thread block)
KTHREAD (kernel thread block)
TEB (thread environment block)
Linux Threads
Linux refers to them as tasks rather than
threads
 Thread creation is done through clone()
system call
 clone() allows a child task to share the
address space of the parent task (process)

Java Threads
Java threads are managed by the JVM
 Java threads may be created by:


Extending Thread class
 Implementing the Runnable interface
Chapter 5: CPU
Scheduling
Basic Concepts
Maximum CPU utilization obtained with
multiprogramming
 CPU–I/O Burst Cycle – Process execution
consists of a cycle of CPU execution and
I/O wait
 CPU burst distribution

CPU Scheduler


Selects from among the processes in memory that are
ready to execute, and allocates the CPU to one of them
CPU scheduling decisions may take place when a
process:






1.
2.
3.
4.
Switches from running to waiting state
Switches from running to ready state
Switches from waiting to ready
Terminates
Scheduling under 1 and 4 is nonpreemptive
All other scheduling is preemptive
Dispatcher

Dispatcher module gives control of the CPU to
the process selected by the short-term
scheduler; this involves:
 switching
 switching
context
to user mode
 jumping to the proper location in the user program to
restart that program

Dispatch latency – time it takes for the
dispatcher to stop one process and start another
running
Scheduling Criteria





CPU utilization – keep the CPU as busy as possible
Throughput – # of processes that complete their
execution per time unit
Turnaround time – amount of time to execute a particular
process
Waiting time – amount of time a process has been
waiting in the ready queue
Response time – amount of time it takes from when a
request was submitted until the first response is
produced, not output (for time-sharing environment)
Scheduling
First Come, First Serve
 Shortest-Job-First

 Pre-emptive
 Non
Pre-emptive
Priority
 Round Robin

Multilevel Queue


Ready queue is partitioned into separate queues:
foreground (interactive)
background (batch)
Each queue has its own scheduling algorithm



foreground – RR
background – FCFS
Scheduling must be done between the queues

Fixed priority scheduling; (i.e., serve all from foreground then from
background). Possibility of starvation.
 Time slice – each queue gets a certain amount of CPU time which it can
schedule amongst its processes; i.e., 80% to foreground in RR
 20% to background in FCFS
Multilevel Feedback Queue


A process can move between the various queues; aging
can be implemented this way
Multilevel-feedback-queue scheduler defined by the
following parameters:





number of queues
scheduling algorithms for each queue
method used to determine when to upgrade a process
method used to determine when to demote a process
method used to determine which queue a process will enter
when that process needs service
Multiple-Processor Scheduling




CPU scheduling more complex when multiple
CPUs are available
Homogeneous processors within a
multiprocessor
Load sharing
Asymmetric multiprocessing – only one
processor accesses the system data structures,
alleviating the need for data sharing
Real-Time Scheduling
Hard real-time systems – required to
complete a critical task within a
guaranteed amount of time
 Soft real-time computing – requires that
critical processes receive priority over less
fortunate ones

Thread Scheduling

Local Scheduling – How the threads
library decides which thread to put onto an
available LWP

Global Scheduling – How the kernel
decides which kernel thread to run next
Chapter 6
Process
Synchronization
Background






Concurrent access to shared data may result in data
inconsistency
Maintaining data consistency requires mechanisms to
ensure the orderly execution of cooperating processes
Producer - “writer”
Consumer - “reader”
Race Condition – outcome of execution depends on the
orrder in which access to data takes place
Critical Section – segment of code in which it is crucial
that no other process is allowed to execute
simultaneously
Critical-Section Problem Solution

Mutual Exclusion - If process Pi is executing in its
critical section, then no other processes can be
executing in their critical sections

Progress - If no process is executing in its critical
section and there exist some processes that wish to
enter their critical section, then the selection of the
processes that will enter the critical section next cannot
be postponed indefinitely

Bounded Waiting - A bound must exist on the number
of times that other processes are allowed to enter their
critical sections after a process has made a request to
enter its critical section and before that request is
granted
Synchronization Hardware


Many systems provide hardware support for critical
section code
Uniprocessors – could disable interrupts


Currently running code would execute without preemption
Generally too inefficient on multiprocessor systems


Operating systems using this not broadly scalable
Modern machines provide special atomic hardware
instructions



Atomic = non-interruptable
Either test memory word and set value
Or swap contents of two memory words
Solutions (cont.)

Test and Set : Shared boolean variable lock., initialized to false.
do { while ( TestAndSet (&lock )); // do nothing
// critical section
lock = FALSE;
//
remainder section
} while ( TRUE);

Swap : Shared Boolean variable lock initialized to FALSE; Each
process has a local Boolean variable key.
do { key = TRUE;
while ( key == TRUE)
Swap (&lock, &key );
//critical section
lock = FALSE;
//remainder section
} while ( TRUE);
Semaphore



Synchronization tool that does not require busy waiting
Semaphore S – integer variable
Two standard operations modify S: wait() and signal()



Originally called P() and V()
Less complicated
Can only be accessed via two indivisible (atomic)
operations
wait (S) {
while S <= 0
; // no-op
S--;
}
signal (S) {
S++;
}
Semaphore as General
Synchronization Tool


Counting semaphore – integer value can range over an unrestricted
domain
Binary semaphore – integer value can range only between 0
and 1; can be simpler to implement



Also known as mutex locks
Can implement a counting semaphore S as a binary semaphore
Provides mutual exclusion
Semaphore S; // initialized to 1
wait (S);
Critical Section
signal (S);
Semaphore Implementation


Must guarantee that no two processes can execute
wait () and signal () on the same semaphore at the same
time
Thus, implementation becomes the critical section
problem where the wait and signal code are placed in
the crtical section.

Could now have busy waiting in critical section implementation




But implementation code is short
Little busy waiting if critical section rarely occupied
Note that applications may spend lots of time in critical
sections and therefore this is not a good solution.
Solution : Semaphore with no busy waiting
Deadlock and Starvation


Deadlock – two or more processes are waiting indefinitely for an
event that can be caused by only one of the waiting processes
Let S and Q be two semaphores initialized to 1
P0
wait (S);
wait (Q);
.
.
.
signal (S);
signal (Q);

P1
wait (Q);
wait (S);
.
.
.
signal (Q);
signal (S);
Starvation – indefinite blocking. A process may never
be removed from the semaphore queue in which it is
suspended.
Classical Problems of
Synchronization
Bounded-Buffer Problem
 Readers and Writers Problem
 Dining-Philosophers Problem

Monitors


A high-level abstraction that provides a convenient and effective mechanism
for process synchronization
Only one process may be active within the monitor at a time
monitor monitor-name
{
// shared variable declarations
procedure P1 (…) { …. }
…
procedure Pn (…) {……}
Initialization code ( ….) { … }
…
}
}

Condition Variables
Chapter 7
Deadlocks
The Deadlock Problem

A set of blocked processes each holding a resource and
waiting to acquire a resource held by another process in
the set.

Four conditions of a deadlock:




Mutual Exclusion
Hold and Wait
No preemption
Circular Wait
Resource Allocation Graph

Two node types
 Process
 Resource

Two edge types
 Request
 Assignment
If graph contains no cycles  no deadlock.
 If graph contains a cycle 
 if only one instance per resource type, then deadlock.
 if several instances per resource type, possibility of deadlock.

Methods for Handling Deadlocks

Deadlock Prevention
Exclusion – must hold for nonshareable
resources
 Hold and Wait – cannot request resources if you are
holding some
 No Preemption
 Circular Wait – impose a total ordering of resource
types and require that each process requests
resources in an increasing order of enumeration
 Mutual

Deadlock Avoidance
 Make
sure any action will result in a safe state
Safe State

System is in safe state if there exists a safe sequence of
all processes.

Sequence <P1, P2, …, Pn> is safe if for each Pi, the
resources that Pi can still request can be satisfied by
currently available resources + resources held by all the
Pj, with j<I.





If Pi resource needs are not immediately available, then Pi can
wait until all Pj have finished.
When Pj is finished, Pi can obtain needed resources, execute,
return allocated resources, and terminate.
When Pi terminates, Pi+1 can obtain its needed resources, and
so on.
If a system is in safe state  no deadlocks.
If a system is in unsafe state  possibility of deadlock.
Methods for Handling Deadlocks
(cont.)

Deadlock Detection


Allow system to enter deadlock state
Detection algorithm

Maintain wait-for graph




Nodes are processes.
Pi  Pj if Pi is waiting for Pj.
Periodically invoke an algorithm that
searches for a cycle in the graph.
Recovery scheme


Process termination
Resource preemption
Chapter 8
Memory Management
Logical vs. Physical Address Space

The concept of a logical address space that is bound to a
separate physical address space is central to proper
memory management
Logical address – generated by the CPU; also referred to as
virtual address
 Physical address – address seen by the memory unit


Logical and physical addresses are the same in compiletime and load-time address-binding schemes; logical
(virtual) and physical addresses differ in execution-time
address-binding scheme

Memory-Management Unit (MMU) - Hardware device that
maps virtual to physical address
 In MMU scheme, the value in the relocation register is
added to every address generated by a user process at
the time it is sent to memory
Memory Management Concepts



Dynamic Loading - Routine is not loaded until it is
called
Dynamic Linking - Linking postponed until execution
time
Swapping - A process can be swapped temporarily out
of memory to a backing store, and then brought back
into memory for continued execution


Backing Store
Roll out, Roll in
Contiguous Allocation


Main memory divided into OS partition and user partition
Single-partition allocation


Relocation register and limit register define logical memory
range and protect users from overwriting each other’s memory
Multiple-partition


Hole – block of available space, scattered throughout memory
Must have some algorithm to determine where to place a new
process



First Fit
Best Fit
Worst Fit
Fragmentation



External Fragmentation – total memory space exists to
satisfy a request, but it is not contiguous
Internal Fragmentation – allocated memory may be
slightly larger than requested memory; this size
difference is memory internal to a partition, but not being
used
Reduce external fragmentation by compaction


Shuffle memory contents to place all free memory together in
one large block
Compaction is possible only if relocation is dynamic, and is done
at execution time
Paging







Logical address space of a process can be noncontiguous; process
is allocated physical memory whenever the latter is available
Divide physical memory into fixed-sized blocks called frames (size
is power of 2, between 512 bytes and 8192 bytes)
Divide logical memory into blocks of same size called pages.
Keep track of all free frames
To run a program of size n pages, need to find n free frames and
load program
Set up a page table to translate logical to physical addresses
Internal fragmentation
Address Translation Scheme

Address generated by CPU is
divided into:
Page number (p) – used as an
index into a page table which
contains base address of each
page in physical memory
 Page offset (d) – combined
with base address to define
the physical memory address
that is sent to the memory unit

Implementation of a page table





Page table is kept in main memory
Page-table base register (PTBR) points to
the page table
Page-table length register (PRLR) indicates
size of the page table
In this scheme every data/instruction access
requires two memory accesses. One for the
page table and one for the data/instruction.
The two memory access problem can be
solved by the use of a special fast-lookup
hardware cache called associative
memory or translation look-aside buffers
(TLBs)
More about page tables

Memory protection
 Valid/invalid

bit
Page table structure
 Hierarchical
 Hashed
Page Tables
 Inverted Page Tables

Shared Pages
Segmentation



Memory-management scheme that supports user view of memory
A program is a collection of segments. A segment is a logical unit
such as main program, procedure, function, etc.
Segment table – maps two-dimensional physical addresses; each
table entry has:
base – contains the starting physical address where the segments
reside in memory
 limit – specifies the length of the segment



Segment-table base register (STBR) points to the segment table’s
location in memory
Segment-table length register (STLR) indicates number of segments
used by a program;
segment number s is legal if s < STLR
Segmentation (cont.)

Relocation.



Sharing.





first fit/best fit
external fragmentation
Protection. With each entry in segment table associate:



shared segments
same segment number
Allocation.


dynamic
by segment table
validation bit = 0  illegal segment
read/write/execute privileges
Protection bits associated with segments; code sharing occurs at segment level
Since segments vary in length, memory allocation is a dynamic storage-allocation
problem