Overview

There are some simple notes of Operating System. Most contents are copied from slides.

Introduction

Isolation

  • 强制隔离以避免对整个系统的可用性、可靠性和安全影响
  • 运行的程序通常是隔离的单元
  • 防止程序 X 破坏或监视程序 Y
    • 读/写内存,使用 100% CPU,更改文件描述符
  • 防止进程干扰操作系统
  • 防止恶意程序、病毒、木马和 bug
    • 错误的过程可能会试图欺骗硬件或内核

Methods

  • 地址空间
    • 一个程序仅寻址其自己的内存
    • 每个程序若无许可,则无法访问不属于自己的内存
  • CPU 硬件中的特权模式/中断机制
    • 防止应用程序访问设备和敏感的 CPU 寄存器
    • 例如地址空间配置寄存器
    • 例如打断一直占用 CPU 的应用程序

Program Call

  • 好处
    • 执行很快
    • 灵活且易于传递和返回复杂数据类型
    • 程序员熟悉的机制
  • 坏处
    • 应用程序不可靠,可能有崩溃的风险

Interrupt

  • CPU 硬件支持中断、异常的处理
  • 中断是异步发生的,是来自处理器外部的 I/O 设备的信号结果
    • 硬件中断不是由任意一条专门的 CPU 指令造成,从这个意义上它是异步的
  • 硬件中断的异常处理程序通常称为中断处理程序
    • I/O 设备通过向处理器芯片的一个引脚发信号,并将异常信号放到系统总线上以触发中断
    • 在当前指令执行完后,处理器从系统总线读取异常号,保存现场,切换到 Kernel Mode
    • 调用中断处理程序,当中断处理程序完成后,它将控制返回到下一条本来要执行的指令
  • Timer 可以稳定定时地产生中断
    • 防止应用程序一直占用 CPU
    • 让 OS Kernel 能周期性地进行资源管理

Exception Handling

Basic principles and concepts

  • System Call: 应用程序主动向操作系统发出服务请求。
  • Exception: 由于非法指令或其他原因导致当前指令执行失败,(如:内存出错)后的处理请求。
  • Hardware Interrupt: 来自硬件设备的处理请求。

Summary

  • 源头:
    • 中断:外设
    • 异常:应用程序意想不到的行为
    • 系统调用:应用程序请求 OS 服务
  • 相应方式
    • 中断:异步
    • 异常:同步
    • 系统调用:异步或同步
  • 处理机制
    • 中断:持续且对应用程序透明
    • 异常:杀死或重新继续执行
    • 系统调用:等待和持续

Interrupt Handling

Establish Interrupt Mechanisms

  • 建立中断服务例程
    • 初始化
      • 设置时钟中断触发次数
      • 设置 sie 的 TI 使能 STIE 位
    • 服务例程
      • 调用 OpenSBI 提供的接口设置下次时钟中断触发时间
  • 让 CPU 能响应中断
    • 硬件设置
      • sstatus: 保存全局中断的使能位
      • sie: 指出 CPU 目前能处理或忽略的中断
      • stvec: 中断入口地址
    • 初始化
      • 设定 sie 的 TI 使能 STIE 位
      • 设定 sstatus 的使能中断 SIE 位
      • 实现中断服务总控函数
      • 设置 stvec 指向中断服务总控函数的入口地址
  • 相应并处理中断
    • 硬件设置
      • sepc: 保存中断的指令地址
      • pc: 设置为 stvec
      • scause: 设置中断的来源
      • sstatus: SIE 位置零以禁用中断
      • stval: 保存了中断相关的附加信息
    • 软件保存
      • x[0-32]: 通用寄存器
      • pc: 设置为 stvec
      • scause: 设置为中断的来源
      • sstatus: SIE 位置零以禁用中断
      • stval: 保存了中断相关的附加信息
    • 产生中断后
      • 硬件设置
      • 软件保存被打断现场
      • 执行软件实现的中断服务例程
      • 软件恢复被打断现场
      • 继续执行
    • 硬件中断服务例程可被打断
      • 不同硬件中断源可能在硬件中断处理时出现
      • 中断请求会保持到 CPU 做出响应
      • 硬件中断服务例程中需要临时禁止中断请求
  • 保存或恢复现场
    • 还需要为被中断的服务保存和恢复当时程序运行时的上下文
      • x[0-32]: 通用寄存器
      • sstatus: 系统系状态
      • sepc: 触发异常/中断的指令地址
      • scause: 指示发生异常/中断的种类
      • stval: 保存了发生异常/中断的附加信息
    • SAVE_ALL 寄存器

System Call

When an application wants to print a string, it will trigger the system call write().

  • 系统调用
    • 在 RISC-V 中,ecallsret 指令用于系统调用
    • 堆栈切换和特权级的转换
  • 函数调用
    • 在 RISC-V 中,callret 指令用于系统调用
    • 无堆栈切换和特权级的转换

Expenditure of System Call

开销会超过函数调用,主要来源:

  • 切换内核堆栈
  • 验证参数
  • 切换页表
  • 拷贝数据

Implementation of System Call

  • 应用发起请求
  • std lib 发出系统调用请求
    • 发出设置系统调用号和参数,发出 ecall
  • 硬件设置
    • sepc:保存请求后的指令地址
    • pc: 设置为 stvec
    • scause: 设置为 ecall from u-mode
    • sstauts: SIE 位置零以禁用中断
    • stval: 保存了相关的附件信息
  • 软件保存被打断现场
  • 执行软件实现的中断服务例程
  • 软件恢复现场
  • 应用继续执行

Memory Management

In a uniprogramming system, main memory is divided into two parts: one part for the operating system (resident monitor, kernel) and other part for the program currently being executed. In a multiprogramming system, the “user” part of memory must be further subdivided to accommodate multiple processes. The task of subdivision is carried out dynamically by the operating system and is known as memory management.

Effective memory management is vital in a multiprogramming system. If only a few processes are in memory, then for much of the time all of the processes will be waiting for I/O (input/output), and the processor will be idle. Thus, memory needs to be allocated to ensure a reasonable supply of ready processes to consume available processor time.

In this chapter, some key concepts about contiguous memory allocation and non-contiguous memory allocation will be discussed.

Contiguous Memory Allocation

The Memory Hierarchy

A typical hierarchy is illustrated in figure below.

The Memory Hierarchy

As one goes down the hierarchy, the following occur:

  • Decreasing cost per bit
  • Increasing capacity
  • Increasing access time
  • Decreasing frequency of access to the memory by the processor

Cache Memory

Although cache memory is invisible to the OS, it interacts with other memory management hardware.

Cache Principles

Cache memory is intended to provide memory access time approaching that of the fastest memories available, and at the same time support a large memory size that has the price of less expensive types of semiconductor memories. The concept is illustrated in the Figure below

Single Cache

There is a relatively large and slow main memory together with a smaller, faster cache memory.

The next figure depicts the use of multiple levels of cache.

Three-level Cache Organization

Main memory consists of up to addressable words, with each word having a unique -bit address. For mapping purposes, this memory is considered to consist of a number of fixed length blocks of K words each. That is there are blocks. Cache consists of C slots (lines) of K words each, and the number of slots is considerably less than the number of main memory blocks ().

Cache/Main Memory Structure

Memory Management Methods

  • Relocation
  • Segmentation
  • Paging
  • Virtual Memory

Address Space

The range of addresses available to a computer program.

  • A logical address is a reference to a memory location independent of the current assignment of data to memory; a translation must be made to a physical address before the memory access can be achieved.

  • A relative address is a particular example of logical address, in which the address is expressed as a location relative to some known point, usually a value in a processor register.

  • A physical address, or absolute address is an actual location in main memory.

Generating Logical Address

  • Compile
  • Assemble
  • Link
  • Relocation

Procedures

  • CPU
    • ALU: 需要逻辑地址的内存内容
    • MMU:进行逻辑地址和物理地址的转换
    • CPU 控制逻辑:给总线发送物理地址请求
  • 内存: 发送物理地址的内容给 CPU 或接受 CPU 数据到物理地址
  • 操作系统:建立逻辑地址与物理地址的映射

Address Check

Address Check

Memory Partitioning

The principal operation of memory management is to bring processes into main memory for execution by the processor.

Fixed Partitioning

In most schemes for memory management, we can assume the OS occupies some fixed portion of main memory, and the rest of main memory is available for use by multiple processes. The simplest scheme for managing this available memory is to partition it into regions with fixed boundaries.

Description

Main memory is divided into a number of static partitions at system generation time. A process may be loaded into a partition of equal or greater size.

Strengths

Simple to implement; little operating system overhead.

Weaknesses

Inefficient use of memory due to internal fragmentation; maximum number of active processes is fixed.

Example of Fixed Partitioning

With unequal-size partitions, there are two possible ways to assign processes to partitions. The simplest way is to assign each process to the smallest partition within which it will fit. In this case, a scheduling queue is needed for each partition to hold swapped-out processes destined for that partition.

Internal Fragmentation

Occurs when memory is divided into fixed-size partitions (e.g., page frames in main memory, physical blocks on disk). If a block of data is assigned to one or more partitions, then there may be wasted space in the last partition. This will occur if the last portion of data is smaller than the last partition.

External Fragmentation

Occurs when memory is divided into variable-size partitions corresponding to the blocks of data assigned to the memory (e.g., segments in main memory). As segments are moved into and out of the memory, gaps will occur between the occupied portions of memory.

Dynamic Partitioning

Description

Partitions are created dynamically, so each process is loaded into a partition of exactly the same size as that process. Data structure is used in dynamic partitioning to maintain allocated blocks and empty blocks.

Strengths

No internal fragmentation; more efficient use of main memory.

Weaknesses

Inefficient use of processor due to the need for compaction to counter external fragmentation.

Placement Algorithms

Three placement algorithms that might be considered are best-fit, first-fit, and next-fit. All, of course, are limited to choosing among free blocks of main memory that are equal to or larger than the process to be brought in. Best-fit chooses the block that is closest in size to the request. First-fit begins to scan memory from the beginning and chooses the first available block that is large enough. Next-fit begins to scan memory from the location of the last placement and chooses the next available block that is large enough.

  • Best-fit: 分配 n 字节的分区时,查找并使用不小于 n 的最小空闲分区

    • 原理及实现
      • 空闲分区列表按大小排序
      • 分配时查找一个合适的分区
      • 释放时查找并合并邻近的空闲分区
    • 优点
      • 大部分分配块的尺寸较小时效果较好
      • 可避免大的空闲分区被拆分
      • 可减小外部碎片的大小
      • 实现简单
    • 缺点
      • 容易产生外部碎片
      • 释放分区速度较慢
  • Worst-fit: 分配 n 字节,使用尺寸不小于 n 的最大空闲分区

    • 原理及实现

      • 空闲分区列表按由大到小排序
      • 分配时,选最大的分区
      • 释放时,尽可能与邻近分区合并,并调整空闲分区列表顺序
    • 优点

      • 中等大小的分配请求较多时,效果最好
      • 不会出现太多的小碎片
    • 缺点

      • 释放分区速度较慢
      • 容易产生外部碎片
      • 容易破坏大的空闲分区
  • First-fit: 分配 n 个字节,使用第一个可用的空间比 n 大的空闲块。

    • 原理及实现
      • 空闲分区列表按地址顺序排序
      • 分配时,选第一个合适的分区
      • 释放时,尽可能与邻近分区合并
    • 优点
      • 实现简单
      • 在高地址空间有大块的空闲分区
    • 缺点
      • 分配大块空间速度较慢
      • 容易产生外部碎片

Example

Allocation of 16-Mbyte Block: Example of Memory Configuration Worst-fit: 36M

Defragmentation

通过调整进程占用的分区位置来减少或避免分区碎片。

Compaction

通过移动分配给进程的内存分区,以合并外部碎片。
碎片紧凑的条件:所有的应用程序可动态重定位。

Swapping in/out

通过抢占并回收处于等待状态进程的分区,以增大可用内存空间

Buddy System

Both fixed and dynamic partitioning schemes have drawbacks. A fixed partitioning scheme limits the number of active processes and may use space inefficiently if there is a poor match between available partition sizes and process sizes. A dynamic partitioning scheme is more complex to maintain and includes the overhead of compaction. An interesting compromise is the buddy system. In a buddy system, memory blocks are available of size words, , where

  • small size block that allocated

  • largest size block that is allocated; generally is the size of the entire memory available for allocation

To begin, the entire space available for allocation is treated as a single block of size . If a request of size such that is made, then the entire block is allocated. Otherwise, the block is split into two equal buddies of size . This process continues until the smallest block greater than or equal to is generated and allocated to the request. At any time, the buddy system maintains a list of holes (unallocated blocks) of each size . A hole may be removed from the list by splitting it in half to create two buddies of size in the list. Whenever a pair of buddies on the list both become unallocated,they are removed from that list and coalesced into a single block on the list. Presented with a request for an allocation of size such that .
The Figure below gives an example using a 1-Mbyte initial block.

Example of Buddy System

Implementation

  • 数据结构
    • 空闲块按大小和起始地址组织成二维数组
    • 初始状态:只有一个大小为 的空闲块
  • 分配过程
    • 由小到大在空闲块数组中找最小的可用空闲块
    • 如空闲块过大,对可用空闲块进行二等分,直到得到合适的可用空闲块
  • 释放过程
    • 把释放的块放入空闲块数组
    • 合并满足合并条件的空闲块
  • 合并条件
    • 大小相同为
    • 地址相邻
    • 低地址空闲块起始地址为 的倍数

Non-contiguous Memory Allocation

  • 设计目标:提高内存利用效率和管理灵活性
  • 优点
    • 允许一个程序的使用非连续的物理地址空间
    • 允许共享代码与数据
    • 支持动态加载和动态链接

Segmentation

Segmented Address Space

进程的段地址空间由多个段组成:主代码段、子模块代码段、公用库代码段、堆栈段(stack)、堆数据(heap)、初始化数据段和符号表等。
段式存储管理的目的在于更细粒度和灵活的分离与共享

Access Mechanisms

  • 段的概念

    • 段表示访问方式和储存数据等属性相同的一段地址空间。
    • 段地址在逻辑空间上连续,在物理空间上不连续
    • 若干个段组成逻辑地址空间
  • 段访问: 逻辑地址由段基址和段内偏移组成的二元组构成

Paging

  • Frame: 把物理地址空间划分为大小相同的基本分配单位(2 的 次方)
  • Page: 把逻辑地址空间也划分为相同大小的基本分配单位
  • Frame 和 Page 的大小必须是相同的

Frame

物理内存被划分为相同大小的帧,用二元组 表示,其中 代表帧号(共有 个帧),每个帧有 个字节。因此物理地址计算公式为

Page

进程逻辑地址空间被划分为大小相等的页,通常来讲,页内偏移等于段内偏移但页号不等于帧号。逻辑地址由二元组 表示,虚拟地址计算公式为

Address Mapping

  • 逻辑地址中的页号是连续的
  • 物理地址中的帧号不是连续的
  • 不是所有的页都有相应的帧对应

Page Tables

Page tables save the information of the mapping from logical addresses to physical addresses.

每一个进程都有一个页表,每一个页面对应一个页表项,通过 Page Table Base Register (PTBR) 实现页面到页表项之间的映射。

Page Table Entries

Page Table Entry Information

  • Frame Number: It gives the frame number in which the current page you are looking for is present. The number of bits required depends on the number of frames.Frame bit is also known as address translation bit.

  • Present/Absent Bit: Present or absent bit says whether a particular page you are looking for is present or absent. In case if it is not present, that is called Page Faults. It is set to 0 if the corresponding page is not in memory. Used to control page faults by the operating system to support virtual memory. Sometimes this bit is also known as valid/invalid bits.

  • Protection Bit: Protection bit says that what kind of protection you want on that page. So, these bit for the protection of the page frame (read, write etc).

  • Referenced bit: Referenced bit will say whether this page has been referred in the last clock cycle or not. It is set to 1 by hardware when the page is accessed.

  • Modified Bit: Modified bit says whether the page has been modified or not. Modified means sometimes you might try to write something on to the page. If a page is modified, then whenever you should replace that page with some other page, then the modified information should be kept on the hard disk or it has to be written back or it has to be saved back. It is set to 1 by hardware on write-access to page. Sometimes this modified bit is also called as the Dirty bit.

  • Caching Enable/Disabled: a bit enables or disable caching of the page.

Performance Problem

  1. 访问一个内存单元需要先访问页表,再访问数据。
  2. 页表可能会非常大

How to solve: Caching or Indirection

Translation Lookaside Buffer

Translation lookaside buffer (TLB) functions in the same way as a memory cache and contains those page table entries that have been most recently used. The organization of the resulting paging hardware is illustrated in Figure below. Given a virtual address, the processor will first examine the TLB. If the desired page table entry is present (TLB hit), then the frame number is retrieved and the real address is formed. If the desired page table entry is not found (TLB miss), then the processor uses the page number to index the process page table and examine the corresponding page table entry. If the “present bit” is set, then the page is in main memory, and the processor can retrieve the frame number from the page table entry to form the real address. The processor also updates the TLB to include this new page table entry. Finally, if the present bit is not set, then the desired page is not in main memory and a memory access fault, called a page fault, is issued.

Use of a Translation Lookaside Buffer

Comparison between Paging and Segmentation

Logical Addresses

Hierarchical Page Table

Take a two-level hierarchical page table as an example:

A Two-Level Hierarchical Page Table

Address Translation in a Two-Level Paging System

Virtual Memory

We now move to a discussion of virtual memory. An analysis of this topic is complicated by the fact that memory management is a complex interrelationship between processor hardware and operating system software. We will focus first on the hardware aspect of virtual memory, looking at the use of paging, segmentation, and combined paging and segmentation. Then we will look at the issues involved in the design of a virtual memory facility in operating systems.

Preliminary Concepts

Overlay

  • 解释:应用程序手动把需要的指令和数据保存在内存中

  • 目标:在较小的可用内存中运行较大的程序

  • 特点:

    • 只能发生在没有调用关系的模块间
    • 程序员必须给出模块间的逻辑覆盖结构
    • 发生在运行程序的内部模块间
  • 方法:根据程序逻辑结构,将程序划分为若干功能相对独立的模块,将不会同时执行的模块共享同一块内存区域

    • 必要部分(常用功能)的代码和数据常驻内存
    • 可选部分(不常用功能)放在其他程序模块中,只在需要用到的时候装入内存
    • 不存在调用关系模块可共享一块内存区域
  • 示例
    An Overlay Example

  • 不足

    • 增加编程困难
      • 需要划分功能模块并确定模块间的覆盖关系
      • 增加了编程的复杂度
    • 增加执行时间
      • 从外存装入覆盖模块
      • 时间换空间

Swapping

  • 目标:增加正在运行或需要运行的程序的内存
  • 特点:
    • 以进程为单位
    • 不需要模块间的逻辑覆盖结构
    • 发生在内存进程间
  • 实现方法:
    • 可将暂时不能运行的程序放到外存
    • 换入换出的基本单位:整个进程的地址空间
    • 换出 (swap out):把一个进程的整个地址空间保存到外存
    • 换入 (swap in):将外存中某进程的地址空间读入到内存
  • 问题与解答
    • 何时需要交换:当内存空间不够或又不够的可能时换出
    • 交换区大小:存放所有用户进程的所有内存映像的拷贝
    • 换出后再换入时要放在原处吗(重定位问题):采用动态地址映射的方法

The Principle of Locality

Intuitively, the principle of locality makes sense. Consider the following line of reasoning:

  1. Except for branch and call instructions, which constitute only a small fraction of all program instructions, program execution is sequential. Hence, in most cases, the next instruction to be fetched immediately follows the last instruction fetched.

  2. It is rare to have a long uninterrupted sequence of procedure calls followed by the corresponding sequence of returns. Rather, a program remains confined to a rather narrow window of procedure-invocation depth. Thus, over a short period of time, references to instructions tend to be localized to a few procedures.

  3. Most iterative constructs consist of a relatively small number of instructions repeated many times. For the duration of the iteration, computation is therefore confined to a small contiguous portion of a program.

  4. In many programs, much of the computation involves processing data structures, such as arrays or sequences of records. In many cases, successive references to these data structures will be to closely located data items.

To summarize, the principle of locality states that program and data references within a process tend to cluster. Hence, the assumption that only a few pieces of a process will be needed over a short period of time is valid. Also, it should be possible to make intelligent guesses about which pieces of a process will be needed in the near future, which avoids thrashing.

Example

for (int j = 0; j < 1024; j++)
  for (int i = 0; i < 1024; i++) A[i][j] = 0;

for (int i = 0; i < 1024; i++)
  for (int j = 0; j < 1024; j++) A[i][j] = 0;

In the first code, page faults occur, and in the second code, page faults occur.

Virtual Storage

  • 思路

    • 把不常用的部分内存块暂存到外
  • 原理

    • 装载程序时
      • 只将当前指令执行需要的部分页面或段载入内存
    • 指令执行中需要的指令或数据不在内存(即缺页或缺段)时
      • 处理器通知操作系统将相应的页或段调入内存
    • 操作系统将内存中暂时不用的页或段保存到外存
  • 实现方式

    • 虚拟页式存储
    • 虚拟段式存储
  • 基本特征

    • 不连续性
      • 物理内存分配非连续
      • 虚拟地址空间使用非连续
    • 大用户空间
      • 提供给用户的虚拟内存可大于实际的物理内存
    • 部分交换
      • 虚拟存储只对部分虚拟地址空间进行调入和调出

Virtual Pages

页式存储管理的增加请求掉和和页面替换

  • 当用户程序装载到内存运行时,只装入部分页面就启动程序运行
  • 进程在运行中发现有需要的代码或数据不在内存中就向系统发出缺页异常请求
  • 操作系统在处理缺页异常时,将外存中的相应页面载入内存使得进程能继续执行 Address Translation in Virtual Pages

Page Table Entries

Page Faults

Procedures

  1. 在内存中有空闲物理帧时,分配一物理帧 f,转第 5 步
  2. 根据页面替换算法选择将被替换的物理帧对应的逻辑页 q
  3. 若 q 被修改过则将它写回内存
  4. 修改 q 的 present bit 为 0
  5. 将需要访问的页 p 载入物理帧 f
  6. 修改 p 的 present bit 为 1,帧号为 f
  7. 重新执行产生缺页的指令

Effective Memory-access Time

  • Hit Ratio: The percentage of times that a particular page number is found in the TLB.

  • TLB Hit: The page is found in TLB

  • TLB Miss: The page is not found in the TLB so that accessing memory to get the page table entries and the data is needed.

Example

  • TLB lookup takes 5 nano sec.
  • Memory access time is 100 nano sec.
  • Hit ratio is 80%.

What is the effective memory access time?

Local Replacement

Introduction

  • 置换算法实现的功能:当出现缺页异常,需调入新页面而内存已满时,置换算法选择被置换的物理页面
  • 设计目标
    • 尽可能减少页面调入调出的次数
    • 把短期内不再访问的页面调出
  • 页面锁定 (frame locking)
    • 描述必须常驻内存的逻辑页面
    • 操作系统的关键部分
    • 要求响应速度的代码和数据
    • 页表中的锁定标志位 (lock bit)
  • 评价方法
    • 模拟页面置换行为,记录产生缺页的次数
    • 更少的缺页次数,更好的性能

Algorithms

Optimal (OPT)

  • Selects for replacement that page for which the time to the next reference is the longest (based on future information), and results in the fewest number of page faults.
  • Impossible to implement
  • Benchmark algorithm (a standard to judge real-world algorithms)

Least recently used (LRU)

  • Replaces the page in memory that has not been referenced for the long time (based on past information). By the principle of locality, this should be the page least likely to be referenced in the near future.
  • Does nearly as well as the optimal policy.
  • Implement by linked list or stack, ordered by visit time, with high time complexity.
  • LRU can be degenerated into FIFO if a block have not been referenced since it was fetched into memory.

First-in-first-out (FIFO)

  • Replacing the page that has been in memory the longest: A page fetched into memory a long time ago may have now fallen out of use
  • Process as a circular buffer
  • May cause belady anomaly
  • Easy to implement but performs poorly (Belady Anomaly)

Clock

  • Organize frames in a circular buffer with which a pointer is associated.
  • First loaded or subsequently referenced, use bit = 1.
  • When replace, find a frame with use bit = 0, and reset use bit of frames that each scan encounters to 0.
  • If all of the frames have a use bit of 1, then the pointer will make one complete cycle through the buffer, setting all the use bits to 0, and stop at its original position, replacing the page in that frame.
  • An implementation of LRU using FIFO mechanism.

Improved Clock

The clock algorithm can be made more powerful by increasing the number of bits that it employs like

  1. Not accessed recently, not modified (u = 0; m = 0)
  2. Accessed recently, not modified (u = 1; m = 0)
  3. Not accessed recently, modified (u = 0; m = 1)
  4. Accessed recently, modified (u = 1; m = 1)
  • Beginning at the current position of the pointer, scan the frame buffer. During this scan, make no changes to the use bit. The first frame encountered with (u = 0; m = 0) is selected for replacement.
  • If step 1 fails, scan again, looking for the frame with (u = 0; m = 1). The first such frame encountered is selected for replacement. During this scan, set the use bit to 0 on each frame that is bypassed.
  • If step 2 fails, the pointer should have returned to its original position and all of the frames in the set will have a use bit of 0. Repeat step 1 and, if necessary, step 2. This time, a frame will be found for the replacement.

Least Frequently Used

  • Replace the block in the set that has experienced the fewest references.
  • could be implemented by associating a counter with each block.
  • When a block is brought in, it is assigned a count of 1; with each reference to the block, its count is incremented by 1. When replacement is required, the block with the smallest count is selected.
  • Make poor replacement choices when certain blocks are referenced relatively infrequently overall, but frequently in some short intervals of repeated references due to locality, thus building up high reference counts.

FBR

  • To overcome this difficulty with LFU, a technique known as frequency-based replacement is proposed.
    • A simple version in the Figure (a): The blocks are organized in a stack and the top part of the stack is designated for the new section, which means the referenced block is moved to the top of the new section if there is a cache hit. Besides, if the block is not in the new section (i.e. the old section), its reference count increments by 1.
    • A further refinement addresses this problem, seeing Figure (b): Divide the stack into three sections (new, middle and old). As before, reference counts are not incremented on blocks in the new section. However, only blocks in the old section are eligible for replacement. Supposing a sufficiently large middle section, this allows relatively frequently referenced blocks a chance to build uo their counts before becoming eligible for replacement (buffer section). Simulation studies indicates that this refined policy is significantly better than LRU or LFU. Note that sizes of new, middle and old sections are hyperparameters which should be fine tuned in daily use.

Frequency-Based Replacement

Example

OPT/LRU/FIFO/CLOCK

Algorithm Compare

Improved Clock

Time012345678910
Requestcawdbwebawbcd
Frame 0 (u,m)a(00)a(00)a(11)a(11)a(11)a(01)a(01)a(11)a(11)a(11)a(01)
Frame 1 (u,m)b(00)b(00)b(00)b(00)b(11)b(01)b(11)b(11)b(11)b(11)b(01)
Frame 2 (u,m)c(00)c(10)c(10)c(10)c(10)e(10)e(10)e(10)e(10)e(10)d(10)
Frame 3 (u,m)d(00)d(00)d(00)d(10)d(10)d(00)d(00)d(00)d(00)c(10)c(00)
Page Fault

LFU

Notice that ax means a has been referenced for x times.

Time012345678910
Requestc7a1d14b5e18b1a19b20c20d17
Frame 0a8a8a9a9a9e18e18e18e18e18d17
Frame 1b5b5b5b5b10b10b11a19a19a19a19
Frame 2c6c13c13c13c13c13c13c13b20b20b20
Frame 3d2d2d2d16d16d16d16d16d16c20c20
Page Fault

Belady Anomaly

Belady anomaly is the phenomenon in which increasing the number of page frames results in an increase in the number of page faults for certain memory access patterns. This phenomenon is commonly experienced when using the first-in first-out (FIFO) page replacement algorithm.

In FIFO, the page fault may or may not increase as the page frames increase, but in Optimal and stack-based algorithms like LRU, as the page frames increase the page fault decreases.

Global Replacement

A global replacement policy considers all unlocked pages in main memory as candidates for replacement, regardless of which process owns a particular page.

  • 全局置换算法要解决的问题

    • 进程在不同阶段的内存需求是变化的
    • 分配给进程的内存也需要在不同阶段有所变化
    • 全局置换算法需要确定分配给进程的物理页面数
  • CPU 利用率与并发进程数的关系

    • 进程数少时,提高并发进程数可提高 CPU 利用率
    • 并发进程导致内存访问增加
    • 并发进程的内存访问降低了访存的局部性特征
    • 局部性特征的下降会导致缺页率上升和 CPU 利用率下降

Working Set Strategy

Working Set

The working set with parameter for a process at virtual time , which we designate as , is the set of pages of that process that have been referenced in the last virtual time units.The variable is a window of virtual time over which the process is observed. (当前时刻 前的 时间窗口中的所有访问页面所组成的集合)

The working set size will be a non-decreasing function of the window size. The result is illustrated in Figure below, which shows a sequence of page references for a process.

Working Set Defined by Window Size

Note that the larger the window size, the larger is the working set, which can be expressed in the following relationship: The working set is also a function of time. If a process executes over time units and uses only a single page, . A working set can also grow as large as the number of pages N of the process, if many different pages are rapidly addressed and if the window size allows. Thus,

Typical Graph of Working Set Size

Figure above indicates the way in which the working set size can vary over time for a fixed value of . For many programs, periods of relatively stable working set sizes alternate with periods of rapid change.

  • When a process first begins executing, it gradually builds up to a working set as it references new pages.
  • Eventually by the principle of locality, the process should stabilize on a certain set of pages.
  • Subsequent transient periods reflect a shift of the program to a new locality.
  • During the transition phase, some of the pages from the old locality remain with the window, causing a surge in the size of the working as new pages are referenced.
  • As the window slides past these page references, the working set size declines until it contains only those pages from the new locality.

Resident Set

That portion of a process that is actually in main memory at a given time. (在当前时刻,进程实际驻留在内存当中的页面集合)

  • Resident Set 和 Working Set 的关系
    • Working Set 是进程在运行过程中固有性质
    • Resident Set 取决于系统分配给进程物理页面的数目和页面置换算法
  • Resident Set 与 Page Fault Rate 的关系
    • ,缺页较少
    • Working Set 发生剧烈抖动时,缺页较多
    • 进程 Resident Set 大小达到一定数目后,PFF 不会显著下降

Algorithm

Basic Idea

  1. Monitor the working set of each process.
  2. Periodically remove from the resident set of a process those pages that are not in its working set. This is essentially an LRU policy.
  3. A process may execute only if its working set is in main memory (i.e., if its resident set includes its working set).

Implementation

  • 窗口大小:当前时刻前 个访问页面
  • 访存链表
    • 访存时,换出不在工作集的页面,更新访存链表
    • 缺页时,换入页面,更新访存链表

Example

Working Set Algorithm Example

Page Fault Frequency Strategy

Page Fault Rate (PFR)

The reciprocal of time between page faults is the page fault rate.

  • 影响 PFR 的因素
    • 页面置换算法
    • 分配给进程的物理页面数目
    • 页面大小
    • 程序写法

Algorithm

Basic Idea

  • It requires a use bit to be associated with each page in memory. The bit is set to 1 when that page is accessed. When a page fault occurs, the OS notes the virtual time since the last page fault for that process; this could be done by maintaining a counter of page references.
  • A threshold F is defined. If the amount of time since the last page fault is less than F, then a page is added to the resident set of the process.
  • Otherwise, discard all pages with a use bit of 0, and shrink the resident set accordingly. At the same time, reset the use bit on the remaining pages of the process to 0.

Example

The threshold F=2

PBR Example

Thrashing

A phenomenon in virtual memory schemes, in which the processor spends most of its time swapping pieces rather than executing instructions.

  • 产生原因:随着驻留内存进程数的增加,由于页面数有限,导致缺页现象频繁发生。
  • 操作系统需要在并发水平和缺页率间达到一个平衡,即选择一个适当进程数目和进程需要的页面数

FBR

Local vs. Global Replacement

Replacement algorithms can be local or global.

  • When a process incurs a page fault, a local page replacement algorithm selects for replacement some page that belongs to that same process (or a group of processes sharing a memory partition).
  • A global replacement algorithm is free to select any page in memory.
  • With a local replacement strategy, the number of frames allocated to a process does not change.
  • With global replacement, a process may happen to select only frames allocated to other processes, thus increasing the number of frames allocated to it (assuming that other processes do not choose its frames for replacement).
  • Global replacement generally results in greater system throughput and is therefore the more common method.

Processes and Threads

This chapter examines some concepts related to process management, which are found in a number of contemporary operating systems. We show that the concept of process is more complex and subtle than presented so far and in fact embodies two separate and potentially independent concepts: one relating to resource ownership, and another relating to execution.

Processes

Definition

进程是指一个具有一定独立功能的程序在一个数据集合上的一次动态执行过程。

Composition

  • 代码
  • 数据
  • 状态寄存器:CR0, IP
  • 通用寄存器:AX, BX, CX
  • 进程占用系统资源

Features

  • 动态性:可动态地创建、结束进程
  • 并发性:进程可以被独立调度并占用处理机运行
  • 独立性:不同进程的工作不相互影响
  • 制约性:因访问共享数据/资源或进程间同步而产生制约

Relations and Discrepancies between programs

Relations

  • Processes are the abstract of Programs in executive states
    • Program = Static executive file
    • Process = Program + executive states
  • Multiple executions of the some program are regarded as different processes.
  • Necessary resources in executing processes: Memory and CPU

Discrepancies

  • 进程是动态的,程序是静态的
    • 程序是有序代码的集合
    • 进程是程序的执行
  • 进程是暂时的,程序时永久的
    • 进程是一个状态变化的过程
    • 程序可长久保存
  • 进程与程序的组成不同:进程的组成包括程序、数据和进程控制块

Process Control Block

操作系统用 PCB 来进程的基本状态,每个进程都在操作系统中有一个唯一对应的PCB

Structure

At any given point in time, while the program is executing, this process can be uniquely characterized by a number of elements, including the following:

  • Identifier: A unique identifier associated with this process, to distinguish it from all other processes.
  • State: If the process is currently executing, it is in the running state.
  • Priority: Priority level relative to other processes.
  • Program counter: The address of the next instruction in the program to be executed.
  • Memory pointers: Include pointers to the program code and data associated with this process, plus any memory blocks shared with other processes.
  • Context data: These are data that are present in registers in the processor while the process is executing.
  • I/O status information: Includes outstanding I/O requests, I/O devices assigned to this process, a list of files in use by the process, and so on.
  • Accounting information: May include the amount of processor time and clock time used, time limits, account numbers, and so on.

Simplified Process Control Block

Control Information

  • 调度和状态信息: 调度进程和处理及使用情况

  • 进程间通信信息: 进程间通信相关的各种标识

  • 储存管理信息

  • 进程所用资源

  • 有关数据结构的连接信息

Implementation

  • 链表:同一状态进程的 PCB 组成链表
  • 索引表:统一状态进程的 PCB 被同一索引指向

Process States

Life-cycle

  • 进程创建
    • 系统初始化时
    • 用户请求创建一个新进程
    • 被其他进程调用
  • 进程执行
    • 内核选择一个就绪进程,让其占用处理机并执行
  • 进程等待
    • 请求并等待系统服务,无法马上完成
    • 启动某种无法马上完成的操作
    • 需要的资源没有达到
  • 进程抢占
    • 被高优先级进程抢占
    • 在规定时间内未完成
  • 进程唤醒
    • 被阻塞进程需要的资源可被满足
    • 被阻塞进程等待的时间到达
    • 被阻塞的进程只能被别的进程或操作系统唤醒
  • 进程结束
    • 自愿结束
      • 正常退出
      • 错误退出
    • 非自愿结束
      • 致命错误n
      • 被其他进程所杀

操作系统在 中断 中处理进程
Sleep() 为例,创建 -> 就绪 -> 运行 -> 等待 -> 操作系统时钟中断 ...

Process Creation

Processes Switching

Processes Switching

Processes Termination

A Two-State Process Model

Two-state Process Model

A Five-State Model

Split the Not Running state of the two-state process model into two states: Ready and Blocked and add two additional states for good measure, we can get a five-state model as follows:

  1. Running: The process that is currently being executed.
  2. Ready: A process that is prepared to execute when given the opportunity.
  3. Blocked/Waiting: A process that cannot execute until some event occurs, such as the completion of an I/O operation.
  4. New: A process that has just been created but has not yet been admitted to the pool of executable processes by the OS. Typically, a new process has not yet been loaded into main memory, although its process control block has been created.
  5. Exit: A process that has been released from the pool of executable processes by the OS, either because it halted or because it aborted for some reason.

Five-state Process Model

Figure above indicates the types of events that lead to each state transition for a process; the possible transitions are as follows:

  • Null -> New: A new process is created to execute a program.
  • New -> Ready: The OS will move a process from the New state to the Ready state when it is prepared to take on an additional process.
  • Ready -> Running: When it is time to select a process to run, the OS chooses one of the processes in the Ready state.
  • Running -> Exit: The currently running process is terminated by the OS if the process indicates that it has completed or if it aborts.
  • Running -> Ready
    • The running process has reached the maximum allowable time for uninterrupted execution;
    • The OS preempted a process if there is another process with higher priority level being blocked.
  • Running -> Blocked: A process is put in the Blocked state if it requests something for which it must wait. For example, a process may request a service from OS that OS is not prepared to perform immediately or a process may initiate that must be completed before the process can continue.
  • Blocked -> Ready: A process in the Blocked state is moved to the Ready state when the event for which it has been waiting occurs.
  • Ready -> Exit: For clarity, this transition is not shown on the state diagram. In some systems, a parent may terminate a child process at any time. Also, if a parent terminates, all child processes associated with that parent may be terminated.
  • Blocked -> Exit: The comments under the preceding item apply.

Suspended Processes

There are two independent concepts here: whether a process is waiting on an event (blocked or not), and whether a process has been swapped out of main memory (suspended or not). To accommodate this combination, we need four states:

  1. Ready: The process is in main memory and available for execution.
  2. Blocked: The process is in main memory and awaiting an event.
  3. Blocked/Suspend: The process is in secondary memory and awaiting an event.
  4. Ready/Suspend: The process is in secondary memory but is available for execution as soon as it is loaded into main memory.

Two Suspended States

Important new transitions are the following:

  • Blocked -> Blocked/Suspend:
  • Blocked/Suspend -> Ready/Suspend:
  • Ready/Suspend -> Ready:
  • Ready -> Ready/Suspend:
  • Blocked/Suspend -> Blocked:

Queueing Model

Queueing Model

Process Switching

On the face of it, the function of process switching would seem to be straightforward. At some time, a running process is interrupted, and the OS assigns another process to the Running state and turns control over to that process. However, several design issues are raised.

  • First, what events trigger a process switch?
  • Another issue is that we must recognize the distinction between mode switching and process switching.
  • Finally, what must the OS do to the various data structures under its control to achieve a process switch?

When to Switch Processes:

  • Interrupt
    • Clock interrupt
    • I/O interrupt
    • Memory fault
  • Trap
  • Supervisor call

进程切换的要求:

  • 切换前,保存进程上下文
  • 切换后,恢复进程上下文
  • 快速切换

进程生命周期的信息:

  • 寄存器
  • CPU 状态
  • 内存地址空间

Example

以进程 P0、进程 P1 为例:
P0 执行 --> 中断或系统调用 --> P0 保存 --> P1 恢复 --> P1 执行 --> 中断或系统调用 --> P1 保存 --> P0 恢复 --> P0 执行
图示: Context Switch 在进程切换的过程中,内核为每一个进程维护了一个对应的进程控制块(PCB),内核将相同状态进程的 PCB 放在同一队列。

Instructions in Linux

  • 创建进程:fork or exec
  • 父进程等待子进程:wait
  • 进程的有序终止:exit
  • 优先级控制:nice
  • 进程调试支持:ptrace
  • 定时:sleep
    Instruction and State

Threads

Introduction

多进程的问题:

  1. 如何通信、如何共享数据
  2. 系统开销较大,涉及进程的创建、进程的终止等

解决思路即在进程内部增加一类实体,满足以下特性:

  1. 实体之间可以并发执行
  2. 实体之间共享相同的地址空间

Definition

线程是进程的一部分,描述指令流执行状态,它是进程中指令执行流的最小单位,是 CPU 调度的基本单位。 线程间的堆栈 (Heap) 资源不能共享
线程是资源调度的基本单位,即线程描述在进程资源环境中指令流的执行状态。因此同一个进程内的线程共享其资源。

Multithreading

Multithreading refers to the ability of an OS to support multiple, concurrent paths of execution within a single process.

multithreading

The left half of Figure depicts single-threaded approaches.

  • MS-DOS is an example of an OS that supports a single-user process and a single thread.
  • Other operating systems, such as some variants of UNIX, support multiple user processes, but only support one thread per process.

The right half part of Figure states two multithreaded approaches.

  • A Java runtime environment is an example of a system of one process with multiple threads.
  • Windows, Solaris, and many modern versions of UNIX use multiple processes, each of which supports multiple threads.

In a multithreaded environment, a process is defined as the unit of resource allocation and a unit of protection. The following are associated with processes:

  • A virtual address space that holds the process image.
  • Protected access to processors, other processors, files, and I/O resources.

With a process, there may be one or more threads, each with following:

  • A thread execution state (Running, Ready, etc.)
  • A saved thread context when not running; one way to review a thread is as an independent program counter operating within a process
  • An execution stack
  • Some per-thread static storage for local variables
  • Access to the memory and resources of its process, shared with all other threads in that process.

The Figure below illustrates the distinction between threads and processes from the point of view of process management.

Single Threaded and Multithreaded Process Models

The key benefits of threads derive from the performance implications:

  1. It takes far less time to create a new thread in an existing process, than to create a brand-new process.

  2. It takes less time to terminate a thread than a process.

  3. It takes less time to switch between two thread s within the same process than to switch between processes.

  4. Threads enhance efficiency in communication between different executing programs.

  • 多线程的优点
    • 一个进程中可以同时存在多个线程
    • 多线程之间可以并发
    • 线程间共享地址空间和文件等资源
  • 多线程的缺点
    • 一个线程崩溃,会导致同属一个进程下的其他线程崩溃

Types of Threads

There are two broad categories of thread implementation: user-level threads (ULTs) and kernel-level threads (KLTs). The latter are also referred to in the literature as kernel-supported threads or lightweight processes.

User-level Threads

由一组用户级的线程库函数来完成线程的管理,包括线程的创建、终止、同步和调度等。

User-level Threads

  • 优点

    • 不依赖于操作系统
    • 用户空间实现线程机制
      • 每个进程有私有的线程控制块(TCB)列表
      • TCB 列表又线程库函数维护
    • 同一进程内的用户线程切换速度快
    • 允许每个进程拥有自己的线程调度算法
  • 不足

    • 线程发起系统调用而阻塞时,整个线程进入等待。
      • 除非当前运行线程主动放弃,否则其所在进程的其他线程无法抢占 CPU
    • 不支持基于线程的处理机抢占,除非当前运行线程主动放弃
    • 只能按进程分配 CPU 时间,即在多线程的多进程中,每个线程分配的时间片较少

Kernel-level Threads

内核通过系统调用实现的线程机制,由内核实现线程的创建、终止和管理。

Kernel-level Threads

优点:

  • 由内核维护 PCB 和 TCB
  • 线程执行系统调用而被阻塞不影响其他线程
  • 线程的创建、终止和切换相对较大:因为通过系统调用或内核函数,在内核中实现
  • 基于线程数量的时间分配:线程数量多的进程 CPU 执行时间多

Lightweight Process

内核支持的用户线程,一个进程可有一个或多个轻量级进程,每个 Lightweight 进程由一个单独的内核进程(即内核线程池)来支持,线程由内核线程池(独立内核进程)执行,但线程与内核线程池之间未绑定。 Lightweight Process

Relationship between Threads and Processes

Threads: ProcessesDescriptionExample Systems
1:1Each thread of execution is a unique process with its own address space and resources.Traditional UNIX implementations
M:1A process defines an address space and dynamic resource ownership. Multiple threads may be created and executed within that process.Windows NT, Solaris, Linux, OS/2, OS/390, MACH
1:MA thread may migrate from one process environment to another. This allows a thread to be easily moved among distinct systems.Ra (Clouds), Emerald
M:NIt combines attributes of M:1 and 1:M cases.TRIX

Scheduling

Uniprocessor Schduling

In a multiprogramming system, multiple processes exist concurrently in main memory. Each process alternates between using a processor and waiting for some event to occur, such as the completion of an I/O operation. The processor or processors are kept busy by executing one process while the others processes wait.

Types of Processor Scheduling

  • Long-term scheduling: The decision to add to the pool of processes to be executed.
  • Medium-term scheduling: The decision to add to the number of processes that are partially or fully in main memory.
  • I/O scheduling: The decision as to which process's pending I/O request shall be handled by an available I/O device.

Time Division Multiplexing

  • 进程切换: CPU 资源占用者的切换
    • 保存当前进程在 PCB 的上下文
    • 恢复下一个进程的执行上下文
  • 处理机调度:从就绪队列中挑选下个占用 CPU 资源的 进程,从多个 CPU 中挑选就绪进程可用的 CPU 资源
  • 调度程序:挑选就绪进程的内核函数
  • 调度形式
    • 非抢占系统:当前进程主动放弃 CPU
    • 可抢占系统:中断请求被服务例程响应完成时或当前进程被抢占
  • 调度策略:如何从就绪队列中选择下一个执行进程

Principles of Scheduling Algorithms

  • CPU 使用率:处于忙状态的时间百分比
  • 吞吐量 (Throughput):单位时间完成的进程数量
  • 周转时间 (Turnaround Time):进程从初始化到结束的总时间
  • 等待时间 (Waiting Time):进程在就绪队列的总时间
  • 响应时间 (Response Time):从提交请求到产生相应所花费的总时间

调度算法需要得到更快的服务,也需要得到高吞吐量,而带宽和延迟不可兼得

  • 处理机调度策略的相应时间目标

    • 减少响应时间
    • 减少平均响应时间的波动(操作系统的计算延迟)
  • 处理机调度策略的吞吐量目标

    • 增加吞吐量:减少开销、系统资源的高效开销
    • 减少等待时间,即减少每个进程的等待时间
    • 操作系统需要保证吞吐量不受用户交互的影响
  • 处理机调度的公平性目标

    • 公平不等于公正(例如:富人和穷人缴同等额度的税)

    • 公平性即保证每个进程占用相同的 CPU 的时间、保证每个进程的等待时间相同

    • 公平通常会增加平均响应时间,降低效率

Scheduling Criteria

  • User Oriented, Performance Related
    • Turnaround time(周转时间):waiting time plus service time
    • Response time(响应时间):from the submission of a request until the response begins to be received
    • Deadlines
  • User Oriented, Other
    • Predictability
  • System Oriented, Performance Related
    • Throughput
    • Processor utilization
  • System Oriented, Other
    • Fairness
    • Enforcing priorities
    • Balancing resources

Scheduling Algorithms

Table below presents some summary information about the various scheduling policies that are examined in this subsection. The selection function determines which process, among ready processes, is selected next for execution. The function may be based on priority, resource requirements, or the execution characteristics of the process. In the latter case, three quantities are significant:

time spent in system so far, waiting time spent in execution so far total service time required by the process

The decision mode specifies the instants in time at which the selection function is exercised. There are two general categories:

  • Nonpreemptive: Once a process is in the Running state, it continues to execute until it terminates or it blocks itself to wait for I/O or to request some OS service.

  • Preemptive: The current running process may be interrupted and move to the Ready state by the OS, which happens when a new process arrives, when an interrupt occurs that places a blocked process in the Ready state, or periodically, based on a clock interrupt.

FCFSSPNHRRNRR
Selection Functionconstant
Decision ModeNon-preemptiveNon-preemptiveNon-preemptivePreemptive
ThroughputNot emphasizedHighHighMay be low if quantum is too small
Response TimeMay be high Provides good response time for short processesProvides good response time for short processesProvides good response timeProvides good response time for short processes
OverheadPenalizes short processes; penalizes I/O-bound processesPenalizes long processesGood balanceFair treatment
StarvationNoPossibleNoNo

First Come First Served, FCFS

先来先服务算法:依据进程进入就绪状态的先后顺序排列(超市结账排队)

  • 优点:简单
  • 缺点:
    • 平均等待时间波动较大:短进程可能排在长进程后面;
    • I/O 和 CPU 资源利用较低:CPU 密集进程不使用 I/O,I/O 密集进程不使用 CPU
  • 示例:3 个进程,计算时间分别为 12, 3, 3
    • 任务到达顺序:P1, P2, P3,周转时间 (Turnaround Time)为
    • 任务到达顺序:P2, P3, P1,周转时间为

Shortest Process Next, SPN

短进程优先算法:选择就绪队列中执行时间最短的进程占用 CPU 进入运行状态

  • 就绪队列按预期的执行时间来排序

  • 具有最优平均周转时间

  • 连续的短进程流会使长进程无法获得 CPU 资源,对长进程不公平

  • 需要预估下一个 CPU 计算的持续时间

    • 通过问用户的方式。用户欺骗就杀死进程
    • 基于历史的记录学习(类似于动量梯度下降,即
  • How to Estimate Processing Time

    • The simplest calculation of estimating the processing time of each process would be where is processor execution time for the ith instance of process, is predicted value for the ith instance

    • To avoid recalculating the entire summation each time, we can get Typically,we would like to give greater weight to more recent instances, because these are more likely to reflect future behavior.

    • A common technique for predicting a future value on the basis of a time series of past values is exponential averaging:

    where is a constant weighting factor () that determines the relative weight given to more recent observations relative to older observations. To see this more clearly, consider an expansion

Highest Response Ratio Next, HRRN

最高响应比优先算法:选择就绪队列中相应比 R 值最高的进程

where response ratio
time spent waiting for the processor
expected service time

特点:在短进程优先算法的基础上进行改进;不可抢占;关注进程的等待时间;防止无限期推迟

Round Robin, RR

时间片轮转算法:A straightforward way to reduce the penalty that short jobs suffer with FCFS is to use preemption based on a clock.

  • 时间片 (slices of time): 分配处理及资源的基本时间单位
  • 算法思路:时间片结束后,按 FCFS (First come, first service) 算法切换到下一个就绪进程;每隔 个时间片进程执行一个时间片 q。(When the interrupt occurs, the currently running process is placed in the ready queue, and the next ready job is selected on a FCFS basis.)
  • 示例:
    • P1: 53, P2: 8, P3: 68, P4: 24 (时间片为20)
    • P1 0-20,P2 20-28,P3 28-48,P4 48-68(P2 已结束)
    • P1 68-88,P3 88-108,P4 108-112 (P4 已结束)
    • P1 112-125,P3 125-145 (P1 已结束)
    • P3 145-153 (P3已结束)
    • 等待时间
    • 平均等待时间:
  • 时间片太大:等待时间太长,极端情况会退化成 FCFS
  • 时间片太小:产生大量上下文切换且开销太大
  • 经验规则:维持上下文切换开销处于 1% 以内

Multiple Queues, MQ

  • 多级队列调度算法
    • 就绪队列被划分为几个对立的子队列,是上述几种算法的综合
    • 每个队列都有自己的调度策略
    • 队列之间的调度:
      • 固定优先级:先处理前台,再处理后台
      • 时间片轮转:每个队列都能得到一个确定且能够调度其进程的 CPU 总时间

Multilevel Feedback Queues, MLFQ

  • 多级反馈队列算法
    • 进程可在不同队列间移动的多级队列算法
    • CPU 密集型进程优先级下降得很快
    • 对 I/O 密集型进程有利

Fair-Share Scheduling

公平共享调度算法:公平共享调度控制用户对系统资源的访问

  • 一些用户组比其他用户组更重要
  • 保证不重要的组无法垄断资源
  • 未使用的资源按比例分配
  • 没有达到资源使用率目标的组获得更高的优先级

A Comparison of Scheduling Policies

A Comparison of Scheduling Polices

Real-time Scheduling

  • 定义:正确性依赖于其 时间功能 两方面的操作系统

  • 实时操作系统的性能指标

    • 时间约束的及时性
    • 速度和平均性能相对不重要
  • 分类

    • 强实时操作系统
      • 要求在指定时间内必须完成重要的任务
    • 弱实时操作系统
      • 重要进程有高优先级
  • Hard Deadline: 错过任务实现会导致灾难性后果;必须验证,在最坏的情况下能够满足实现

  • Soft deadline: 通常能满足任务时限,尽力保证满足任务时限;若不能满足则降低要求

  • 可调度性: 表示一个实时操作系统能够满足任务时限要求

  • 实时调度算法

    • 速率单调调度算法 (Rate Monotonic Scheduling): 通过 周期 安排优先级,周期越短优先级越高,执行周期最短的任务
      • 示例:Rate Monotonic Scheduling
    • 最早截止时间优先算法 (Earliest Deadline First, EDF): 截止时间越早优先级越高,执行截止时间最早的任务

Multicore Scheduling

  • 多处理器调度
    • 多个处理机组成一个多处理器系统
    • 处理机间可负载共享
    • 对称多处理机的进程分配:
      • 静态进程分配
      • 动态进程分配
    • 优先级反置 (Priority Inversion): 操作系统中出现高优先级进程长时间等待低优先级进程所占用资源的现象
      • 解决方法: 优先级继承 (Priority Inheritance)、优先级天花板协议 (Priority Ceiling Protocol)

Algorithms

SQMS (Single Queue Multiprocessor Schedule)

  • Lack of scalability.

  • Weak in cache affinity, violates the locality principle.

  • Let process run in the same CPU, load balance will be invalid.

  • Thus either assure the cache affinity or load balance.

MQMS (Multi-Queue Multiprocessor Scheduling)

  • Queues assign to different CPUs, and the scheduling algorithms can be different.

  • 所有进程都可以保持在一个 CPU 上。

  • 容易造成负载不均。

例如:某一个 CPU 任务执行完毕后,该 CPU 处于空转状态。因此只有通过进程的跨 CPU 迁移才能实现负载均衡。

  • 具有可扩展性,队列数可根据 CPU 数进行扩展。

如何进行进程工作的迁移?

  • Work stealing: 进程、页表基址等需要迁移。

SMP and Linux Kernel

Linux Kernel 2.4 的 调度算法:initial queueexpired queue, 单队列机制。

Linux Kernel 2.6 的 调度算法:由 Ingo Molnar 设计,由于 Linux 系统有 140 中优先级,因此用一个 140 维的数组存放。在 active bitarray 中寻找 left-most bit 的位置,并在 APA 中寻找 APA[x], 从 APA[x] 中 dequeue 一个 process,对于当前执行完的 process,重新计算 priority 再 enqueue 到 expired priority array 相应的队列 EPA[x] 中。在一定时间间隔后,要进行 load balance 分析。

操作系统内核和驱动程序属于动态链接库。

Completely Fair Scheduler, CFS

通过计算进程消耗的 CPU 时间而不是优先级来进行调度。占用的资源越少,优先级就越高,但这种做法不一定公平,因为有些进程相比于其他进程更加重要即使执行的时间很长,因此会对每个进程进行赋权,即维护相对公平性。

分配给进程的运行时间 = 调度周期 × 进程权重/总权重之和。调度周期即指把所有 TASK_RUNNING 态进程都调度一遍的时间。 Linux 引入 vruntime 进行计算:实际运行时间 × 1024/进程权重,vruntime 越小表示以前占用 CPU 时间较短,受到了不公平待遇。 例如:系统中只有两个权重为 1 和 2 的进程,假设跳读周期为 30ms

  • A 的 CPU 时间为
  • B 的 CPU 时间为

则在这 30 ms 中 A 运行 10 ms,B 运行 20 ms。

问题:新进程的 vruntime 为 0,则在相当长时间内都会保持抢占 CPU 的优势,因此应该设定一个最小值从而与老进程保持在一个合理的差距范围内。

  • 休眠进程在唤醒后的 vruntime 相比于其他 active 进程较小,因此会强占 CPU,因此在进程重新唤醒后应该对 vruntime 进行一些补偿。在 Linux 中 sched_features 的 WAKEUP_PREEMPT 位决定在休眠后是否主动强占 CPU。

  • 进程从一个 CPU 迁移到另一个 CPU 上时 vruntime 会不会变? 当一个进程从 CPU_x 出去并进入 CPU_y 的运行队列中时,它的 vruntime = vruntime - min_vruntime_x + min_vruntime_y。

  • 同时 vrumtime 可能会溢出(因为类型是 unsigned long,因此在比较 vruntime 的时候应该先减去 min_vruntime。

    • 示例
    unsigned char a = 251, b = 254;
    b += 5;
    signed char c = a - 250, d = b - 250;
    // b 比 a 大 8
    
    

Brain Fuck Scheduler, BFS

一种时间片轮转算法的变种,在多处理机情况的但就绪队列选择,增加了队列互斥访问的开销。

Concurrency

This chapter begins with an introduction to the concept of concurrency and the implications of the execution of multiple concurrent processes. We find that the basic requirement for support of concurrent processes is the ability to enforce mutual exclusion; that is, the ability to exclude all other processes from a course of action while one process is granted that ability.

Next, we will examines two problems that plague all efforts to support concurrent processing: deadlock and starvation. We begin with a discussion of the underlying principles of deadlock and the related problem of starvation. Then we will examine the three common approaches to dealing with deadlock: prevention, detection, and avoidance. We will then look at one of the classic problems used to illustrate both synchronization and deadlock issues: the dining philosophers problem.

Mutual Exclusion and Synchronization

Key Terms

  • Atomic operation: A function or action implemented as a sequence of one or more instructions that appears to be indivisible; that is, no other process can see an intermediate state or interrupt the operation. The sequence of instruction is guaranteed to execute as a group, or not execute at all, having no visible effect on system state. Atomicity guarantees isolation from concurrent processes.
  • Critical section: A section of code within a process that requires access to shared resources, and that must not be executed while another process is in a corresponding section of code.
  • Deadlock: A situation in which two or more processes are unable to proceed because each is waiting for one of the others to do something.
  • Livelock: A situation in which two or more processes continuously change their states in response to changes in the other process(es) without doing any useful work.
  • Mutual exclusion: The requirement that when one process is in a critical section that accesses shared resources, no other process may be in a critical section that accesses any of those shared resources.
  • Race condition: A situation in which multiple threads or processes read and write a shared data item, and the final result depends on the relative timing of their execution.
  • Starvation: A situation in which a runnable process is overlooked indefinitely by the scheduler; although it is able to proceed, it is never chosen.

Critical Section

entry section
    critical section
exit section
    remainder section
  • Critical Section: 进程中访问临界资源的一段需要互斥执行的代码

  • Entry Section: 检查可否进入临界区的一段代码,如可进入,设置相应“正在访问临界区”标志

  • Exit Section: 清除“正在访问临界区”标志

  • Remainder Section: 代码中的其余部分

访问规则:空闲则入,忙则等待,有限等待

Mutual Exclusion Software Support

Software approaches can be implemented for concurrent processes that execute ona single-processor or a multiprocessor machine with shared main memory. These approaches usually assume elementary mutual exclusion at the memory access level. That is, simultaneous accesses (reading and/or writing) to the same location in main memory are serialized by some sort of memory arbiter, although the order of access granting is not specified ahead of time.

Dekker’s Algorithm

As mentioned earlier, any attempt at mutual exclusion must rely on some fundamental exclusion mechanism in the hardware. The most common of these is the constraint that only one access to a memory location can be made at a time. Using this constraint, we reserve a global memory location labeled turn. A process (P0 or P1) wishing to execute its critical section first examines the contents of turn. If the value of turn is equal to the number of the process, then the process may proceed to its critical section. Otherwise, it is forced to wait. Our waiting process repeatedly reads the value of turn until it is allowed to enter its critical section. This procedure is known as busy waiting, or spin waiting, because the thwarted process can do nothing productive until it gets permission to enter its critical section. Instead, it must linger and periodically check the variable; thus it consumes processor time (busy) while waiting for its chance.
After a process has gained access to its critical section, and after it has completed that section, it must update the value of turn to that of the other process.
In formal terms, there is a shared global variable: int turn = 0;

First Attempt

// PROCESS 0 
while(turn != 0)
    /* do nothing */ ;
/* critical section*/;
turn = 1;

// PROCESS 1
while(turn != 1)
    /* do nothing */ ;
/* critical section*/;
turn = 0;

Second Attempt

// PROCESS 0 
while(flag[1] != 0)
    /* do nothing */ ;
flag[0] = 1;
/* critical section*/;
flag[0] = 0;

// PROCESS 1
while(flag[0] != 0)
    /* do nothing */ ;
flag[1] = 1;
/* critical section*/;
flag[1] = 0;

Third Attempt

// PROCESS 0 
flag[0] = 1;
while(flag[1] != 0)
    /* do nothing */ ;
/* critical section*/;
flag[0] = 0;

// PROCESS 1
flag[1] = 1;
while(flag[0] != 0)
    /* do nothing */ ;
/* critical section*/;
flag[1] = 0;

Fourth Attempt

// PROCESS 0 
flag[0] = 1;
while(flag[1] != 0) {
    flag[0] = 0;
    /* delay */ ;
    flag[0] = 1;
}
/* critical section*/;
flag[0] = 0;

// PROCESS 1
flag[1] = 1;
while(flag[0] != 0) {
    flag[1] = 0;
    /* delay */ ;
    flag[1] = 1;
}
/* critical section*/;
flag[1] = 0;

A Correct Solution

bool flag [2];
int turn;
void P0() {
    while (true) {
        flag[0] = true;
        while (flag[1]) {
            if (turn == 1) {
                flag[0] = false;
                while (turn == 1) 
                    /* do nothing */;
                flag[0] = true;
            }
        }
        /* critical section */;
        turn = 1;
        flag[0] = false;
        /* remainder */;
    }
}
void P1( )
{
    while (true) {
        flag[1] = true;
        while (flag[0]) {
            if (turn == 0) {
                flag[1] = false;
                while (turn == 0) 
                    /* do nothing */;
                flag[1] = true;
            }
        }
        /* critical section */;
        turn = 0;
        flag[1] = false;
        /* remainder */;
    }
}
void main () {
    flag[0] = false;
    flag[1] = false;
    turn = 1;
    parbegin (P0, P1);
}

Peterson’s Algorithm

Dekker’s algorithm solves the mutual exclusion problem, but with a rather complex program that is difficult to follow and whose correctness is tricky to prove. Peterson has provided a simple, elegant solution.

bool flag[2];
int turn;
void P0() {
    while (true) {
        flag[0] = true;
        turn = 1;
        while (flag[1] && turn == 1) /* do nothing */;
        /* critical section */;
        flag[0] = false;
        /* remainder */;
    }
}
void P1() {
    while (true) {
        flag[1] = true;
        turn = 0;
        while (flag[0] && turn == 0) /* do nothing */;
        /* critical section */;
        flag[1] = false;
        /* remainder */
    }
}
void main() {
    flag[0] = false;
    flag[1] = false;
    parbegin (P0, P1);
}

Mutual Exclusion Hardware Support

Atomic Operation

原子操作指不存在任何中断或失败的操作

  • 要么操作成功完成
  • 要么操作没有执行
  • 不会出现部分执行的情况

操作系统需要利用同步机制在并发执行的同时,保证一些操作是原子操作。

  • 用两个原子指令实现一个锁(lock)
    • Lock.acquire()
      • 在锁被释放前一直等待,然后获得锁
      • 如果两个线程都在等待同一个锁,并且同时发现锁被释放了,那么只有一个能够获得锁
    • Lock.release()
      • 解锁并唤醒任何等待中的进程

Interrupt Disabling

禁用硬件中断:没有中断,没有上下文切换,因此没有并发

local_irq_save(unsigned long flags); 
critical section
local_irq_restore(unsigned long flags); 

Because the critical section cannot be interrupted, mutual exclusion is guaranteed. The price of this approach, however, is high. The efficiency of execution could be noticeably degraded because the processor is limited in its ability to interleave processes. Another problem is that this approach will not work in a multiprocessor architecture. When the computer includes more than one processor, it is possible (and typical) for more than one process to be executing at a time. In this case, disabled interrupts do not guarantee mutual exclusion.

  • 进入临界区:禁止所有中断,并保存标志
  • 离开临界区:使能所有中断,并恢复标志
  • 缺点:
    • 禁用中断后,进程无法被停止,可能导致其他进程处于饥饿状态

    • 临界区可能很长,可能存在对硬件的影响

Special Machine Instructions

At the hardware level, as was mentioned, access to a memory location excludes any other access to that same location. With this as a foundation, processor designers have proposed several machine instructions that carry out two actions atomically, such as reading and writing or reading and testing, of a single memory location with one instruction fetch cycle. During execution of the instruction, access to the memory location is blocked for any other instruction referencing that location.

Compare & Swap Instruction

int compare_and_swap (int *word, int testval, int newval) {
    int oldval;
    oldval = *word
    if (oldval == testval) *word = newval;
    return oldval;
}

This atomic instruction therefore has two parts: A compare is made between a memory value and a test value; if the values are the same, a swap occurs. The entire compare&swap function is carried out atomically—that is, it is not subject to interruption. Example:

/* program mutualexclusion */
const int n = /* number of processes */;
int bolt;
void P(int i)
{
    while (true) {
    while (compare_and_swap(bolt, 0, 1) == 1)
        /* do nothing */;
    /* critical section */;
    bolt = 0;
    /* remainder */;
    }
}
void main() {
    bolt = 0;
    parbegin (P(1), P(2), ... ,P(n));
}

A shared variable bolt is initialized to 0. The only process that may enter its critical section is one that finds bolt equal to 0. All other processes attempting to enter their critical section go into a busy waiting mode. When a process leaves its critical section, it resets bolt to 0; at this point one and only one of the waiting processes is granted access to its critical section.

Exchange Instruction

void exchange (int *register, int *memory) {
    int temp;
    temp = *memory;
    *memory = *register;
    *register = temp;
}

Example:

/* program mutualexclusion */
int const n = /* number of processes */;
int bolt;
void P(int i)
{
    while (true) {
        int keyi = 1;
        do exchange (&keyi, &bolt)
        while (keyi != 0);
        /* critical section */;
        bolt = 0;
        /* remainder */;
    }
}
void main() {
    bolt = 0;
    parbegin (P(1), P(2), ..., P(n));
}

A shared variable bolt is initialized to 0. Each process uses a local variable key that is initialized to 1. The only process that may enter its critical section is one that finds bolt equal to 0. It excludes all other processes from the critical section by setting bolt to 1.
When a process leaves its critical section, it resets bolt to 0, allowing another process to gain access to its critical section.
Note the following expression always holds because of the way in which the variables are initialized and because of the nature of the exchange algorithm: If bolt = 0, then no process is in its critical section. If bolt = 1, then exactly one process is in its critical section, namely the process whose key value equals 0.

Test & Set Instruction

bool test_and_set (bool *target) ‏{
    boolean rv = *target;
    *target = true;
    return rv:
}

Example:

/* program mutualexclusion */
const int n = /* number of processes */;
int bolt;
void P(int i)
{
    while (true) {
    while (test_and_set(bolt) == 1)
        /* do nothing */;
    /* critical section */;
    bolt = 0;
    /* remainder */;
    }
}
void main() {
    bolt = 0;
    parbegin (P(1), P(2), ... ,P(n));
}

忙等待锁 (Busy Waiting Lock):

Lock::Acquire() {
   while (test_and_set(bolt));
     //spin
}
Lock::Release() {
    bolt = 0;
}

无忙等待锁:

class Lock {
   int bolt = 0;
   WaitQueue q;
}

Lock::Acquire() {
   while (test_and_set(bolt)) {
      // add this TCB to wait queue q;
      schedule();
   }
}

Lock::Release() {
   bolt = 0;
   remove one thread t from q;
   wakeup(t);
}

Advantages

  • It is applicable to any number of processes on either a single processor or multiple processors sharing main memory.
  • It is simple and therefore easy to verify.
  • It can be used to support multiple critical sections; each critical section can be defined by its own variable.

Disadvantages

  • Busy waiting is employed: Thus, while a process is waiting for access to a critical section, it continues to consume processor time.
  • Starvation is possible: When a process leaves a critical section and more than one process is waiting, the selection of a waiting process is arbitrary. Thus, some process could indefinitely be denied access.
  • Deadlock is possible: Consider the following scenario on a single-processor system. Process P1 executes the special instruction (e.g., compare&swap, exchange) and enters its critical section. P1 is then interrupted to give the processor to P2, which has higher priority. If P2 now attempts to use the same resource as P1, it will be denied access because of the mutual exclusion mechanism. Thus, it will go into a busy waiting loop. However, P1 will never be dispatched because it is of lower priority than another ready process, P2.

Summary

  • 锁是一种高级的同步抽象方法
    • 互斥可以使用锁来实现
    • 需要硬件支持
  • 常用的三种同步实现方法
    • 禁用中断(仅限于单处理器)
    • 软件方法(复杂)
    • 原子操作指令(单处理器或多处理器均可)

Semaphore

Key Terms

  • Semaphore: An integer value used for signaling among processes. Only three operations may be performed on a semaphore, all of which are atomic: initialize, decrement, and increment. The decrement operation may result in the blocking of a process, and the increment operation may result in the unblocking of a process. Also known as a counting semaphore or a general semaphore.

  • Binary semaphore: A semaphore that takes on only the values 0 and 1.

  • Mutex: Similar to a binary semaphore. A key difference between the two is that the process that locks the mutex (sets the value to 0) must be the one to unlock it (sets the value to 1).

  • Condition variable: A data type that is used to block a process or thread until a particular condition is true.

  • Monitor: A programming language construct that encapsulates variables, access procedures, and initialization code within an abstract data type. The monitor’s variable may only be accessed via its access procedures and only one process may be actively accessing the monitor at any one time. The access procedures are critical sections. A monitor may have a queue of processes that are waiting to access it.

  • Event flags: A memory word used as a synchronization mechanism. Application code may associate a different event with each bit in a flag. A thread can wait for either a single event or a combination of events by checking one or multiple bits in the corresponding flag. The thread is blocked until all of the required bits are set (AND) or until at least one of the bits is set (OR).

  • Mailboxes/messages: A means for two processes to exchange information and that may be used for synchronization.

  • Spinlocks: Mutual exclusion mechanism in which a process executes in an infinite loop waiting for the value of a lock variable to indicate availability.

Notice that:

  • A mutex can be released only by the thread that had acquired it.
  • A binary semaphore can be signaled by any thread (or process).

Introduction

信号量是操作系统提供的一种协调共享资源访问的方法,由一个整形(sem)变量和两个原子操作组成。

  • P():减少

    • sem -1
    • 若 sem <0,进入等待,否则继续
  • V():增加

    • sem +1
    • 若 sem 0,则唤醒一个进程
  • 信号量是被保护的整数变量:初始化完成后,只能通过 P(),V() 进程修改且 PV 是原子操作。

  • P() 可能阻塞,V() 不会阻塞

For signaling, special variables called semaphores are used.

  • To transmit a signal via semaphore s, a process executes the primitive semSignal (s).
  • To receive a signal via semaphore s, a process executes the primitive semWait (s); if the corresponding signal has not yet been transmitted, the process is suspended until the transmission takes place.

To achieve the desired effect, we can view the semaphore as a variable that has an integer value upon which only three operations are defined:

  1. A semaphore may be initialized to a nonnegative integer value.
  2. The semWait operation decrements the semaphore value. If the value becomes negative, then the process executing the semWait is blocked. Otherwise, the process continues execution.
  3. The semSignal operation increments the semaphore value. If the resulting value is less than or equal to zero, then a process blocked by a semWait operation, if any, is unblocked.
struct semaphore {
    int count;
    queueType queue;
};
void semWait(semaphore s) // P 操作
{
    s.count--;
    if (s.count < 0) {
    /* place this process in s.queue */;
    /* block this process */;
    }
}
void semSignal(semaphore s) { // V 操作
    s.count++;
    if (s.count<= 0) {
    /* remove a process P from s.queue */;
    /* place process P on ready list */;
    }
}

Categories

  • Binary Semaphore, may only take on the value 0 and 1.

    • Three Operations:

      1. binary semaphore may be initialized to 0 or 1.
      2. The semWaitB operations checks the semaphore value. If the value is zero, then the process executing the semWaitB is blocked.
      3. The semSignalB operation checks to see if any processes are blocked on this semaphore (semaphore value equals 0). If so, then a process blocked by a semWaitB operation is unblocked. If no processes are blocked, then the value of the semaphore is set to one.
      struct semaphore {
        int count;
        queueType queue;
      };
      void semWait(semaphore s){
        s.count--;
        if (s.count < 0) {
          /*place this process in s.queue*/;
          /*block this process*/;
        }
      }
      void semSignal(semaphore s) {
        s.count++;
        if (s.count<= 0) {
          /*remove a process P from s.queue*/;
          /*place process P on ready list*/;
        }
      }
      
  • The nonbinary semaphore is often referred to as either a counting semaphore or a general semaphore.

Applications

Mutual Exclusion

每个 Critical Section 设置一个信号量,初值为 1。

  • Consider n processes, identified in the array P(i), all of which need access to the same resource. Each process has a critical section used to access the resource.
  • In each process, a semWait (s) is executed just before its critical section.
  • If the value of s becomes negative, the process is blocked. If the value is 1, then it is decremented to 0 and the process immediately enters its critical section.
/* program mutualexclusion */
const int n = /* number of processes */;
semaphore s = 1;
void P(int i) {
  while (true) {
    semWait(s);
    /* critical section */;
    semSignal(s);
    /* remainder */;
  }
}
void main() {
  parbegin (P(1), P(2), . . . , P(n));
}

必须成对使用 PV 操作。

Conditional Synchronization

用信号量实现条件同步:每个条件同步设置一个信号量

condition = new Semaphore(0);
// Thread A
...M...
condition -> P();
.. N ..

// Thread B
.. X ..
condition -> V();
.. Y ..

The Producer/Consumer Problem

  • There are one or more producers generating some type of data (records, characters) and placing these in a buffer.(一个或多个生产者在生成数据后放在一个缓冲区里)
  • There is a single consumer that is taking items out of the buffer one at a time.(单个消费者从缓冲区取出数据处理)
  • The system is to be constrained to prevent the overlap of buffer operations, that is only one agent (producer or consumer) may access the buffer at any one time.(系统限制同时访问)

The problem is to make sure that the producer won’t try to add data into the buffer if it’s full, and that the consumer won’t try to remove data from an empty buffer.

Note that variable n is the number of items in the buffer and semaphore s is used to enforce mutual exclusion.

/* program producerconsumer */
semaphore n = 0, s = 1;
void producer() {
  while (true) {
    produce();
    semWait(s);
    append();
    semSignal(s);
    semSignal(n);
  }
}
void consumer() {
  while (true) {
    semWait(n);
    semWait(s);
    take();
    semSignal(s);
    consume();
  }
}
void main() {
  parbegin (producer, consumer);
}

Now suppose the semWait (n) and semWait (s) operations are accidentally reversed. This produces a serious, indeed a fatal, flaw. If the consumer ever enters its critical section when the buffer is empty (n.count = 0), then no producer can ever append to the buffer and the system is deadlocked. This is a good example of the subtlety of semaphores and the difficulty of producing correct designs.

Finally, let us add a new and realistic restriction to the producer/consumer problem: namely, that the buffer is finite. The buffer is treated as a circular storage and pointer values must be expressed modulo the size of the buffer. The following relationships hold: |Block on:|Unblock on:| |:--:|:--:| |Producer: insert in full buffer |Consumer: item inserted| |Consumer: remove from empty buffer| Producer: item removed| The producer and consumer functions can be expressed as follows (variable in and out are initialized to 0 and n is the size of the buffer):

Producer and Consumer with Finite Buffer

/* program boundedbuffer */
const int sizeofbuffer = /* buffer size */;
semaphore s = 1, n = 0, e = sizeofbuffer;
void producer() {
  while (true) {
    produce();
    semWait(e);
    semWait(s);
    append();
    semSignal(s);
    semSignal(n);
  }
}
void consumer() {
  while (true) {
    semWait(n);
    semWait(s);
    take();
    semSignal(s);
    semSignal(e);
    consume();
  }
}
void main() {
  parbegin (producer, consumer);
}

Implementation of Semaphores

Implementation of Semaphores

Monitor

Monitor Outline A monitor is a software module consisting of one or more procedures, an initialization sequence, and local data. The chief characteristics of a monitor are the following:

  1. The local data variables are accessible only by the monitor’s procedures and not by any external procedure
  2. A process enters the monitor by invoking one of its procedures.
  3. Only one process may be executing in the monitor at a time; any other processes that have invoked the monitor are blocked, waiting for the monitor to become available.

管程是一种用于多线程互斥访问共享资源的程序结构,任一时刻最多只有一个线程执行管程代码。

Condition Variable

A monitor supports synchronization by the use of condition variables that are contained within the monitor and accessible only within the monitor. Condition variables are a special data type in monitors, which are operated on by two functions:

  • cwait (c): Suspend execution of the calling process on condition c. The monitor is now available for use by another process.
  • csignal (c): Resume execution of some process blocked after a cwait on the same condition. If there are several such processes, choose one of them; if there is no such process, do nothing.

Structure of Monitor:

Structure of Monitor

Although a process can enter the monitor by invoking any of its procedures, we can think of the monitor as having a single entry point that is guarded so only one process may be in the monitor at a time. Other processes that attempt to enter the monitor join a queue of processes blocked waiting for monitor availability.

Once a process is in the monitor, it may temporarily block itself on condition x by issuing cwait (x); it is then placed in a queue of processes waiting to reenter the monitor when the condition changes, and resume execution at the point in its program following the cwait (x) call. If a process that is executing in the monitor detects a change in the condition variable x, it issues csignal (x), which alerts the corresponding condition queue that the condition has changed.

Implementation:

class Condition {
    int numWaiting = 0;
    WaitQueue q;
}
Condition::Wait(lock){
    numWaiting++;
    Add this thread t to q;
    release(lock);
    schedule(); //need mutex
    require(lock);
}
Condition::Signal(){
    if (numWaiting > 0) {
        Remove a thread t from q;
        wakeup(t); //need mutex
        numWaiting--;
    }
}

The Producer/Consumer Problem Implemented in Monitor

/* program producerconsumer */
monitor boundedbuffer;
char buffer [N]; /* space for N items */
int nextin, nextout; /* buffer pointers */
int count; /* number of items in buffer */
cond notfull, notempty; /* condition variables for synchronization */

void append (char x) {
  if (count == N) cwait(notfull); /* buffer is full; avoid overflow */
  buffer[nextin] = x;
  nextin = (nextin + 1) % N;
  count++;
  /* one more item in buffer */
  csignal (notempty); /*resume any waiting consumer */
}
void take (char x) {
  if (count == 0) cwait(notempty); /* buffer is empty; avoid underflow */
  x = buffer[nextout];
  nextout = (nextout + 1) % N;
  count--; /* one fewer item in buffer */
  csignal (notfull); /* resume any waiting producer */
}

monitor()
{ /* monitor body */
  nextin = 0; nextout = 0; count = 0; /* buffer initially empty */
}

void producer() {
  char x;
  while (true) {
    produce(x);
    append(x);
  }
}
void consumer() {
  char x;
  while (true) {
    take(x);
    consume(x);
  }
}
void main() {
  parbegin (producer, consumer);
}

Readers/Writers Problem

The readers/writers problem is defined as follows: There is a data area shared among a number of processes. The data area could be a file, a block of main memory, or even a bank of processor registers. There are a number of processes that only read the data area (readers) and a number that only write to the data area (writers). The conditions that must be satisfied are as follows:

  1. Any number of readers may simultaneously read the file.
  2. Only one writer at a time may write to the file.
  3. If a writer is writing to the file, no reader may read it.

Thus, readers are processes that are not required to exclude one another, and writers are processes that are required to exclude all other processes, readers and writers alike.

Writers Have Priority

/* program readersandwriters */
int readcount,writecount; semaphore x = 1, y = 1, z = 1, wsem = 1, rsem = 1;
void reader() {
  while (true) {
    semWait (z);
    semWait (rsem);
    semWait (x);
    readcount++;
    if (readcount == 1) semWait (wsem);
    semSignal (x);
    semSignal (rsem);
    semSignal (z);
    READUNIT();
    semWait (x);
    readcount--;
    if (readcount == 0) semSignal (wsem);
    semSignal (x);
  }
}
void writer () {
  while (true) {
    semWait (y);
    writecount++;
    if (writecount == 1) semWait (rsem);
    semSignal (y);
    semWait (wsem);
    WRITEUNIT();
    semSignal (wsem);
    semWait (y);
    writecount--;
    if (writecount == 0) semSignal (rsem);
    semSignal (y);
  }
}
void main() {
  readcount = writecount = 0;
  parbegin (reader, writer);
}

State of the Process Queues for Program:

  • Readers only in the system
    • wsem set
    • no queues
  • Writers only in the system
    • wsem and rsem set
    • writers queue on wsem
  • Both readers and writers with read first
    • wsem set by reader
    • rsem set by writer
    • all writers queue on wsem
    • one reader queues on rsem
    • other readers queue on z
  • Both readers and writers with write first
    • wsem set by writer
    • rsem set by writer
    • writers queue on wsem
    • one reader queues on rsem
    • other readers queue on z

Writers Have Priority

Readers Have Priority

/* program readersandwriters */
int readcount;
semaphore x = 1,wsem = 1;
void reader() {
  while (true) {
    semWait (x);
    readcount++;
    if(readcount == 1) semWait (wsem);
    semSignal (x);
    READUNIT();
    semWait (x);
    readcount--;
    if(readcount == 0) semSignal (wsem);
    semSignal (x);
  }
}
void writer() {
  while (true){
    semWait (wsem);
    WRITEUNIT();
    semSignal (wsem);
  }
}
void main() {
  readcount = 0;
  parbegin (reader,writer);
}

Readers Have Priority

Deadlock and Starvation

Deadlock can be defined as the permanent blocking of a set of processes that either compete for system resources or communicate with each other. A set of processes is deadlocked when each process in the set is blocked awaiting an event (typically the freeing up of some requested resource) that can only be triggered by another blocked process in the set. Deadlock is permanent because none of the events is ever triggered. Unlike other problems in concurrent process management, there is no efficient solution in the general case. Illustration of Deadlock:

Illustration of Deadlock

死锁问题:由于竞争资源或者通信关系,两个或更多线程在执行中出现,永远相互等待只能由其他进程引发的事件。

General Categories of Resources

Reusable Resource

  • 资源不能被删除且在任何时刻只能有一个进程使用

  • 进程释放资源后,其他进程可以重用

Consumable Resource

  • 资源的创建和销毁(I/O 缓冲区的中断、信号、信息等)
  • 可能出现死锁

Resource Allocation Graphs

A useful tool in characterizing the allocation of resources to processes is the resource allocation graph.The resource allocation graph is a directed graph that depicts a state of the system of resources and processes, with each process and each resource represented by a node. A graph edge directed from a process to a resource indicates a resource that has been requested by the process but not yet granted. Within a resource node, a dot is shown for each instance of that resource.

Examples of Resource Allocation Graphs

The Conditions for Deadlock

  1. Mutual exclusion. Only one process may use a resource at a time. No process may access a resource unit that has been allocated to another process.
  2. Hold and wait. A process may hold allocated resources while awaiting assignment of other resources.
  3. No preemption. No resource can be forcibly removed from a process holding it.
  4. Circular wait. A closed chain of processes exists, such that each process holds at least one resource needed by the next process in the chain

Deadline Avoidance

  • Deadlock Prevention: 确保系统永远不会进入死锁状态。
  • Deadlock Avoidance:在使用前进行判断,只允许不会出现死锁的进程请求资源
  • Deadlock Detection & Recovery:在检测到运行系统进入死锁状态后,进行恢复
  • 安全状态和死锁的关系:
    • 系统处于安全状态,一定没有死锁
    • 系统处于不安全状态,可能出现死锁
      • 避免死锁就是确保系统不会进入不安全状态

Banker's Algorithm

Process Initiation Denial

Consider a system of processes and different types of resources. Let us define the following vectors and matrices:

  • Resource . Total amount of each resource in the system.
  • Available . Total amount of each resource not allocated to any process.
  • Claim . = requirement of process for resource .
  • Allocation . = current allocation to process of resource .

The following relationships hold:

  1. , for all : All resources are either available or allocated.

  2. , for all : No process can claim more than the total amount of resources in the system.

  3. , for all : No process is allocated more resources of any type than the process originally claimed to need.

With these quantities defined, we can define a deadlock avoidance policy that refuses to start a new process if its resource requirements might lead to deadlock. Start a new process only if

Resource Allocation Denial

  • A safe state is one in which there is at least one sequence of resource allocations to processes that does not result in a deadlock (i.e., all of the processes can be run to completion).

  • An unsafe state is, of course, a state that is not safe.

In terms of the matrices and vectors introduced earlier, the condition to be met for process is: Safe State:

Safe State

Unsafe State:

Unsafe State

It is unsafe because each process will need at least one additional unit of R1, and there are none available. Thus, on the basis of deadlock avoidance, the request by P1 should be denied and P1 should be blocked. It is important to point out that Figure (b) is not a deadlocked state. It merely has the potential for deadlock. If P1 were run from this state, it would subsequently release one unit of R1 and one unit of R3 prior to needing these resources again. If that happened, the system would return to a safe state.

Thus, the deadlock avoidance strategy does not predict deadlock with certainty; it merely anticipates the possibility of deadlock and assures that there is never such a possibility.

Deadlock Detection

The Allocation matrix and Available vector described in the previous section are used. In addition, a request matrix is defined such that represents the amount of resources of type requested by process .

The algorithm proceeds by marking processes that are not part of a deadlocked set. Initially, all processes are unmarked. Then the following steps are performed:

  1. Mark each process that has a row in the Allocation matrix of all zeros. A process that has no allocated resources cannot participate in a deadlock.
  2. Initialize a temporary vector to equal the Available vector.
  3. Find an index such that process is currently unmarked and the ith row of is less than or equal to . That is, , for . If no such row is found, terminate the algorithm.
  4. If such a row is found, mark process and add the corresponding row of the allocation matrix to . That is, set , for . Return to step 3.

Example I

Example of Deadlock

  1. Mark P4, because P4 has no allocated resources.
  2. Set
  3. The request of process P3 is less than or equal to , so mark P3 and set
  4. No other unmarked process has a row in Q that is less than or equal to W. Therefore, terminate the algorithm.

The algorithm concludes with P1 and P2 unmarked, indicating these processes are deadlocked.

Example II

Example Two of Deadlock

  1. 由于已分配进程中找不到全 0 向量,所以不能 mark。
  2. 计算 向量为全 0。
  3. T0 的资源请求 向量 小于等于
  4. T2 的资源请求 向量 小于等于
  5. T3 的资源请求 向量 小于等于
  6. T4 的资源请求 向量 小于等于
  7. T1 的资源请求 向量 小于等于

因此执行序列 T0, T2, T3, T4, T1 合法。

Deadlock Recovery

Once deadlock has been detected, some strategy is needed for recovery. The following are possible approaches, listed in the order of increasing sophistication:

  1. Abort all deadlocked processes.
  2. Back up each deadlocked process to some previously defined checkpoint, and restart all processes.
  3. Successively abort deadlocked processes until deadlock no longer exists.
  4. Successively preempt resources until deadlock no longer exists.

Dining Philosophers Problem

  • 5个哲学家围绕一张圆桌而坐
    • 桌子上放着5支叉子
    • 每两个哲学家之间放一支
  • 哲学家的动作包括思考和进餐
    • 进餐时需同时拿到左右两边的叉子
    • 思考时将两支叉子放回原处

如何保证哲学家们的动作有序进行?如:不出现有人永远拿不到叉子

Solution 1

DPP Solution 1
不正确,五个人同时拿左边的叉子,所有人的右边都没有叉子了(死锁)。

Solution 2

DPP Solution 2
互斥访问正确,但每次只允许一人进餐。

Solution Three

DPP Solution 3
没有死锁,可有多人同时就餐。

A Solution Using a Monitor

DPP Monitor

Interprocess Communication

  • 进程通信是进程进行通信和同步的机制
  • IPC 提供2个基本操作
    • 发送操作:send(message)
    • 接收操作:receive(message)
  • 进程通信流程
    • 在通信进程间建立通信链路
    • 通过 send/receive 交换消息
  • 进程链路特征
    • 物理 (如,共享内存,硬件总线)
    • 逻辑 (如,逻辑属性)

Direct Communication

  • 进程必须正确的命名对方

    • send (P, message) – 发送信息到进程P
    • receive(Q, message) – 从进程 Q接受消息
  • 通信链路的属性

    • 自动建立链路
    • 一条链路恰好对应一对通信进程
    • 每对进程之间只有一个链接存在
    • 链接可以是单向的,但通常为双向的

Indirect Communication

  • 通过操作系统维护的消息队列实现进程间的消息接收和发送
    • 每个消息队列都有一个唯一的标识
    • 只有共享了相同消息队列的进程,才能够通信
  • 通信链路的属性
    • 只有共享了相同消息队列的进程,才建立连接
    • 连接可以是单向或双向
    • 消息队列可以与多个进程相关联
    • 每对进程可以共享多个消息队列
  • 通信流程
    • 创建一个新的消息队列
    • 通过消息队列发送和接收消息
    • 销毁消息队列
  • 基本通信操作
    • send(A, message) – 发送消息到队列 A
    • receive(A, message) – 从队列 A 接受消息

Blocking and Non-blocking Communication

  • 进程通信可划分为阻塞(同步)或非阻塞(异步)
  • 阻塞通信:阻塞发送和接受
  • 非阻塞通信:非阻塞发送和接受
    没有消息发送时,接收者在请求接收消息后,接收不到任何消息

Signal

  • 进程间的软件中断通知和处理机制
  • 信号的接收处理
    • 捕获 (catch):执行进程指定的信号处理函数被调用
    • 忽略 (ignore):执行操作系统指定的缺省处理
    • 屏蔽 (mask):禁止进程接收和处理信号
  • 不足
    • 传送的信息量小,只有一个信号类型

Pipe

  • 进程间基于内存文件的通信机制
    • 子进程从父进程继承文件描述符
    • 缺省文件描述符:0 stdin, 1 stdout, 2 stderr
  • 进程不知道(或不关心)的另一端
    • 可能从键盘、文件、程序读取
    • 可能写入到终端、文件、程序

Message Queue

  • 消息队列是由操作系统维护的以字节序列为基本单位的间接通信机制
    • 每个消息 Message 是一个字节序列
    • 相同标识的消息组成按先进先出顺序组成一个消息队列 Message Queues

Shared Memory

  • 共享内存是把同一个物理内存区域同时映射到多个进程的内存地址空间的通信机制
  • 进程
    • 每个进程都有私有内存地址空间
    • 每个进程的内存地址空间需明确设置共享内存段
  • 线程
    • 同一进程中的线程总是共享相同的内存地址空间
  • 优点
    • 快速、方便地共享数据
  • 不足
    • 必须用额外的同步机制来协调数据访问

File System

Files and File Systems

From the user’s point of view, one of the most important parts of an operating system is the file system. The file system permits users to create data collections, called files, with desirable properties, such as:

  • Long-term existence: Files are stored on disk or other secondary storage and do not disappear when a user logs off.

  • Sharable between processes: Files have names and can have associated access permissions that permit controlled sharing.

  • Structure: Depending on the file system, a file can have an internal structure that is convenient for particular applications. In addition, files can be organized into a hierarchical or more complex structure to reflect the relationships among files.

Any file system provides not only a means to store data organized as files, but a collection of functions that can be performed on files. Typical operations include the following:

  • Create
  • Delete
  • Open
  • Close
  • Read
  • Write

File Structure

Four terms are in common use when discussing files:

Field

A field is the basic element of data. Depending on the file design, fields may be fixed length or variable length. In the latter case, the field often consists of fwo or three subfields: the actual value to be stored, the name of the fields, and in some cases, the length of the field. In other cases of variable-length fields, the length of the field is indicated by the use of special demarcation (boundary) symbols between fields.

Record

A record is a collection of related fields that can be treated as a unit by some application program. A record will be of variable length if some of its fields are of variable length or if the number of fields may vary. In the latter case, each field is usually accompanied by a field name. In either case, the entire record usually includes a length field.

File

A file is a collection of similar records. The file is treated as a single entity by users and applications and may be referenced by name. Files have file names and may be created and deleted.

Database

A database is a collection of related data. The essential aspects of a database are that the relationships that exist among elements of data are explicit.

File Management Systems

A file management system is that set of system software that provides services to users and applications in the use of files. Typically, the only way a user or application may access files is through the file management system.
The following objectives for a file management system are as follows:

  • To meet the data management needs and requirements of the user, which include storage of data and the ability to perform the aforementioned operations
  • To guarantee, to the extend possible, that the data in the file are valid
  • To optimize performance, both from the system point of view in terms of overall throughput, and from the user’s point of view in terms of response time
  • To provide I/O support for a variety of storage device types
  • To minimize or eliminate the potential for lost or destroyed data
  • To provide a standardized set of I/O interface routines to user processes
  • To provide I/O support for multiple users, in the case of multiple-user systems

With respect to the first point, meeting user requirements, the extent of such requirements depends on the variety of applications and the environment in which the computer system will be used. For an interactive, general-purpose system, the following constitute a minimal set of requirements:

  1. Each user should be able to create, delete, read, write, and modify files.
  2. Each user may have controlled access to other users’ files.
  3. Each user may control what types of accesses are allowed to the user’s files
  4. Each user should be able to move data between files.
  5. Each user should be able to back up and recover the user’s files in case of damage.
  6. Each user should be able to access his or her files by name rather than by numeric identifier.

File System Architecture

File System Software Architecture

File Management Functions

Element of File Management

File Organization and Access

In choosing a file organization, several criteria are important:

  • Short access time
  • Ease of update
  • Economy of storage
  • Simple maintenance
  • Reliability

Five Fundamental Organizations

  1. The pile
  2. The sequential file
  3. The indexed sequential file
  4. The indexed file
  5. The direct, or hashed, file

Common File Organization

File Directories

Contents

Associated with any file management system and collection of files is a file directory. The directory contains information about the files, including attributes, location, and ownership. Much of this information, especially that concerned with storage, is managed by the operating system. The directory is itself a file, accessible by various file management routines. Although some of the information in directories is available to users and applications, this is generally provided indirectly by system routines.

Structure

To understand the requirements for a file structure, it is helpful to consider the types of operations that may be performed on the directory:

  • Search: When a user or application references a file, the directory must be searched to find the entry corresponding to that file.
  • Create file: When a new file is created, an entry must be added to the directory.
  • Delete file: When a file is deleted, an entry must be removed from the directory.
  • List directory: All or a portion of the directory may be requested. Generally, this request is made by a user and results in a listing of all files owned by that user, plus some of the attributes of each file (e.g., type, access control information, usage information).
  • Update directory: Because some file attributes are stored in the directory, a change in one of these attributes requires a change in the corresponding directory entry.

Naming

Users need to be able to refer to a file by a symbolic name. Clearly, each file in the system must have a unique name in order that file references be unambiguous. On the other hand, it is an unacceptable burden on users to require they provide unique names, especially in a shared system.

The use of a tree-structured directory minimizes the difficulty in assigning unique names. Any file in the system can be located by following a path from the root or master directory down various branches until the file is reached. The series of directory names, culminating in the file name itself, constitutes a pathname for the file.

Tree-structured Directory

File Sharing

In a multi-user system, there is almost always a requirement for allowing files to be shared among a number of users. Two issues arise: access rights and the management of simultaneous access.

Access Rights

The following list is representative of access rights that can be assigned to a particular user for a particular file:

  • None: The user may not even learn of the existence of the file, much less access it. To enforce this restriction, the user would not be allowed to read the user directory that includes this file.
  • Knowledge: The user can determine that the file exists and who its owner is. The user is then able to petition the owner for additional access rights.
  • Execution: The user can load and execute a program but cannot copy it. Proprietary programs are often made accessible with this restriction.
  • Reading: The user can read the file for any purpose, including copying and execution. Some systems are able to enforce a distinction between viewing and copying. In the former case, the contents of the file can be displayed to the user, but the user has no means for making a copy.
  • Appending: The user can add data to the file, often only at the end, but cannot modify or delete any of the file’s contents. This right is useful in collecting data from a number of sources.
  • Updating: The user can modify, delete, and add to the file’s data. This normally includes writing the file initially, rewriting it completely or in part, and removing all or a portion of the data. Some systems distinguish among different degrees of updating.
  • Changing protection: The user can change the access rights granted to other users. Typically, this right is held only by the owner of the file. In some systems, the owner can extend this right to others. To prevent abuse of this mechanism, the file owner will typically be able to specify which rights can be changed by the holder of this right.
  • Deletion: The user can delete the file from the file system.

One user is designated as the owner of a given file, usually the person who initially created the file. The owner has all of the access rights listed previously and may grant rights to others. Access can be provided to different classes of users:

  • Specific user: Individual users who are designated by user ID
  • User groups: A set of users who are not individually defined. The system must have some way of keeping track of the membership of user groups.
  • All: All users who have access to this system. These are public files.

Simultaneous Access

When access is granted to append or update a file to more than one user, the operating system or file management system must enforce discipline. A brute-force approach is to allow a user to lock the entire file when it is to be updated. A finer grain of control is to lock individual records during update. Essentially, this is the readers/writers problem. Issues of mutual exclusion and deadlock must be addressed in designing the shared access capability.

Record Blocking

Given the size of a block, there are three methods of blocking that can be used:

  1. Fixed blocking: Fixed-length records are used, and an integral number of records are stored in a block. There may be unused space at the end of each block. This is referred to as internal fragmentation.
  2. Variable-length spanned blocking: Variable-length records are used and are packed into blocks with no unused space. Thus, some records must span two blocks, with the continuation indicated by a pointer to the successor block.
  3. Variable-length unspanned blocking: Variable-length records are used, but panning is not employed. There is wasted space in most blocks because of the inability to use the remainder of a block if the next record is larger than the remaining unused space. Fixed blocking is the common mode for sequential files with fixed-length records. Variable-length spanned blocking is efficient of storage and does not limit the size of records. However, this technique is difficult to implement. Records that span two blocks require two I/O operations, and files are difficult to update, regardless of the organization. Variable-length unspanned blocking results in wasted space and limits record size to the size of a block. The record-blocking technique may interact with the virtual memory hardware, if such is employed. In a virtual memory environment, it is desirable to make the page the basic unit of transfer. Pages are generally quite small, so it is impractical to treat a page as a block for unspanned blocking. Accordingly, some systems combine multiple pages to create a larger block for file I/O purposes. This approach is used for VSAM files on IBM mainframes.

Secondary Storage Management

On secondary storage, a file consists of a collection of blocks. The operating system or file management system is responsible for allocating blocks to files. This raises two management issues. First, space on secondary storage must be allocated to files, and second, it is necessary to keep track of the space available for allocation. We will see that these two tasks are related; that is, the approach taken for file allocation may influence the approach taken for free space management. Further, we will see that there is an interaction between file structure and allocation policy.

File Allocation

File System Methods

Volumes

Volume: A collection of addressable sectors in secondary memory that an OS or application can use for data storage. The sectors in a volume need not be consecutive on a physical storage device; instead, they need only appear that way to the OS or application. A volume may be the result of assembling and merging smaller volumes.

Reliability

Consider the following scenario:

  1. User A requests a file allocation to add to an existing file.
  2. The request is granted and the disk and file allocation tables are updated in main memory but not yet on disk.
  3. The system crashes and subsequently restarts.
  4. User B requests a file allocation and is allocated space on disk that overlaps the last allocation to user A.
  5. User A accesses the overlapped portion via a reference that is stored inside A’s file.

This difficulty arose because the system maintained a copy of the disk allocation table and file allocation table in main memory for efficiency. To prevent this type of error, the following steps could be performed when a file allocation is requested:

  1. Lock the disk allocation table on disk. This prevents another user from causing alterations to the table until this allocation is completed.
  2. Search the disk allocation table for available space. This assumes a copy of the disk allocation table is always kept in main memory. If not, it must first be read in.
  3. Allocate space, update the disk allocation table, and update the disk. Updating the disk involves writing the disk allocation table back onto disk. For chained disk allocation, it also involves updating some pointers on disk.
  4. Update the file allocation table and update the disk.
  5. Unlock the disk allocation table.

Unix File Management

In the UNIX file system, six types of files are distinguish:

  1. Regular, or ordinary: Contains arbitrary data in zero or more data blocks. Regular files contain information entered in them by a user, an application program, or a system utility program. The file system does not impose any internal structure to a regular file, but treats it as stream of bytes.
  2. Directory: Contains a list of file names plus pointers to associated inodes (index nodes), described later. Directories are hierarchically organized. Directory files are actually ordinary files with special write protection privileges on only the file system can write into them, while read access is available to user programs.
  3. Special: Contains no data but provides a mechanism to map physical devices to file names. The file names are used to access peripheral devices, such as terminals and printers. Each I/O device is associated with a special file.
  4. Named pipes: A pipe is an interprocess communications facility. A pipe file buffers data received in its input so a process that reads from the pipe's output receives the data on a first-in-first-out basis.
  5. Links: In essence, a link is an alternative file name for an existing file.
  6. Symbolic Links: This is a data file that contains the name of the file to which it is linked.

Inodes

Modern UNIX operating systems support multiple file systems but map all of these into a uniform underlying system for supporting file systems and allocating disk space to files. All types of UNIX files are administered by the OS by means of inodes. An inode is a control structure that contains the key information needed by the operating system for a particular file. Several file names may be associated with a single inode, but an active inode is associated with exactly one file, and each file is controlled by exactly on inode.
The FreeBSD inode structure includes the following data elements:

  • The type and access mode of the file
  • The file's owner and group-access identifiers
  • The time when the file was created, when it was most recently read and written, and when its inode was most recently updated by the system
  • The size of the file by bytes
  • A sequence of block pointers, explained in the next subsection
  • The number of physical blocks used by the file, including blocks used to hold indirect pointers and attributes
  • The number of directory entries that reference the file
  • The kernel and user-settable flags that describe the characteristics of the file
  • The generation number of teh file (a random selected number assigned to the inode each time that the latter is allocated to a new file ; the generation number is use dto detect references to delete files)
  • The blocksize of the data blocks references by the inode (typically the same as, but sometimes larger than, the file system blocksize)
  • The size of the extended attribute information
  • Zero or more extended attribute entries

Structure of FreeBSD Inode and File

File Allocation

File allocation is done on a block basis. Allocation is dynamic, as needed, rather than using preallocation. Hence the blocks of a file on disk are not necessarily contiguous. An indexed method is used to keep track of each file, with part of the index stored in the inode for the file. In all UNIX implementations, the inode includes a number of direct pointers and three indirect pointers (single, double, triple).

Directories

Directories are structured in a hierarchical tree. Each directory can contains file and/or other directories. A directory inside another directory is referred to as a subdirectory. As was mentioned, a directory is simply a file that contains a list of file names plus pointers to associated inodes. Each directory entry (dentry) contains a name for the associated file or subdirectory plus an integer called the i-number (index number). When the file or directory is accessed, its i-number is used as an index into the inode table.

Volume Structure

A UNIX file system resides on a single logical disk or disk partition and is laid out with the following elements:

  • Boot block: Contains code required to boot the operating system
  • Superblock: Contains attributes and information about the file system, such as partition size, and inode table size
  • Inode table: The collection of inodes for each file
  • Data block: Storage space available for data files and subdirectories

Linux Virtual File System

Linux includes a versatile and powerful file-handling facility, designed to support a wide variety of file management systems and file structures. The approach taken in Linux is to make use of a virtual file system (VFS), which presents a single, uniform file system interface to user processes. The VFS defines a common file model that is capable of representing any conceivable file system’s general feature and behavior.

Linux Virtual File System Context

Linux Virtual File System Concept

VFS is an object-oriented scheme. Because it is written in C, rather than a language that supports object programming (such as C++ or Java), VFS objects are implemented simply as C data structures. Each object contains both data and pointers to file-system-implemented functions that operate on data. The four primary object types in VFS are as follows:

  • Superblock object: Represents a specific mounted file system
  • Inode object: Represents a specific file
  • Dentry object: Represents a specific directory entry
  • File object: Represents an open file associated with a process

The Superblock Object

The superblock object stores information describing file system. Typically, the superblock corresponds to the file system superblock or file system control block, which is stored in a special sector on disk.
The superblock object consists of a number of data items. Examples include the following:

  • The device this file system is mounted on
  • The basic block size of the file system
  • Dirty flag, to indicate that the superblock has been changed but not written back to disk
  • File system type
  • Flags, such as a read-only flag
  • Pointer to the root of the file system directory
  • List of open files
  • Semaphore for controlling access to the file system
  • List of superblock operations

The Inode Object

An inode is associated with each file. The inode object holds all the information about a named file except its name and the actual data contents of the file. Items contained in an inode object include owner, group, permissions, access times for a file, size of data it holds, and number of links.
The inode object also includes an inode operations object that describes the file system’s implemented functions that the VFS can invoke on an inode. The methods defined for the inode object include the following:

  • create: Creates a new inode for a regular file associated with a dentry object in some directory
  • lookup: Searches a directory for an inode corresponding to a file name
  • mkdir: Creates a new inode for a directory associated with a dentry object in some directory

The Dentry Object

A dentry (directory entry) is a specific component in a path. The component may be either a directory or a file name. Dentry objects facilitate quick lookups to files and directories, and are used in a dentry cache for that purpose. The dentry object includes a pointer to the inode and superblock. It also includes a pointer to the parent dentry and pointers to any subordinate dentrys.

The File Object

The file object is used to represent a file opened by a process. The object is created in response to the open() system call, and destroyed in response to the close() system call. The file object consists of a numbers of items, including the following:

  • Dentry object associated with the file
  • File system containing the file
  • File objects usage counter
  • User's user ID
  • User's group ID
  • File pointer, which is the current position in the file from which the current position in the file from which the next operation will take place

The file object also includes an inode operations object that describes the file system's implemented functions that the VFS can invoke on a file object. The methods defined for the file object include read, write, open, release, and lock.

Caches

The VFS employs three caches to improve performance:

  1. Inode cache: Because every file and directory is represented by a VFS inode, a directory listing command or a file access command causes a number of inodes to be accessed. The inode cache stores recently visited inodes to make access quicker.
  2. Directory cache: The directory cache stores the mapping between the full directory names and their inode numbers. This speeds up the process of listing a directory.
  3. Buffer cache: The buffer cache is independent of the file systems and it integrated into the mechanisms that the Linux kernel uses to allocate and read and write data buffers. As the real life systems read data from the underlying physical disks, this results in requests to the block device drivers to read physical blocks from the device that they control. So, if the same data is needed often, it will be retrieved from the buffer cache rather than read from the disk.