CS计算机代考程序代写 cache Java jvm algorithm x86 interpreter flex compiler file system PowerPoint Presentation

PowerPoint Presentation

Chapter 16: Virtual Machines

1

Chapter 16: Virtual Machines
Overview
History
Benefits and Features
Building Blocks
Types of Virtual Machines and Their Implementations
Virtualization and Operating-System Components
Examples

2

Chapter Objectives

To explore the history and benefits of virtual machines
To discuss the various virtual machine technologies
To describe the methods used to implement virtualization
To show the most common hardware features that support virtualization and explain how they are used by operating-system modules

3

Overview
Fundamental idea – abstract the hardware of a single computer into several different execution environments
Several components
Host – underlying hardware system that runs the virtual machines
Virtual machine manager (VMM) or hypervisor – creates and runs virtual machines by providing interface that is identical to the host
(Except in the case of paravirtualization)
Guest – process provided with virtual copy of the host
The Guest is usually an operating system
Single physical machine can run multiple operating systems concurrently, each in its own virtual machine

The software component that provides virtual environments is a VMM.

System Models

Non-virtual machine
Virtual machine

Implementation of VMMS
Vary greatly, with options including:
Type 0 Hypervisors – Hardware-based solutions that provide support for virtual machine creation and management via firmware
IBM LPARs and Oracle LDOMs are examples
Type 1 Hypervisors – Operating-system-like software built to provide virtualization
Including VMware ESX, Joyent SmartOS, and Citrix XenServer
Also includes general-purpose operating systems that provide standard functions as well as VMM functions
Including Microsoft Windows Server with HyperV and RedHat Linux with KVM
Type 2 Hypervisors – Applications that run on standard operating systems but provide VMM features to guest operating systems
Including VMware Workstation and Fusion, Parallels Desktop, and Oracle VirtualBox

Implementation of VMMs (cont.)
Other variations include:
Paravirtualization – Technique in which the guest operating system is modified to work in cooperation with the VMM to optimize performance
Programming-environment virtualization – VMMs do not virtualize real hardware but instead create an optimized virtual system
Used by Oracle Java and Microsoft.Net
Emulators – Allow applications written for one hardware environment to run on a very different hardware environment, such as a different type of CPU
Application containment – Not virtualization at all but rather provides virtualization-like features by segregating applications from the operating system, making them more secure, manageable
Including Oracle Solaris Zones, BSD Jails, and IBM AIX WPARs
Much variation due to breadth, depth and importance of virtualization in modern computing

FreeBSD jails
FreeBSD jails mainly aim at three goals:
Virtualization: Each jail is a virtual environment running on the host machine with its own files, processes, user and superuser accounts. From within a jailed process, the environment is almost indistinguishable from a real system.
Security: Each jail is sealed from the others, thus providing an additional level of security.
Ease of delegation: The limited scope of a jail allows system administrators to delegate several tasks which require superuser access without handing out complete control over the system.

FreeBSD jails
The FreeBSD jail does not however achieve true virtualization; it does not allow the virtual machines to run different kernel versions than that of the base system.
All virtual servers share the same kernel and hence expose the same bugs and potential security holes.
There is no support for clustering or process migration, so the host kernel and host computer is still a single point of failure for all virtual servers.
It is possible to use jails to safely test new software, but not new kernels.

History
First appeared in IBM mainframes in 1972
Allowed multiple users to share a batch-oriented system
Formal definition of virtualization helped move it beyond IBM
A VMM provides an environment for programs that is essentially identical to the original machine
Programs running within that environment show only minor performance decreases
The VMM is in complete control of system resources
In late 1990s Intel CPUs fast enough for researchers to try virtualizing on general purpose PCs
Xen and VMware created technologies, still used today
Virtualization has expanded to many OSs, CPUs, VMMs
For a robust open-source VMM check out www.virtualbox.org

Benefits and Features
Host system protected from VMs, VMs protected from each other
A virus less likely to spread
Sharing is provided though via shared file system volume, or via virtual or physical network communications
Common Feature: Freeze, suspend a running VM
Take Snapshot of a given state enabling
Clone by creating copy on another machine and running both original and copy
Great for OS research, better system development efficiency
Run multiple, different OSs on a single machine
Consolidation – running two or more separate systems and running them in VMs on one system.
app development – test the app on multiple operating systems

Benefits and Features (cont.)
Templating – create an OS + application VM, provide it to customers, use it to create multiple instances of that combination
Live migration – move a running VM from one host to another!
No interruption of user access (more about this later)
All those features taken together -> Cloud Computing
Using web-based APIs, programs tell cloud infrastructure (servers, networking, storage) to create thousands of new guests, VMs, virtual desktops
This is used by many multiuser gaming and photo-sharing sites today
A Virtual Environment might include 100 physical servers, each running 20 virtual servers. Without virtualization, 2,000 servers would require several system administrators. VMMs make it possible for one or two system administrators to manage this environment.

Building Blocks
Generally difficult to provide an exact duplicate of underlying machine
Especially if only dual-mode operation available on CPU
But getting easier over time as CPU features and support for VMM improves
Most VMMs implement Virtual CPU (VCPU) to represent state of CPU per guest as guest believes it to be
When guest context switched onto CPU by VMM, information from VCPU loaded and stored
Just as a general-purpose OS would use a PCB
We will examine one technique that is used

Building Block – Trap and Emulate
Dual mode CPU means guest executes in user mode
Kernel runs in kernel mode
Not safe to let guest kernel run in kernel mode too
So VM needs two modes – virtual user mode and virtual kernel mode
Both of which run in real user mode
Actions in guest that usually cause switch to kernel mode must cause switch to virtual kernel mode

Trap-and-Emulate (cont.)
How does switch from virtual user mode to virtual kernel mode occur?
Attempting a privileged instruction in user mode causes an error -> trap
VMM gains control, analyzes error, executes operation as attempted by guest
Returns control to guest in user mode
Most virtualization products use this at least in part
User mode instructions in guest runs at same speed as if not a guest
But kernel mode privileged runs slower due to trap-and-emulate
Especially a problem when multiple guests running, each needing trap-and-emulate
CPUs adding hardware support, and more CPU modes and guest state management to improve virtualization performance

Trap-and-Emulate
Virtualization Implementation

Building Block – Binary Translation
Some CPUs don’t have clean separation between privileged and nonprivileged instructions
Earlier Intel x86 CPUs are among them
Earliest Intel CPU designed for a calculator
Backward compatibility means difficult to improve
Consider Intel x86 popf instruction
Loads CPU flags register from contents of the stack
If CPU in privileged mode -> all flags replaced
If CPU in user mode -> only some flags replaced
No trap is generated

Binary Translation (cont.)
Other similar problem instructions we will call special instructions
Caused trap-and-emulate method considered impossible until 1998
Binary translation solves the problem
Basics are simple, but implementation very complex
If guest VCPU is in user mode, guest can run instructions natively
If guest VCPU in kernel mode (guest believes it is in kernel mode)
VMM examines every instruction guest is about to execute by reading a few instructions ahead of program counter
Non-special-instructions run natively
Special instructions translated into new set of instructions that perform equivalent task (for example changing the flags in the VCPU)

Binary Translation (cont.)
Implemented by translation of code within VMM
Code reads native instructions dynamically from guest, on demand, generates native binary code that executes in place of original code
Performance of this method would be poor without optimizations
Products like VMware use caching
Translate once, and when guest executes code containing special instruction cached translation used instead of translating again
Testing showed booting Windows XP as guest caused 950,000 translations, at 3 microseconds each, or 3 second (5 %) slowdown over native

Binary Translation Virtualization Implementation

Nested Page Tables
Memory management another general challenge to VMM implementations
How can VMM keep page-table state for both guests believing they control the page tables and VMM that does control the tables?
Common method (for trap-and-emulate and binary translation) is Nested Page Tables (NPTs)
Each guest maintains page tables to translate virtual to physical addresses
VMM maintains per guest NPTs to represent guest’s page-table state
Just as VCPU stores guest CPU state
When guest on CPU -> VMM makes that guest’s NPTs the active system page tables
Guest tries to change page table -> VMM makes equivalent change to NPTs and its own page tables
Can cause many more TLB* misses -> much slower performance

*Translation Lookaside Buffer

Building Blocks – Hardware Assistance
All virtualization needs some HW support
More support -> more feature rich, stable, better performance of guests
Intel added new VT-x instructions in 2005 and AMD the AMD-V instructions in 2006
Generally define more CPU modes – “guest” and “host”
VMM can enable host mode, define characteristics of each guest VM, switch to guest mode and guest(s) on CPU(s)
In guest mode, guest OS thinks it is running natively, sees devices (as defined by VMM for that guest)
Access to virtualized device, privileged instructions cause trap to VMM
CPU maintains VCPU, context switches it as needed

Nested Page Tables

Types of Virtual Machines and Implementations
Many variations as well as HW details
Assume VMMs take advantage of HW features
HW features can simplify implementation, improve performance
Whatever the type, a VM has a lifecycle
Created by VMM
Resources assigned to it (number of cores, amount of memory, networking details, storage details)
In type 0 hypervisor, resources usually dedicated
Other types dedicate or share resources, or a mix
When no longer needed, VM can be deleted, freeing resources
Steps simpler, faster than with a physical machine install
Can lead to virtual machine sprawl with lots of VMs, history and state difficult to track

videos

5:12

Types of VMs – Type 0 Hypervisor
Old idea, under many names by HW manufacturers
“partitions”, “domains”
A HW feature implemented by firmware
OS need to nothing special, VMM is in firmware that loads at boot time
Smaller feature set than other types
Each guest has dedicated HW
I/O a challenge – difficult to have enough devices, controllers to dedicate to each guest
Sometimes VMM implements a control partition running daemons that other guests communicate with for shared I/O
Can provide virtualization-within-virtualization (guest itself can be a VMM with guests (see next slide)
Other types have difficulty doing this……

Type 0 Hypervisor

Types of VMs – Type 1 Hypervisor
Commonly found in company datacenters
In a sense becoming “datacenter operating systems”
Datacenter managers control and manage OSs in new, sophisticated ways by controlling the Type 1 hypervisor
Consolidation of multiple OSs and apps onto less HW
Move guests between systems to balance performance
Snapshots and cloning
Special purpose operating systems that run natively on HW
Rather than providing system call interface, create run and manage guest OSs
Can run on Type 0 hypervisors but not on other Type 1s
Run in kernel mode
Guests generally don’t know they are running in a VM
Implement device drivers for host HW because no other component can
Also provide other traditional OS services like CPU and memory management

Types of VMs – Type 1 Hypervisor (cont.)
Another variation is a general-purpose OS that also provides VMM functionality
RedHat Enterprise Linux with KVM, Windows with Hyper-V, Oracle Solaris
Perform normal duties as well as VMM duties
Typically less feature rich than dedicated Type 1 hypervisors
In many ways, treat guest OSs as just another process
Albeit with special handling when guest tries to execute special instructions

Types of VMs – Type 2 Hypervisor
Less interesting from an OS perspective
Very little OS involvement in virtualization
VMM is simply another process, run and managed by host
Even the host doesn’t know they are a VMM running guests
Tend to have poorer overall performance because can’t take advantage of some HW features
But also a benefit because require no changes to host OS and run on a variety of general purpose OSs
User could have Type 2 hypervisor on an Apple laptop, run multiple guests, all on standard host OS such as Windows, Linux, MacOS

Types of VMs – Paravirtualization
Does not fit the definition of virtualization – VMM not presenting an exact duplication of underlying hardware
But still useful!
VMM provides services that guest must be modified to use
Leads to increased performance
Less needed as hardware support for VMs grows
Xen, leader in paravirtualized space, adds several techniques
including clean and simple device abstractions resulting in
Efficient I/O
Good communication between guest and VMM about device I/O

Xen I/O via Shared Circular Buffer

Types of VMs – Paravirtualization (cont.)
Xen, leader in paravirtualized space, adds several techniques (Cont.)
Memory management does not include nested page tables
Each guest has own read-only tables
Guest uses hypercall (call to hypervisor) when page-table changes needed
Paravirtualization allowed virtualization of older x86 CPUs (and others) without binary translation
Guest had to be modified to use run on paravirtualized VMM
But on modern CPUs Xen no longer requires guest modification -> no longer paravirtualization

Programming Environment Virtualization
Also not-really-virtualization but using same techniques, providing similar features
Programming language is designed to run within custom-built virtualized environment
For example Oracle Java has many features that depend on running in Java Virtual Machine (JVM)
In this case, virtualization is defined as providing APIs that define a set of features made available to a language and programs written in that language to provide an improved execution environment
JVM compiled to run on many systems (including some smart phones)
Programs written in Java run in the JVM no matter the underlying system
Similar to interpreted languages

Types of VMs – Emulation
Another (older) way for running one operating system on a different operating system
Virtualization requires underlying CPU to be same as guest was compiled for
Emulation allows guest to run on different CPU
Necessary to translate all guest instructions from guest CPU to native CPU
Emulation, not virtualization
Useful when host system has one architecture, guest compiled for other architecture
Company replacing outdated servers with new servers containing different CPU architecture, but still want to run old applications
Performance challenge – order of magnitude slower than native code
New machines faster than older machines so can reduce slowdown
Very popular – especially in gaming where old consoles emulated on new

Types of VMs – Application Containment
Some goals of virtualization are segregation of apps, performance and resource management, easy start, stop, move, and management of them
Can do those things without full-fledged virtualization
If applications compiled for the host operating system, don’t need full virtualization to meet these goals
Oracle containers / zones for example create virtual layer between OS and apps
Only one kernel running – host OS
OS and devices are virtualized, providing resources within zone with impression that they are only processes on system
Each zone has its own applications; networking stack, addresses, and ports; user accounts, etc
CPU and memory resources divided between zones
Zone can have its own scheduler to use those resources

videos

VMM vs Containers:

Solaris 10 with Two Zones

Virtualization and Operating-System Components

Now look at operating system aspects of virtualization
CPU scheduling, memory management, I/O, storage, and unique VM migration feature
How do VMMs schedule CPU use when guests believe they have dedicated CPUs?
How can memory management work when many guests require large amounts of memory?

OS Component – CPU Scheduling
Even single-CPU systems act like multiprocessor ones when virtualized
One or more virtual CPUs per guest
Generally VMM has one or more physical CPUs and number of threads to run on them
Guests configured with certain number of VCPUs
Can be adjusted throughout life of VM
When enough CPUs for all guests -> VMM can allocate dedicated CPUs, each guest much like native operating system managing its CPUs
Usually not enough CPUs -> CPU overcommitment
VMM can use standard scheduling algorithms to put threads on CPUs
Some add fairness aspect

OS Component – CPU Scheduling (cont.)
Cycle stealing by VMM and oversubscription of CPUs means guests don’t get CPU cycles they expect
Consider timesharing scheduler in a guest trying to schedule 100ms time slices -> each may take 100ms, 1 second, or longer
Poor response times for users of guest
Time-of-day clocks incorrect
Some VMMs provide application to run in each guest to fix time-of-day and provide other integration features

OS Component – Memory Management
Also suffers from oversubscription -> requires extra management efficiency from VMM
For example, VMware ESX guests have a configured amount of physical memory, then ESX uses 3 methods of memory management
Double-paging, in which the guest page table indicates a page is in a physical frame but the VMM moves some of those pages to backing store
Install a pseudo-device driver in each guest (it looks like a device driver to the guest kernel but really just adds kernel-mode code to the guest)
Balloon memory manager communicates with VMM and is told to allocate or deallocate memory to decrease or increase physical memory use of guest, causing guest OS to free or have more memory available
Deduplication by VMM determining if same page loaded more than once, memory mapping the same page into multiple guests

OS Component – I/O
Easier for VMMs to integrate with guests because I/O has lots of variation
Already somewhat segregated / flexible via device drivers
VMM can provide new devices and device drivers
But overall I/O is complicated for VMMs
Many short paths for I/O in standard OSes for improved performance
Less hypervisor needs to do for I/O for guests, the better
Possibilities include direct device access, DMA pass-through, direct interrupt delivery
Again, HW support needed for these
Networking also complex as VMM and guests all need network access
VMM can bridge guest to network (allowing direct access)
And / or provide network address translation (NAT)
NAT address local to machine on which guest is running, VMM provides address translation to guest to hide its address

OS Component – Storage Management
Both boot disk and general data access need be provided by VMM
Need to support potentially dozens of guests per VMM (so standard disk partitioning not sufficient)
Type 1 – storage guest root disks and config information within file system provided by VMM as a disk image
Type 2 – store as files in file system provided by host OS
Duplicate file -> create new guest
Move file to another system -> move guest
Physical-to-virtual (P-to-V) convert native disk blocks into VMM format
Virtual-to-physical (V-to-P) convert from virtual format to native or disk format
VMM also needs to provide access to network attached storage (just networking) and other disk images, disk partitions, disks, etc

OS Component – Live Migration
Taking advantage of VMM features leads to new functionality not found on general operating systems such as live migration
Running guest can be moved between systems, without interrupting user access to the guest or its apps
Very useful for resource management, maintenance downtime windows, etc
The source VMM establishes a connection with the target VMM
The target creates a new guest by creating a new VCPU, etc
The source sends all read-only guest memory pages to the target
The source sends all read-write pages to the target, marking them as clean
The source repeats step 4, as during that step some pages were probably modified by the guest and are now dirty
When cycle of steps 4 and 5 becomes very short, source VMM freezes guest, sends VCPU’s final state, sends other state details, sends final dirty pages, and tells target to start running the guest
Once target acknowledges that guest running, source terminates guest

Live Migration of Guest Between Servers

Disk space cannot be moved as part of a live migration. Disk access is maintained via remote access.

Examples – VMware

VMware Workstation runs on x86, provides VMM for guests
Runs as application on other native, installed host operating system -> Type 2
Lots of guests possible, including Windows, Linux, etc all runnable concurrently (as resources allow)
Virtualization layer abstracts underlying HW, providing guest with is own virtual CPUs, memory, disk drives, network interfaces, etc
Physical disks can be provided to guests, or virtual physical disks (just files within host file system)

VMware Workstation Architecture

Examples – Java Virtual Machine

Example of programming-environment virtualization
Very popular language / application environment invented by Sun Microsystems in 1995
Write once, run anywhere
Includes language specification (Java), API library, Java virtual machine (JVM)
Java objects specified by class construct, Java program is one or more objects
Each Java object compiled into architecture-neutral bytecode output (.class) which JVM class loader loads
JVM compiled per architecture, reads bytecode and executes
Includes garbage collection to reclaim memory no longer in use
Made faster by just-in-time (JIT) compiler that turns bytecodes into native code and caches them
Java on a chip is JVM implemented in hardware executing bytecode operations as native code.

The Java Virtual Machine

3

VMM on Windows 10
https://docs.microsoft.com/en-us/virtualization/hyper-v-on-windows/

Send email to CISC3320.bc@gmail.com
Within the next 90 seconds

End of Chapter 16

54

.MsftOfcThm_Accent1_Fill {
fill:#4472C4;
}
.MsftOfcThm_Accent1_Stroke {
stroke:#4472C4;
}

2_17.eps

(a)

processes

hardware

kernel

(b)

programming
interface

processes

processes

processes

kernelkernel kernel

VM2VM1 VM3

manager

hardware

virtual machine

fg16_02.eps

Privileged Instruction

Operating
System

VCPU

VMM

VMM

Guest

Kernel Mode

User Mode

Emulate Action

R
e
tu

rn

T
ra

p

Update

User Processes

16_03.eps

User Processes

Special Instruction

(VMM Reads Instructions)

Operating
System

VCPU

VMM

VMM

Guest

Kernel Mode

User Mode

Translate
Execute Translation

R
etu

rn

Update

fg16_04.eps

V
M

M
N

e
st

e
d
P

a
g
e
T

a
b
le

D
a
ta

S
tr

u
ct

u
re PML4E

PDPTE

PDE

PTE

Phy Addr

Host Physical Address

OffsetTableDirectoryDirectory PtrPML4

Guest Virtual Address

Kernel Paging Data
Structures

Guest Physical Address

G
u
e
st

1

2 3 4
5

1 1 2 2 3 3 4

54

16_4.pdf

Guest 1 Guest 2

CPUs
memory

CPUs
memory

Hypervisor (in firmware) I/O

CPUs
memory

CPUs
memory

Guest 3 Guest 4

Guest Guest Guest Guest Guest

fg16_06a.eps

Request Producer
Shared pointer
updated by guest OS

Request Consumer
Private pointer
in Xen

Response Producer
Shared pointer
updated by
Xen

Response Consumer
Private pointer
in guest OS

Request queue – Descriptors queued by the VM but not yet accepted by Xen

Outstanding descriptors – Descriptor slots awaiting a response from Xen

Response queue – Descriptors returned by Xen in response to serviced requests

Unused descriptors

2.18.eps

virtual platform
device management

zone 1

global zone

Solaris kernel

network addresses

zone 2

zone management

user programs
system programs
CPU resources

memory resources

user programs
system programs

network addresses
device access
CPU resources

memory resources

user programs
system programs

network addresses
device access
CPU resources

memory resources

device device…

fg16_08.eps

Guest Target running

5 – Send Dirty Pages (repeatedly)

4 – Send R/W Pages

3 – Send R/O Pages

1 – Establish0 – Running
Guest Source

V
M

M
S

o
u
rc

e

7 – Terminate
Guest Source

V
M

M
T

a
rg

e
t

2 – Create
Guest Target

6 – Running
Guest Target

fg16_07.eps

virtualization layer

host operating system
(Linux)

CPU memory

hardware

I/O devices

application application application application

guest operating
system

(free BSD)

virtual CPU
virtual memory
virtual devices

guest operating
system

(Windows NT)

virtual CPU
virtual memory
virtual devices

guest operating
system

(Windows XP)

virtual CPU
virtual memory
virtual devices

b32.eps

host system
(Windows, Linux, etc.)

class loader

Java
interpreter

Java program
.class files

Java API
.class files