Mass-Storage Systems 

 

 

Overview of Mass Storage Structure

 

 

 

 

Magnetic disks provide bulk of secondary storage of modern computers:

 

§  Drives rotate at 60 to 200 times per second

§  Transfer rate is the rate at which data flow between drive and computer

§  Positioning time (random-access time) is time to move disk arm to desired cylinder (seek time)

and time for desired sector to rotate under the disk head (rotational latency)

§  Head crash results from disk head making contact with the disk surface

§  Disks can be removable

§  Disk drives are addressed as large 1-dimensional arrays of logical blocks,

where the logical block is the smallest unit of transfer

§  The 1-dimensional array of logical blocks is mapped into the sectors of the disk sequentially.

Sector 0 is the first sector of the first track on the outermost cylinder.

Mapping proceeds in order through that track, then the rest of the tracks in that cylinder and

then through the rest of the cylinders from outermost to innermost.

 

Disk Scheduling Algorithms

 

The operating system is responsible for using hardware efficiently.

For the disk drives, this means having a fast access time & disk bandwidth.

 

ü Access time has two major components:

Seek time is the time for the disk to move the heads to the cylinder containing the desired sector

Rotational latency time waiting for the disk to rotate the desired sector to the disk head

 

We like to minimize seek time.

 

ü Disk bandwidth is the total number of bytes transferred divided by

the total time between the first request for service and the completion of the last transfer.

 

Several algorithms exist to schedule the servicing of disk I/O requests.

We illustrate them with a Request Queue (cylinder range 0-199):

                  
          98, 183, 37, 122, 14, 124, 65, 67

 

Head pointer:  cylinder 53

 

êFCFS

 

Illustration shows total head movement of 640 cylinders

 

 

 

 

êSSTF

 

Selects the request with the minimum seek time from the current head position

SSTF scheduling may cause starvation of some requests

 

Illustration shows total head movement of 236 cylinders

 

 

 

êSCAN

 

The disk arm starts at one end of the disk, and moves toward the other end,

servicing requests until it gets to the other end of the disk,

where the head movement is reversed and servicing continues.

 

SCAN algorithm sometimes called the elevator algorithm.

 

Illustration shows total head movement of 208 cylinders

 

 

 

êC-SCAN

 

Provides a more uniform wait time than SCAN

ü The head moves from one end of the disk to the other, servicing requests as it goes

ü When it reaches the other end, however, it immediately returns to the beginning of the disk,

without servicing any requests on the return trip.

 

Treats the cylinders as a Circular list that wraps around from the last cylinder to the first one

 

 

êC-LOOK

 

ü Version of C-SCAN

ü Arm only goes as far as the last request in each direction,

then reverses direction immediately, without first going all the way to the end of the disk.

 

 

 

 

Selecting a Disk-Scheduling Algorithm

 

ü SSTF is common and has a natural appeal

ü SCAN and C-SCAN perform better for systems that place a heavy load on the disk

 

 

Network-Attached Storage

 

·       Network-attached storage (NAS) is storage made available over a network

rather than over a local connection (such as a bus)

·       Implemented via remote procedure calls (RPCs) between host and storage

 

 

 

 

Storage Area Network

 

 

 

RAID  (Redundent Arrays of Independent Disks) Structure

 

·       multiple disk drives provides reliability via redundancy

Increases the mean time to failure

·       improve performance and reliability of the storage system by storing redundant data

·       Mirroring or shadowing  keeps duplicate of each disk

·       small number of hot-spare disks are left unallocated,

automatically replacing a failed disk and having data rebuilt onto them

 

 

Stable-Storage Implementation

 

To implement stable storage:

ü Replicate information on more than one nonvolatile storage media with independent failure modes

ü Update information in a controlled manner to ensure that we can recover the stable data after any failure during data transfer or recovery.

 

Tertiary Storage Devices

 

·       Low cost is the defining characteristic of tertiary storage

·       Generally, tertiary storage is built using removable media

·       Common examples of removable media are floppy disks and CD-ROMs; other types are available

 

Removable Disks

 

·       Floppy disk — thin flexible disk coated with magnetic material, enclosed in a protective plastic case

Most floppies hold about 1 MB;

·       Removable magnetic disks can be nearly as fast as hard disks,

but they are at a greater risk of damage from exposure

·       Magneto-optic disk records data on a rigid platter coated with magnetic material

The magneto-optic head flies much farther from the disk surface than a magnetic disk head,

and the magnetic material is covered with a protective layer of plastic or glass; resistant to head crashes

·       Optical disks do not use magnetism; they employ special materials that are altered by laser light

 

 

WORM Disks

 

·       The data on read-write disks can be modified over and over

·       WORM (“Write Once, Read Many Times”) disks can be written only once

·       Thin aluminum film sandwiched between two glass or plastic platters

·       To write a bit, the drive uses a laser light to burn a small hole through the aluminum;

information can be destroyed by not altered

·       Very durable and reliable

·       Read-only disks, such as CD-ROM and DVD, come from the factory with the data pre-recorded

 

 

Magnetic tape

 

·       Was early secondary-storage medium

·       Relatively permanent and holds large quantities of data

·       Access time slow

·       Random access ~1000 times slower than disk

·       Mainly used for backup, storage of infrequently-used data, transfer medium between systems

·       Once data under head, transfer rates comparable to disk

·       20-200GB typical storage

·       Compared to a disk, a tape is less expensive and holds more data, but random access is much slower

·       Large tape installations typically use robotic tape changers that move tapes between tape drives and storage slots in a tape library

 

 

Tape Drives Application Interface

 

·       The basic operations for a tape drive differ from those of a disk drive

·       locate() positions the tape to a specific logical block,

·       The read_position() returns the logical block number where the tape head is

·       The space() operation enables relative motion

·       Tape drives are “append-only” devices; updating a block in the middle of the tape also effectively erases everything beyond that block

·       An EOT mark is placed after a block that is written

 

 

Hierarchical Storage Management (HSM)

 

·       A hierarchical storage system extends the storage hierarchy beyond primary memory and secondary storage to incorporate tertiary storage — usually implemented as a jukebox of tapes or removable disks

·       Small and frequently used files remain on disk

·       Large, old, inactive files are archived to the jukebox

·       HSM are found in supercomputing centers that have enormous volumes of data.

 

Speed

 

·       Two aspects of speed in tertiary storage are bandwidth and latency

Sustained bandwidth – average data rate during a large transfer;

Effective bandwidth – average over the entire I/O time, including seek() or locate(), and cartridge switching

Access latency – amount of time needed to locate data

·       Access time for a disk – move the arm to the selected cylinder and wait for the rotational latency;

< 35 milliseconds

·       Access on tape requires winding the tape reels until the selected block reaches the tape head;

tens or hundreds of seconds

·       Generally random access within a tape cartridge is about a thousand times slower than random access on disk

 

Reliability

 

·       A fixed disk drive is likely to be more reliable than a removable disk or tape drive

·       An optical cartridge is likely to be more reliable than a magnetic disk or tape

·       A head crash in a fixed hard disk generally destroys the data,

whereas the failure of a tape drive or optical disk drive often leaves the data cartridge unharmed

 

 

Cost

 

·       Main memory is much more expensive than disk storage

·       The cost per megabyte of hard disk storage is competitive with magnetic tape.