Chapter 13: Problem 15
Discuss the techniques for allocating file blocks on disk.
Short Answer
Step by step solution
Key Concepts
These are the key concepts you need to understand to accurately answer the question.
/*! This file is auto-generated */ .wp-block-button__link{color:#fff;background-color:#32373c;border-radius:9999px;box-shadow:none;text-decoration:none;padding:calc(.667em + 2px) calc(1.333em + 2px);font-size:1.125em}.wp-block-file__button{background:#32373c;color:#fff;text-decoration:none}
Learning Materials
Features
Discover
Chapter 13: Problem 15
Discuss the techniques for allocating file blocks on disk.
These are the key concepts you need to understand to accurately answer the question.
All the tools & learning materials you need for study success - in one app.
Get started for free
Suppose that we have an ordered file of fixed-length records and an unordered overflow file to handle insertion. Both files use unspanned records. Outline algorithms for insertion, deletion, and modification of a file record and for reorganizing the file. State any assumptions you make.
What are the techniques used to improve performance of disks in RAID?
Suppose we have a sequential (ordered) file of 100,000 records where each record is 240 bytes. Assume that \(B=2400\) bytes, \(s=16 \mathrm{ms}, r d=8.3 \mathrm{ms},\) and \(b t t=0.8 \mathrm{ms}\) Suppose we want to make \(X\) independent random record reads from the file. We could make \(X\) random block reads or we could perform one exhaustive read of the entire file looking for those \(X\) records. The question is to decide when it would be more efficient to perform one exhaustive read of the entire file than to perform \(x\) individual random reads. That is, what is the value for \(X\) when an exhaustive read of the file is more efficient than random \(X\) reads? Develop this as a function of \(X\)
Suppose that we have a hash file of fixed-length records, and suppose that overflow is handled by chaining. Outline algorithms for insertion, deletion, and modification of a file record. State any assumptions you make.
A PARTS file with Part* as hash key includes records with the following Part* val. ues: 2369,3760,4692,4871,5659,1821,1074,7115,1620,2428,3943,4750 \(6975,4981,9208 .\) The file uses eight buckets, numbered 0 to 7. Each bucket is one disk block and holds two records. Load these records into the file in the given order, using the hash function \(h(K)=K \bmod 8 .\) Calculate the average number of block accesses for a random retrieval on Part#.
What do you think about this solution?
We value your feedback to improve our textbook solutions.