Design and implementation of the Chunks feature

dc.contributor.authorNory, Nawar A.
dc.date.accessioned2008-08-12T13:08:31Z
dc.date.available2008-08-12T13:08:31Z
dc.date.graduationmonthAugust
dc.date.issued2008-08-12T13:08:31Z
dc.date.published2008
dc.description.abstractThe recovery-driven design of the file system has been one of the most challenging fields over the major trends in operating systems. This field has assumed considerable importance in the past decades as the disk sizes have been increasing without a comparable increase in the disk I/O bandwidth and seek time. The rapid increase in the storage size is expected to become constant in the future due to the growing market demand and the continuous database size increment of many companies and major businesses. Due to the same reason, the cost of the average file system checking time has increased without a significant improvement in the disk I/O bandwidth and seek time performance. Operating system bugs, power outages, and hardware failures which result in a file system crash were the main reasons behind the innovation of novel recovery approaches such as Journaling and soft-updates. Although such approaches avoided complete file system checking by checking solely inconsistencies in file system metadata, it become inevitable for them to check the entire file system for inconsistencies because of the previously entioned types of problems. One of the emerging recovery-driven designs which considers minimizing file system checking cost is the Chunkfs files ystem. Chunkfs file system introduces an innovative look into the file system design by dividing the file system layout into smaller chunks, each one of which represents a smaller scale file system by itself. In our work we probed an alternative recovery-driven design which is considerably inspired by the Chunkfs concepts and follows the same design guidelines. This recovery-riven design is introduced by adding a new feature to the file system which best utilizes the existing underlying design through considering the block groups as individual chunks, confining their files and directories spanning across different block groups by means of special controlled continuation links. These links provide a fault isolation means by circumscribing the checking of the file system to only these block groups which appear to be dirty after a crash, a method resulting in a moderate reduction in file system checking cost. We also probed different metrics of metadata sizes, and the probable cost of files and directories expansion across the different block groups.
dc.description.advisorDaniel A. Andresen
dc.description.degreeMaster of Science
dc.description.departmentDepartment of Computing and Information Sciences
dc.description.levelMasters
dc.identifier.urihttp://hdl.handle.net/2097/911
dc.language.isoen_US
dc.publisherKansas State University
dc.rights© the author. This Item is protected by copyright and/or related rights. You are free to use this Item in any way that is permitted by the copyright and related rights legislation that applies to your use. For other uses you need to obtain permission from the rights-holder(s).
dc.rights.urihttp://rightsstatements.org/vocab/InC/1.0/
dc.subjectFile system
dc.subject.umiComputer Science (0984)
dc.titleDesign and implementation of the Chunks feature
dc.typeThesis

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
NawarNory2008.pdf
Size:
544.43 KB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.68 KB
Format:
Item-specific license agreed upon to submission
Description: