FrontPage - Lucene-hadoop Wiki: "Hadoop DFS
Hadoop's Distributed File System is designed to reliably store very large files across machines in a large cluster. It is inspired by the [WWW] Google File System. Hadoop DFS stores each file as a sequence of blocks, all blocks in a file except the last block are the same size. Blocks belonging to a file are replicated for fault tolerance. The block size and replication factor are configurable per file. Files in HDFS are 'write once' and have strictly one writer at any time."
Wednesday, December 06, 2006
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment