Wednesday, December 06, 2006

FrontPage - Lucene-hadoop Wiki

FrontPage - Lucene-hadoop Wiki: "Hadoop DFS

Hadoop's Distributed File System is designed to reliably store very large files across machines in a large cluster. It is inspired by the [WWW] Google File System. Hadoop DFS stores each file as a sequence of blocks, all blocks in a file except the last block are the same size. Blocks belonging to a file are replicated for fault tolerance. The block size and replication factor are configurable per file. Files in HDFS are 'write once' and have strictly one writer at any time."

No comments: