Everything about Everyone, in one random access table

0

Posted by Lawrence Sinclair on 18 Sep 2009 at 16:27

HBASE - Sumit Khanna pointed out this element of the data processing space. I like to think about it as enabling one big table, big enough for every fact about everyone who ever lived.

HBase is the Hadoop database. Use it when you need random, realtime read/write access to your Big Data.

HBase ia an open-source, distributed, column-oriented store modeled after the Google paper,Bigtable: A Distributed Storage System for Structured Databy Chang et al. Just as Bigtable leverages the distributed data storage provided by the Google File System, HBase provides Bigtable-like capabilities on top of Hadoop. HBase's goal is the hosting of very large tables -- billions of rows X millions of columns -- atop clusters of commodity hardware

Data Processing Performance Options

1

Posted by Lawrence Sinclair on 14 Sep 2009 at 04:19

Here are a few of my thoughts on technologies and approaches for achieving better data processing performance in the current technology landscape.

RDBMS
Using mySQL or another RDBMS, performance might be addressed with better indexing or by partitioning the data.
MAP-REDUCE NON-RELATIONAL SYSTEMS
A non-relational approach might be to useHadoopor one of its distributions (such asCloudera). This would allow processing to be distributed anywhere from 3 local machines, or a virtually unlimited (hundreds+++) number of machines on the cloud (such as Amazon EC2). But this is best suited for analytic and data processing tasks that can takes several minutes or hours.
THE BEST OF BOTH WORLDS?!
Somewhere in between these two systems isHadoopDBbyDaniel Abadiof Yale. It uses the Hadoop...