Hadoopa – coolinarism
post-template-default,single,single-post,postid-13796,single-format-standard,eltd-core-1.1.1,symple-shortcodes ,symple-shortcodes-responsive,flow-ver-1.3.7,,eltd-smooth-page-transitions,ajax,eltd-blog-installed,page-template-blog-standard,eltd-header-standard,eltd-fixed-on-scroll,eltd-default-mobile-header,eltd-sticky-up-mobile-header,eltd-dropdown-default,wpb-js-composer js-comp-ver-5.2.1,vc_responsive



Apache Hadoop™ was born out of a need to process an avalanche of big data. The web was generating more and more information on a daily basis, and it was becoming very difficult to index over one billion pages of content. In order to cope, Google invented a new style of data processing known as MapReduce. A year after Google published a white paper describing the MapReduce framework, Hadoop was created to apply these concepts to an open-source software framework to support distribution for the Nutch search engine project. A scalable fault-tolerant distributed system for data storage and processing:

  • Flexibility – Store any data, Run any analysis
  • Scalability – Start at 1TB/3-nodes grow to petabytes/1000s of nodes
  • Economics – Cost per TB at a fraction of traditional options

image: wikimedia

t-penned by netize R
No Comments

Sorry, the comment form is closed at this time.