Big Data/BI Zone is brought to you in partnership with:

Eric is living in Chapel Hill, NC. By night, he writes and edits science fiction. On weekends, he spends too much time making plumbers hop on things. Eric has posted 249 posts at DZone. You can read more from them at their website. View Full User Profile

Here's How to Build an Optimal Hadoop Cluster

01.03.2013
| 3809 views |
  • submit to reddit

If you're ringing in the New Year by building a Hadoop cluster, then you might want to take a look at Atlantbh's detailed tutorial:

Amount of data stored in database/files is growing every day, using this fact there become a need to build cheaper, mainatenable and scalable environments capable of storing  big amounts of data („Big Data“). Conventional RDBMS systems became too expensive and not scalable based on today’s needs, so it is time to use/develop new techinques that will be able to satisfy our needs.
One of the technologies that lead in these directions is Cloud computing. There are different implementation of Cloud computing but we selected Hadoop – MapReduce framework with Apache licence based on Google Map Reduce framework.
In this document I will try to explain how to build scalable Hadoop cluster where it is possible to store, index, search and maintain practically unlimited ammounts of data.
This article will cover installation and configuration steps divided into these sections:
  • Network architecture
  • Operating System
  • Hardware requirements
  • Hadoop software installation/setup

You can read the complete tutorial at Atlantbh's blog.

Published at DZone with permission of its author, Eric Gregory.

(Note: Opinions expressed in this article and its replies are the opinions of their respective authors and not those of DZone, Inc.)

Comments

Mladen Antunovic replied on Mon, 2013/01/07 - 4:38am

Thank You for sharing article

Comment viewing options

Select your preferred way to display the comments and click "Save settings" to activate your changes.