Over a million developers have joined DZone.
{{announcement.body}}
{{announcement.title}}

Using Hadoop for Big Data Acceleration

DZone's Guide to

Using Hadoop for Big Data Acceleration

· Database Zone
Free Resource

MongoDB Atlas is a database as a service that makes it easy to deploy, manage, and scale MongoDB. So you can focus on innovation, not operations. Brought to you in partnership with MongoDB.

In this video, D.K. Panda from Ohio State University presents: Accelerating Big Data with Hadoop (HDFS, MapReduce and HBase) and Memcached.

"The SuperMUC has 147,456 cores and a peak performance of about 3 petaflop/s. The main memory will be 288 terabytes together with 12 petabytes hard disk space based on the GPFS file system. The system will use 18,432 Intel Xeon Sandy Bridge-EP processors running in IBM System x iDataPlex servers. It will also use a new form of cooling that IBM developed, called Aquasar, that uses hot water to cool the processors, a design that should cut cooling electricity usage by 40 percent, IBM claims."

Recorded at the 2013 HPC Advisory Council Switzerland Conference.

Learn more at: http://www.hpcadvisorycouncil.com/eve...


MongoDB Atlas is the best way to run MongoDB on AWS — highly secure by default, highly available, and fully elastic. Get started free. Brought to you in partnership with MongoDB.

Topics:

Opinions expressed by DZone contributors are their own.

THE DZONE NEWSLETTER

Dev Resources & Solutions Straight to Your Inbox

Thanks for subscribing!

Awesome! Check your inbox to verify your email so you can start receiving the latest in tech news and resources.

X

{{ parent.title || parent.header.title}}

{{ parent.tldr }}

{{ parent.urlSource.name }}