Running Big Data Analytics on Bare Metal Cloud

DZone 's Guide to

Running Big Data Analytics on Bare Metal Cloud

Go bare metal.

· Big Data Zone ·
Free Resource


Big data analytics can be defined as a collection of technologies and processes used to crunch through complex data sets in order to discover market trends, correlations, and hidden data patterns. It enables organizations to make informed business decisions and aids researchers in verifying their scientific models.

Big Data Analytics has become an essential part of any business, be it financial analysis, retail, advertising, or healthcare. The amount of data worldwide has been growing exponentially and is estimated to jump from 33ZB this year to 175ZB in 2025. Not only does this create immense opportunities, but it also raises a huge demand for the infrastructure to run big data analytics on and poses new challenges for data engineers.

You may also like: OpenStack Stein: Kubernetes and Bare Metal.

What's Unique About an Analytics Workload?

Let's first agree on what defines analytics workload in the first place. According to Curt Monash, "Analytics is the antonym of transactional." While transactional processing (OLTP) is characterized by a short set of discrete operations with a high number of transactions per second and strict data integrity, analytics workloads are typically distinguished by fewer users making much more complex and resource-intensive queries to the data source.

There is massive parallelism going on behind the curtains, and data movement is lowered as much as possible by making computations as close to the data as possible. Data volume is large, the model is complex, and computation is done by a distributed system – all of which place a real burden on the infrastructure that carries out these tasks.Data to the cloud

Considering Transition to a Cloud?

It's no joke to build up and maintain your big data analytics stack, so companies often choose to migrate their analytics workloads to a cloud in order to reduce complexity and increase operational efficiency. As a rule of thumb, there are two main points to consider when preparing for a transition to a cloud — data storage and data processing.

Mountains of Data to Store

It is distributed data storage that you need to consider first and foremost for your big data project. With a reference to Brewer's theorem, it is impossible for a distributed data store to simultaneously provide more than two guarantees: whether that's consistency, availability, or partition tolerance. So, pick two, and you are good to go. The choice, as always, depends on your application.

To keep the big data wheel spinning, a highly scalable, efficient, and cost-effective storage is required. Almost always it is going to be some type of NoSQL database — nowadays you have more than 225 NoSQL databases to choose from.

Remember Brewer's theorem? This is when you start making sacrifices. If the risk of some data becoming unavailable is tolerable for you (sacrifice availability), a highly flexible and easily scalable document database with straightforward querying, like MongoDB, may be your way to go. Whether it's no big deal if your clients may read inconsistent data (sacrifice consistency), you may want to choose a fault-tolerant and linearly scalable database like Cassandra.

There are even some niche use cases when you may consider using a traditional relational database management system like MySQL or PostgreSQL and sacrifice partition tolerance. Although this may validate your hipster identity, it would probably involve database sharding and make working with unstructured data nearly impossible. Let's just leave SQL for querying data warehouses, shall we?

Regardless of what database you choose, most of them run really well on commodity hardware. Although all hyperscale cloud providers are offering managed database services these days — and some of them do not feel shy to give open source the middle finger — there's no necessity to get locked in their ecosystem when there are superior open source products out there.

For instance, you can run a MongoDB cluster on a bare metal cloud with HDD, SSD, or NVMe directly attached storage to skyrocket your I/O operations on each node. And if you are a true speed fan, setting up an in-memory database like Ignite or Redis might be your thing to do.

Let's Process Data! Wait, but How?

Data is the new oil and not without a reason. We love data; it helps us understand things better and reveal actionable insights. In order to do so, we have to process our data one way or another.

First, there was Hadoop, with its batch processing compute framework based on the MapReduce computing paradigm. Life was good, and songs were sung while engineers scaled their big data clusters horizontally and employed massive parallelism. Each node executed the given reduce functions on the mapped data it has been assigned with – this way, enormous data chunks were processed like a breeze.

This was also how Google started its search engine. In time, the Hadoop ecosystem expanded rapidly and introduced additional layers of abstraction to address new issues, as the big data industry became more mature. It is still the most prominent and used tool in the data industry today, in which you can run smoothly on simple commodity servers. Just make sure you have fast, directly attached storage on your nodes since Hadoop MapReduce is disk-bound.

While batch processing is a really powerful concept, we first need to store the data for it to be processed. This creates difficulties when you want to start making real-time predictions with continuous data streaming in. For something like algorithmic stock trading or wildfire monitoring to work, your data has to be processed in a glimpse. Obviously, we need a different paradigm here, and Apache Spark is at the forefront of innovation when it comes to stream processing.

The project was first intended to address Hadoop weaknesses in stream analytics. Spark has no file management system, so it relies on HDFS or any other storage cluster. It reads data from the cluster, performs its operations in a single step, and then writes the data back to the cluster. This can be 100x faster than Hadoop, since Spark operates in memory by default. When choosing the right infrastructure for your Spark cluster, look for something powerful RAM-wise.

Like most great technologies, Spark has evolved and changed a lot. It is now a unified analytics engine with powerful interactive queries, graph processing, and iterative algorithms. For instance, you can easily build machine learning workflows and employ some of the most popular algorithms on Spark to iterate over your data set and build machine learning models. It can even process batch jobs these days. And the best thing about Spark? It's completely free of charge.

To Cloud or Not to Cloud

Okay, you have all these powerful open source tools in your pocket that we have just discussed. Now, you need to choose the right infrastructure for your project. If you bought bare metal servers and hosted them on-premises, you would be able to squeeze out most of the benefits of raw infrastructure, but with a huge upfront capital investment and further maintenance costs.

Although this option is still considerable for large enterprises, small and medium businesses must be much more agile. Renting infrastructure, on the other hand, is a more convenient option, since you pay per usage and need not invest into hardware. Let's say you have eventually decided to move to a cloud. But which one to choose?

The Big Boys

Every hyperscale provider, be it AWS, Azure or GCP, has a wide portfolio of managed services to offer for big data community, ranging from managed databases to integrated machine learning frameworks. It might seem like a one-size-fits-all solution, but most of their managed services have roughly the same functionality as their open source counterparts that are available free of charge. If you still want to abstract from the infrastructure entirely and are fine with being locked-in at a single provider, get ready to receiving ever-increasing invoices and using complex pricing calculators. Funny enough, there's even a role of Cloud Economist to help you to solve your AWS invoice.

Alright, so you want to keep full control over your cloud stack and always have the freedom to choose where it resides. As mentioned earlier, open source technologies like Hadoop and Spark work really well on commodity hardware, so the main question is which infrastructure as a service provider to choose. Historically, a typical cloud offering included virtual machines with overbooked hardware resources that often resulted in fluctuating workloads and increased security risks. Although cloud service ecosystem expanded greatly, the underlying infrastructure services still rely heavily on the hypervisor.

Bare Metal Cloud on the Block

Bare metal cloud is different. You still have fully automated infrastructure provisioning, just with no underlying virtualization layer. This is great for several reasons. First, all servers are strictly single-tenant and you're the only owner of the entire machine. Being single is not much fun in life, but it's great in the cloud: you have no noisy neighbors, no hardware overbooking, no hypervisor overhead and less security risks. Simply put, bare metal cloud is a much cleaner way to host your resource-intensive applications. As for data analytics workloads that require robust infrastructure and enhanced security, bare metal cloud is unbeatable.

Raw Horsepower of Bare Metal

Running your big data cluster on bare metal machines gives you an extra edge. Servers can be easily scaled up and down in minutes via RESTful API, which is critical when running a distributed system. With no virtualization and hardware overbooking you can run your application at maximum capacity and still have smooth and steady workloads. If that's not enough, you can customize server hardware as you like. Simply add a GPU accelerator when building your machine learning model, increase RAM to scale your in-memory database or put NVMe storage into your servers to skyrocket your Hadoop cluster. Where else can you get specialized hardware that easy? By eliminating hypervisor overhead and introducing custom hardware bare metal cloud gives you the most efficient infrastructure on demand. Raw and simple.

Enhanced Privacy and Security

Security is treated very seriously in the big data world and it's understandable that you need to choose your infrastructure accordingly. Dealing with sensitive data often means that you have to store and process personal identifiable information (PII) and stay in line with legal regulation. Making sure your vendor is GDPR compliant or have industry recognized certifications like ISO 27001 is always a good idea. In addition to that, you may also be legally obliged to have a private and isolated infrastructure. Bare metal cloud is inherently single-tenant and you don't pay a cent for that.

Legal compliance is important, yet system security is no less crucial. On a big data distributed computing cluster data should be moved between your nodes privately. To make this happen you typically need a private network subnet. On bare metal cloud a private network interface is assigned automatically for every server. This way you can process your data internally on a fast and secure LAN with 10G bandwidth.

100 Times Cheaper Data Transfer

Sure, you try to keep your computations as close as possible to where the data resides. This is especially true when your data set is huge, since moving computation is cheaper than moving data. Nevertheless, you still need to move your data in and out of the cluster. While hyperscale providers boast about their low data transfer prices which may vary from $50 to more than $100 per terabyte, on bare metal cloud you can transfer data in and out to the Internet for as low as $1 per terabyte. Quite a different, right?

It's Up to You Now

There are still only a handful of bare metal cloud providers in the market and even less those that can offer fully automated AND easily customizable infrastructure.  So if you are about to conquer the world with the next big thing, consider choosing a more flexible, efficient and price-effective cloud platform to level up your big data application.

Further Reading

big data ,analytics ,cloud (add topic) ,cloud computing

Published at DZone with permission of Marius Rimkus . See the original article here.

Opinions expressed by DZone contributors are their own.

{{ parent.title || parent.header.title}}

{{ parent.tldr }}

{{ parent.urlSource.name }}