Over a million developers have joined DZone.
{{announcement.body}}
{{announcement.title}}

Building LinkedIn's Real-Time Data Pipeline

DZone's Guide to

Building LinkedIn's Real-Time Data Pipeline

· Integration Zone ·
Free Resource

SnapLogic is the leading self-service enterprise-grade integration platform. Download the 2018 GartnerMagic Quadrant for Enterprise iPaaS or play around on the platform, risk free, for 30 days.


At the core of many of LinkedIn's analytics applications is a real-time data pipeline built on top of Apache Kafka. This system handles over 10 billion messages writes per day for thousands of production processes. This talk will cover some of the challenges of building and scaling this data pipeline for log data, system metrics, and other high-volume data streams. It will also cover some details of the design of Kafka, as well as some of the particular requirements of Hadoop data loads and real-time processing applications.

About Jay Kreps
Jay is the technical lead for LinkedIn's data team, which is responsible for the site's core data technologies including storage systems, data pipelines, Hadoop, search, social graph, and recommendation systems. He is an original author on several open source projects including Apache Kafka, a real-time distributed messaging system, and Project Voldemort a distributed key-value store. He has a Masters degree in computer science from UC Santa Cruz where he studied machine learning.

With SnapLogic’s integration platform you can save millions of dollars, increase integrator productivity by 5X, and reduce integration time to value by 90%. Sign up for our risk-free 30-day trial!

Topics:

Opinions expressed by DZone contributors are their own.

{{ parent.title || parent.header.title}}

{{ parent.tldr }}

{{ parent.urlSource.name }}