Over a million developers have joined DZone.
{{announcement.body}}
{{announcement.title}}

Developing Real-Time Data Pipelines With Apache Kafka [Video]

DZone's Guide to

Developing Real-Time Data Pipelines With Apache Kafka [Video]

A presentation from SpringOne2GX on using Apache Kafka for transferring data in Spring.

· Big Data Zone
Free Resource

Learn best practices according to DataOps. Download the free O'Reilly eBook on building a modern Big Data platform.

Recorded at SpringOne2GX 2015.

Speaker: Joe Stein

Big Data Track

Slides: http://www.slideshare.net/SpringCentral/developing-realtime-data-pipelines-with-apache-kafka-53201942

Developing Real-Time Data Pipelines with Apache Kafka http://kafka.apache.org/ is an introduction for developers about why and how to use Apache Kafka. Apache Kafka is a publish-subscribe messaging system rethought of as a distributed commit log. Kafka is designed to allow a single cluster to serve as the central data backbone. A single Kafka broker can handle hundreds of megabytes of reads and writes per second from thousands of clients. It can be elastically and transparently expanded without downtime. Data streams are partitioned and spread over a cluster of machines to allow data streams larger than the capability of any single machine and to allow clusters of coordinated consumers. Messages are persisted on disk and replicated within the cluster to prevent data loss. Each broker can handle terabytes of messages. For the Spring user, Spring Integration Kafka and Spring XD provide integration with Apache Kafka.

Find the perfect platform for a scalable self-service model to manage Big Data workloads in the Cloud. Download the free O'Reilly eBook to learn more.

Topics:
java ,spring ,apache kafka

Published at DZone with permission of Pieter Humphrey, DZone MVB. See the original article here.

Opinions expressed by DZone contributors are their own.

{{ parent.title || parent.header.title}}

{{ parent.tldr }}

{{ parent.urlSource.name }}