Apache Kafka is used for building real-time data pipelines and streaming apps. It is horizontally scalable, fault-tolerant, fast, and runs in production in thousands of companies. The Kafka cluster stores streams of records in categories called topics. Each record consists of a key, a value, and a timestamp.
Kafka has four core APIs:
– The Producer API allows an application to publish a stream records to one or more Kafka topics.
– The Consumer API allows an application to subscribe to one or more topics and process the stream of records produced to them.
– The Streams API allows an application to act as a stream processor, consuming an input stream from one or more topics and producing an output stream to one or more output topics, effectively transforming the input streams to output streams.
-The Connector API allows building and running reusable producers or consumers that connect Kafka topics to existing applications or data systems. For example, a connector to a relational database might capture every change to a table.
Behaim draws upon a decade of real-time integration and API experience for Kafka’s installation, setup, configuration, and production deployment. The team is also knowledgeable with implementing Kafka producers, consumers, streams and connectors.