| Sophie Blee-Goldman (Apache Kafka® Committer and Software Engineer, Confluent) provides a full introduction to Kafka Tables, also known as KTables. A KTable is an abstraction of a changelog stream, where each record in the stream is an update on the primary-keyed table. Whereas event streams, which are sequences of key-value pairs that are independent of each other. In contrast, an update stream is also a sequence of key-value pairs, but newly arriving records replace existing ones with the same key. A KTable is an update stream. Like a KStream, you build a KTable with a StreamBuilder, but you use builder.table instead of builder.stream. But unlike a KStream, a KTable can only subscribe to a single topic at once. Since a KTable needs to store its latest value somewhere, it uses a state store. Updates to a KTable may be buffered in a cache, and only when it gets flushed do the updates proceed further down the processor topology. Like a KStream, you can use mapping and filtering operations with a KTable. KTable has a special variant, a GlobalKTable, which can access all of a topic’s partitions at once, unlike a regular KTable, which typically just accesses one.
Use the promo code STREAMS101 to get $101 of free Confluent Cloud usage:
Promo code details:
► Streams and Tables in Apache Kafka: A Primer:
► Interface KTable:
Confluent is pioneering a fundamentally new category of data infrastructure focused on data in motion. Confluent’s cloud-native offering is the foundational platform for data in motion – designed to be the intelligent connective tissue enabling real-time data, from multiple sources, to constantly stream across the organization. With Confluent, organizations can meet the new business imperative of delivering rich, digital front-end customer experiences and transitioning to sophisticated, real-time, software-driven backend operations. To learn more, please visit www.confluent.io.
#kafka #kafkastreams #streamprocessing #apachekafka #confluent