Systems | Development | Analytics | API | Testing

Confluent

CDC and Data Streaming: Capture Database Changes in Real Time with Debezium PostgreSQL Connector

In today's data-driven world, staying ahead means acting on the most up-to-date information. That's where change data capture (CDC) comes in. CDC is a design pattern that tracks your database tables, capturing every row-level insert, update, and delete as it happens. This real-time monitoring allows downstream systems to react to changes instantly, without batch-based updates and resource-intensive full scans.

How to Visualize Real-Time Data from Apache Kafka using Apache Flink SQL and Streamlit

Data visualization is cool, but have you tried setting up a chart of real-time data? In this video, Lucia Cerchie shows you how to create a live visualization of market data. She starts by producing data to a topic in Confluent Cloud from an Alpaca API websocket, then processes that data with Flink SQL, and finally uses a Streamlit component for a real-time visualization.

Product Management in the Dynamic World of Data Streaming

A year in at Confluent, Product Manager Surabhi Singh has learned a lot about data streaming—and even more about herself. In this fast-paced environment, Surabhi is highly motivated and committed to her work strategically planning, coordinating, and delivering product improvements for customers whose business operations depend on Confluent Platform.

What Made Current 2024 Unforgettable? Hear From Our Attendees | Current 2024

In this recap video from Current 2024, attendees share their favorite moments from the event. From insightful talks on data streaming innovation to hands-on workshops and networking opportunities, hear what participants found most valuable.

Shift Left: Headless Data Architecture, Part 2

The headless data architecture is the formalization of a data access layer at the center of your organization. Encompassing both streams and tables, it provides consistent data access for both operational and analytical use cases. Streams provide low-latency capabilities to enable timely reactions to events, while tables provide higher-latency but extremely batch-efficient querying capabilities. You simply choose the most relevant processing head for your requirements and plug it into the data.