Systems | Development | Analytics | API | Testing

Confluent

Why Short-Lived Connections Are Killing Your Performance! | Kafka Developer Mistakes

Constantly starting and stopping Apache Kafka producers and consumers? That’s a recipe for high resource usage and inefficiency. Short-lived connections are heavy on resources, and can slow down your whole cluster. Keep them running to boost performance, cut latency, and get the most out of your Kafka setup.

Securely Query Confluent Cloud from Amazon Redshift with mTLS

Querying databases comes with costs—wall clock time, CPU usage, memory consumption, and potentially actual dollars. As your application scales, optimizing these costs becomes crucial. Materialized views offer a powerful solution by creating a pre-computed, optimized data representation. Imagine a retail scenario with separate customer and product tables. Typically, retrieving product details for a customer's purchase requires cross-referencing both tables.

Why Relying on Default Settings Can Cost You! | Kafka Developer Mistakes

Default settings in Apache Kafka work when you’re getting started, but aren't suited for production. Sticking with defaults, like a seven-day retention policy, or a replication factor of one, can cause storage issues, or data loss in case of failure. Learn why optimizing retention periods, replication factors, and partitions, is crucial for better Kafka performance and reliability.

Why Using Outdated Versions Hurts Your System! | Kafka Client Mistakes

Keeping your Apache Kafka clients up-to-date is critical for maximizing performance, security, and stability. In this video, we discuss why sticking with old versions could be putting you at risk, since it means you’re missing out on dozens of new features, and hundreds of bug fixes and security patches. Learn why upgrading is more than just a “nice-to-have”—it’s essential for a smoother and safer Kafka experience.

Confluent Introduces Enterprise Data Streaming to MongoDB's AI Applications Program (MAAP)

Today, Confluent, the data streaming pioneer, is excited to announce its entrance into MongoDB’s new AI Applications Program (MAAP). MAAP is designed to help organizations rapidly build and deploy modern generative AI (GenAI) applications at enterprise scale.

Deep Dive into Handling Consumer Fetch Requests: Kafka Producer and Consumer Internals, Part 4

Recap: This is the last part of our four chapters: It’s been a long time coming, but we’ve finally arrived at the fourth and final installment of our blog series. In this series, we’ve been peeling back the layers of Apache Kafka to get a deeper understanding of how best to interact with the cluster using producer and consumer clients. At a high level, a fetch request is comprised of two parts: Let’s dive in.

How to Set Up Networking on Confluent Cloud

Setting up network connections can often seem difficult or time consuming. This video provides a wayfinding introduction to help you get networking up and running for all cluster types on Confluent Cloud, showing you your networking options for each cluster type when running on AWS, Azure, or Google Cloud, respectively.

Connect with Confluent Q4 Update: New Program Entrants and SAP Datasphere Hydration

The Connect with Confluent (CwC) Technology Partner Program consistently expands the reach of Confluent’s data streaming platform across an ever-growing landscape of enterprise data systems. In this blog, you’ll meet the latest program entrants who have built fully managed integrations with Confluent and discover new ways to leverage real-time data across your business.

How to migrate from Kafka to Confluent Cloud with limited downtime

In this short video, a Confluent Solutions Engineering will run through the high-level steps on how to get started with your migration. And even better, once you’re done watching, you can download our comprehensive migration kit for a step by step guide of everything I’ve talked about and more.

Seamlessly Connect IoT Data Streams: Integrating Confluent Cloud with AWS IoT Core

Raw data from IoT devices, like GPS trackers or electronic logging devices (ELDs), often lacks meaning on its own. However, if combined with information from other business systems, such as inventory management or customer relationship management (CRM), this data can now provide a richer, more complete picture for more effective decision-making. For example, combining GPS data with inventory levels can optimize logistics and delivery routes.