|
By Confluent
Confluent Manager for Apache Flink provides a cloud-like management experience for on-premises Kubernetes environments.
|
By Yashwanth Dasari
At Confluent, we’re committed to building the world's leading data streaming platform that gives you the ability to stream, connect, process, and govern all your data, and makes it available wherever it’s needed, however it’s needed, in real time. Today, we're excited to announce the release of Confluent Platform 7.8. This release builds upon Apache Kafka 3.8, reinforcing our core capabilities as a data streaming platform.
|
By Will Zhu
At Confluent, we continuously strive to showcase the power of our data streaming platform through real-world applications, exemplified by our Customer Zero initiative. In part 1 of this blog, we present the latest use case of Customer Zero that harnesses the capabilities of generative AI, data streaming, and real-time predictions to enhance lead scoring for sales, helping our team prioritize high-value prospects and address complex challenges within our organization.
|
By Kasun Indrasiri
Earlier this year, we unveiled our vision for Tableflow to feed Apache Kafka streaming data into data lakes, warehouses, or analytical engines with a single click. Since then, many customers have been exploring, experimenting with, and providing valuable feedback on Tableflow Early Access. Our teams have worked tirelessly to incorporate this feedback and are excited to bring Tableflow Open Preview to you in the near future.
|
By Braeden Quirante
Querying databases comes with costs—wall clock time, CPU usage, memory consumption, and potentially actual dollars. As your application scales, optimizing these costs becomes crucial. Materialized views offer a powerful solution by creating a pre-computed, optimized data representation. Imagine a retail scenario with separate customer and product tables. Typically, retrieving product details for a customer's purchase requires cross-referencing both tables.
|
By Pascal Vantrepote
Today, Confluent, the data streaming pioneer, is excited to announce its entrance into MongoDB’s new AI Applications Program (MAAP). MAAP is designed to help organizations rapidly build and deploy modern generative AI (GenAI) applications at enterprise scale.
|
By Confluent Staff
Recap: This is the last part of our four chapters: It’s been a long time coming, but we’ve finally arrived at the fourth and final installment of our blog series. In this series, we’ve been peeling back the layers of Apache Kafka to get a deeper understanding of how best to interact with the cluster using producer and consumer clients. At a high level, a fetch request is comprised of two parts: Let’s dive in.
|
By Greg Murphy
The Connect with Confluent (CwC) Technology Partner Program consistently expands the reach of Confluent’s data streaming platform across an ever-growing landscape of enterprise data systems. In this blog, you’ll meet the latest program entrants who have built fully managed integrations with Confluent and discover new ways to leverage real-time data across your business.
|
By Braeden Quirante
Raw data from IoT devices, like GPS trackers or electronic logging devices (ELDs), often lacks meaning on its own. However, if combined with information from other business systems, such as inventory management or customer relationship management (CRM), this data can now provide a richer, more complete picture for more effective decision-making. For example, combining GPS data with inventory levels can optimize logistics and delivery routes.
|
By Confluent
IT teams urged to resolve 'data delays' as UK executives struggle to access and use relevant business data.
|
By Confluent
Producer retries in Apache Kafka can make or break message delivery, especially during broker events like updates or failures. Use the idempotent producer, and configure delivery timeouts, in order to avoid common pitfalls that lead to lost messages or broken ordering.
|
By Confluent
Picking the wrong partition key in Apache Kafka? That’s a fast track to performance headaches—think unbalanced loads, slowdowns, and broken message ordering. Choosing the right partitioning strategy keeps your data flowing smoothly and avoids hot partitions.
|
By Confluent
Constantly starting and stopping Apache Kafka producers and consumers? That’s a recipe for high resource usage and inefficiency. Short-lived connections are heavy on resources, and can slow down your whole cluster. Keep them running to boost performance, cut latency, and get the most out of your Kafka setup.
|
By Confluent
Default settings in Apache Kafka work when you’re getting started, but aren't suited for production. Sticking with defaults, like a seven-day retention policy, or a replication factor of one, can cause storage issues, or data loss in case of failure. Learn why optimizing retention periods, replication factors, and partitions, is crucial for better Kafka performance and reliability.
|
By Confluent
Keeping your Apache Kafka clients up-to-date is critical for maximizing performance, security, and stability. In this video, we discuss why sticking with old versions could be putting you at risk, since it means you’re missing out on dozens of new features, and hundreds of bug fixes and security patches. Learn why upgrading is more than just a “nice-to-have”—it’s essential for a smoother and safer Kafka experience.
|
By Confluent
Setting up network connections can often seem difficult or time consuming. This video provides a wayfinding introduction to help you get networking up and running for all cluster types on Confluent Cloud, showing you your networking options for each cluster type when running on AWS, Azure, or Google Cloud, respectively.
|
By Confluent
In this short video, a Confluent Solutions Engineering will run through the high-level steps on how to get started with your migration. And even better, once you’re done watching, you can download our comprehensive migration kit for a step by step guide of everything I’ve talked about and more.
|
By Confluent
Kafka Connectors provide a way to get data flowing between your sources and sinks, and Confluent Cloud. This video reviews your options and shows you how to get started quickly.
|
By Confluent
This video provides a tour of the main Confluent Cloud interface, and shows you how to set up your first cluster.
|
By Confluent
Data visualization is cool, but have you tried setting up a chart of real-time data? In this video, Lucia Cerchie shows you how to create a live visualization of market data. She starts by producing data to a topic in Confluent Cloud from an Alpaca API websocket, then processes that data with Flink SQL, and finally uses a Streamlit component for a real-time visualization.
|
By Confluent
Traditional messaging middleware like Message Queues (MQs), Enterprise Service Buses (ESBs), and Extract, Transform and Load (ETL) tools have been widely used for decades to handle message distribution and inter-service communication across distributed applications. However, they can no longer keep up with the needs of modern applications across hybrid and multi cloud environments for asynchronicity, heterogeneous datasets and high volume throughput.
|
By Confluent
Why a data mesh? Predicated on delivering data as a first-class product, data mesh focuses on making it easy to publish and access important data across your organization. An event-driven data mesh combines the scale and performance of data in motion with product-focused rigor and self-service capabilities, putting data at the front and center of both operational and analytical use-cases.
|
By Confluent
When it comes to fraud detection in financial services, streaming data with Confluent enables you to build the right intelligence-as early as possible-for precise and predictive responses. Learn how Confluent's event-driven architecture and streaming pipelines deliver a continuous flow of data, aggregated from wherever it resides in your enterprise, to whichever application or team needs to see it. Enrich each interaction, each transaction, and each anomaly with real-time context so your fraud detection systems have the intelligence to get ahead.
|
By Confluent
Many forces affect software today: larger datasets, geographical disparities, complex company structures, and the growing need to be fast and nimble in the face of change. Proven approaches such as service-oriented (SOA) and event-driven architectures (EDA) are joined by newer techniques such as microservices, reactive architectures, DevOps, and stream processing. Many of these patterns are successful by themselves, but as this practical ebook demonstrates, they provide a more holistic and compelling approach when applied together.
|
By Confluent
Data pipelines do much of the heavy lifting in organizations for integrating, transforming, and preparing data for subsequent use in data warehouses for analytical use cases. Despite being critical to the data value stream, data pipelines fundamentally haven't evolved in the last few decades. These legacy pipelines are holding organizations back from really getting value out of their data as real-time streaming becomes essential.
|
By Confluent
In today's fast-paced business world, relying on outdated data can prove to be an expensive mistake. To maintain a competitive edge, it's crucial to have accurate real-time data that reflects the status quo of your business processes. With real-time data streaming, you can make informed decisions and drive value at a moment's notice. So, why would you settle for being simply data-driven when you can take your business to the next level with real-time data insights??
|
By Confluent
Data pipelines do much of the heavy lifting in organizations for integrating and transforming and preparing the data for subsequent use in downstream systems for operational use cases. Despite being critical to the data value stream, data pipelines fundamentally haven't evolved in the last few decades. These legacy pipelines are holding organizations back from really getting value out of their data as real-time streaming becomes essential.
|
By Confluent
Shoe retail titan NewLimits relies on a jumble of homegrown ETL pipelines and batch-based data systems. As a result, sluggish and inefficient data transfers are frustrating internal teams and holding back the company's development velocity and data quality.
- December 2024 (11)
- November 2024 (10)
- October 2024 (24)
- September 2024 (27)
- August 2024 (15)
- July 2024 (9)
- June 2024 (22)
- May 2024 (18)
- April 2024 (7)
- March 2024 (18)
- February 2024 (13)
- January 2024 (6)
- December 2023 (9)
- November 2023 (10)
- October 2023 (14)
- September 2023 (28)
- August 2023 (8)
- July 2023 (2)
Connect and process all of your data in real time with a cloud-native and complete data streaming platform available everywhere you need it.
Data streaming enables businesses to continuously process their data in real time for improved workflows, more automation, and superior, digital customer experiences. Confluent helps you operationalize and scale all your data streaming projects so you never lose focus on your core business.
Confluent Is So Much More Than Kafka:
- Cloud Native: 10x Apache Kafka® service powered by the Kora Engine.
- Complete: A complete, enterprise-grade data streaming platform.
- Everywhere: Availability everywhere your data and applications reside.
Apache Kafka® Reinvented for the Data Streaming Era