Analytics

Automating Data Pipelines in CDP with CDE Managed Airflow Service

When we announced the GA of Cloudera Data Engineering back in September of last year, a key vision we had was to simplify the automation of data transformation pipelines at scale. By leveraging Spark on Kubernetes as the foundation along with a first class job management API many of our customers have been able to quickly deploy, monitor and manage the life cycle of their spark jobs with ease. In addition, we allowed users to automate their jobs based on a time-based schedule.

Which Sources Drive The Highest Conversion Rates?

When it comes to conversion rates, channel sources should be at the center of your focus. After all, you want to know which of your marketing investments is driving the most conversions so you can double down or make adjustments to your strategy. In this episode of Data Snacks, we show you: How to track contact and customer conversions by source Other metrics that matter when it comes to conversions What you can do to increase your conversion rates by source

The Citizen Integrator: Key to Business Agility

With the rapidly changing pace of innovative technology, companies must be able to pivot quickly or perish. The ability to adapt to change is critical to a company’s success. A key factor in the ability to pivot is access to real-time information to facilitate data-driven decisions. Traditionally, that data has existed across multiple systems with no simple method for bringing it all together meaningfully.

Announcing the GA of Cloudera DataFlow for the Public Cloud

Are you ready to turbo-charge your data flows on the cloud for maximum speed and efficiency? We are excited to announce the general availability of Cloudera DataFlow for the Public Cloud (CDF-PC) – a brand new experience on the Cloudera Data Platform (CDP) to address some of the key operational and monitoring challenges of standard Apache NiFi clusters that are overloaded with high-performant flows.

Cloudera DataFlow for the Public Cloud: A technical deep dive

We just announced Cloudera DataFlow for the Public Cloud (CDF-PC), the first cloud-native runtime for Apache NiFi data flows. CDF-PC enables Apache NiFi users to run their existing data flows on a managed, auto-scaling platform with a streamlined way to deploy NiFi data flows and a central monitoring dashboard making it easier than ever before to operate NiFi data flows at scale in the public cloud.

Transforming Customer Data for Salesforce

CRM (customer relationship management) software is the lifeblood of any modern B2C company. By monitoring and storing all of your interactions with prospects and customers—from their first visit to your website to their most recent purchase—CRM software makes it dramatically easier to segment your customer base, identify hidden trends in the data, make smarter predictions, and forecasts, and much more.

Cloudera DataFlow for the Public Cloud

Cloudera DataFlow for the Public Cloud takes away the operational and monitoring challenges by providing cloud-native flow management capabilities powered by Apache NiFi. It is a purposely built framework to modernize the data flow user experience so that the NiFi developers and administrators can be prepared to easily handle sophisticated data flows in production.

Dining with data: A Q&A with OpenTable's Senior Vice President of Data and Analytics Grant Parsamyan

For more than 20 years, OpenTable has connected foodies and novice diners with the restaurants they love. But how does its technology work on the back end? To make a long story short: data. Beyond the app and website, OpenTable provides restaurants with software that manages their floor plans, phone reservations, walk-ins, shift scheduling, turn times, and more.

Why Modernizing the First Mile of the Data Pipeline Can Accelerate all Analytics

Every enterprise is trying to collect and analyze data to get better insights into their business. Whether it is consuming log files, sensor metrics, and other unstructured data, most enterprises manage and deliver data to the data lake and leverage various applications like ETL tools, search engines, and databases for analysis. This whole architecture made a lot of sense when there was a consistent and predictable flow of data to process.