Systems | Development | Analytics | API | Testing

BI

The Data Cloud & Public Sector With Deloitte

Welcome to Data Cloud Now, where we are dedicated to illustrating how the Data Cloud is pushing the possible forward, each and every day. In this episode we are exploring the impact it’s having within the public sector. Explore the impact the Data Cloud is having within the Public Sector with special guests Deloitte’s Monica McEwen & Frank Farrall along with Snowflake’s Jeff Frazier.

Developing More Accurate and Complex Machine-Learning Models with Snowpark for Python

Sophos protects people online with a suite of cybersecurity products. Hear Konstantin Berlin, Head of Artificial Intelligence at Sophos, explain how the Snowflake Data Cloud helps Sophos increase the accuracy of their machine-learning models by allowing data scientists to process large and complex data sets independent of data engineers. Through Snowpark, data scientists can run Python scripts along with SQL without having to move data across environments, significantly increasing the pace of innovation.

3-Minute Recap: Unlocking the Value of Cloud Data and Analytics

DBTA recently hosted a roundtable webinar with four industry experts on “Unlocking the Value of Cloud Data and Analytics.” Moderated by Stephen Faig, Research Director, Unisphere Research and DBTA, the webinar featured presentations from Progress, Ahana, Reltio, and Unravel. You can see the full 1-hour webinar “Unlocking the Value of Cloud Data and Analytics” below. Here’s a quick recap of what each presentation covered.

Get Ready for the Next Generation of DataOps Observability

I was chatting with Sanjeev Mohan, Principal and Founder of SanjMo Consulting and former Research Vice President at Gartner, about how the emergence of DataOps is changing people’s idea of what “data observability” means. Not in any semantic sense or a definitional war of words, but in terms of what data teams need to stay on top of an increasingly complex modern data stack.

Ep 59: New Zealand's Crown Research Institute CDAO, Jan Sheppard on Treating Data as a Treasure

Treating data as a treasure is a foundational principle for Jan Sheppard, the Chief Data and Analytics officer at New Zealand’s Crown Research Institute of Environmental Science and Research (ESR.) This agency leads ongoing research in public health, environmental health, and forensics for the country of New Zealand. Like many other CDAOs, her role is relatively new. But the unique values she applies to data can be traced back many hundreds of years to the indigenous Maori people of her country. Through her work, Jan recognizes the profound impact data can have on people and their environments for generations to come.

Talend's contributions to Apache Beam

Apache Beam is an open-source, unified programming model for batch and streaming data processing pipelines that simplifies large-scale data processing dynamics. The Apache Beam model offers powerful abstractions that insulate you from low-level details of distributed data processing, such as coordinating individual workers, reading from sources and writing to sinks, etc.

Building an automated data pipeline from BigQuery to Earth Engine with Cloud Functions

Over the years, vast amounts of satellite data have been collected and ever more granular data are being collected everyday. Until recently, those data have been an untapped asset in the commercial space. This is largely because the tools required for large scale analysis of this type of data were not readily available and neither was the satellite imagery itself. Thanks to Earth Engine, a planetary-scale platform for Earth science data & analysis, that is no longer the case.

Analyzing satellite images in Google Earth Engine with BigQuery SQL

Google Earth Engine (GEE) is a groundbreaking product that has been available for research and government use for more than a decade. Google Cloud recently launched GEE to General Availability for commercial use. This blog post describes a method to utilize GEE from within BigQuery’s SQL allowing SQL speakers to get access to and value from the vast troves of data available within Earth Engine.

How to simplify and fast-track your data warehouse migrations using BigQuery Migration Service

Migrating data to the cloud can be a daunting task. Especially moving data from warehouses and legacy environments requires a systematic approach. These migrations usually need manual effort and can be error-prone. They are complex and involve several steps such as planning, system setup, query translation, schema analysis, data movement, validation, and performance optimization.

Scaling Kafka Brokers in Cloudera Data Hub

This blog post will provide guidance to administrators currently using or interested in using Kafka nodes to maintain cluster changes as they scale up or down to balance performance and cloud costs in production deployments. Kafka brokers contained within host groups enable the administrators to more easily add and remove nodes. This creates flexibility to handle real-time data feed volumes as they fluctuate.