Systems | Development | Analytics | API | Testing

Latest News

Using BigQuery ML and BigQuery GIS together to predict NYC taxi trip cost

In this article, I’ll walk you through the process of building a machine learning model using BigQuery ML. As a bonus, we’ll have the chance to use BigQuery’s support for spatial functions. We’ll use the New York City taxicab dataset, with the goal of predicting taxi fare, given both pick-up and drop-off locations for each ride — imagine that we are designing a trip planner.

What's happening in BigQuery: integrated machine learning, maps, and more

In this month’s installment of What’s Happening in BigQuery, we’re sharing new features intended to make your life easier: some make BigQuery more performant and more cost effective, while others, like BigQuery ML, enable groundbreaking analysis tools in a cloud data warehouse that’s a first of its kind. First off, we just finished Next ‘18, our annual event focused on all things cloud.

Talend Unveils Bulk Data Uploader to Enable Large-Scale, Real-Time Analytics on Microsoft Azure SQL Data Warehouse

Redwood City, CA - August 16, 2018 - Talend(NASDAQ: TLND), a global leader in cloud integration solutions, unveiled a new bulk data uploader for Microsoft Azure SQL Data Warehouse, which helps empower thousands of users across an organization with a fast, flexible and secure cloud data warehouse.

Going Serverless with Talend through CI/CD and Containers

Continuous integration, delivery and deployment, known as CI/CD, has become such a critical piece in every successful software project that we cannot deny the benefits it can bring to your project. At the same time, containers are everywhere right now and are very popular among developers. In practice, CI/CD delivery allows users to gain confidence in the applications they are building by continuously test and validate them.

3 things you should never measure in BI

When I speak to people who are thinking about implementing BI, they are often overwhelmed by all the things they could measure. Many start by wanting to measure everything, which doesn’t necessarily help them. That’s because there’s an inherent cost in measuring things – everything you report and track creates an ongoing burden that your organization has to maintain. That’s why it’s important to be selective about what you measure from the get-go.

How to Develop a Data Processing Job Using Apache Beam - Streaming Pipelines

In our last blog, we talked about developing data processing jobs using Apache Beam. This time we are going to talk about one of the most demanded things in modern Big Data world nowadays – processing of Streaming data. The principal difference between Batch and Streaming is the type of input data source. When your data set is limited (even if it’s huge in terms of size) and it is not being updated along the time of processing, then you would likely use a batching pipeline.