Systems | Development | Analytics | API | Testing

Blog

Stop being an Excel slave

All marketers have to deal with essentially the same problem when managing their online campaigns and looking for the right attribution model: the data they need for evaluation is scattered across various systems, and collecting and collating it takes too much time. If they decide to use solutions intended to help them to handle this problem, they often find such solutions are not, in themselves, enough.

3 things you should never measure in BI

When I speak to people who are thinking about implementing BI, they are often overwhelmed by all the things they could measure. Many start by wanting to measure everything, which doesn’t necessarily help them. That’s because there’s an inherent cost in measuring things – everything you report and track creates an ongoing burden that your organization has to maintain. That’s why it’s important to be selective about what you measure from the get-go.

How to Develop a Data Processing Job Using Apache Beam - Streaming Pipelines

In our last blog, we talked about developing data processing jobs using Apache Beam. This time we are going to talk about one of the most demanded things in modern Big Data world nowadays – processing of Streaming data. The principal difference between Batch and Streaming is the type of input data source. When your data set is limited (even if it’s huge in terms of size) and it is not being updated along the time of processing, then you would likely use a batching pipeline.

Talend and Splunk: Aggregate, Analyze and Get Answers from Your Data Integration Jobs

Log management solutions play a crucial role in an enterprise's layered security framework— without them, firms have little visibility into the actions and events occurring inside their infrastructures that could either lead to data breaches or signify a security compromise in progress. Splunk is the “Google for log files” heavyset enterprise tool that was the first log analysis software and has been the market leader ever since.

Talend & Apache Spark: Debugging & Logging

So far, our journey on using Apache Spark with Talend has been a fun and exciting one. The first three posts on my series provided an overview of how Talend works with Apache Spark, some similarities between Talend and Spark Submit, the configuration options available for Spark jobs in Talend and how to tune Spark jobs for performance. If you haven’t already read them you should do so before getting started here.

Yellowfin Bytes: Text Replacement and Regular Expressions

The third edition of Yellowfin Bytes brings you an exciting inclusion made to the Data Transformation module. Users have been exploring Yellowfin’s very own lite version of an ETL tool in the form of this module, ever since its introduction in Yellowfin 7.4 last year. We have since been enhancing this functionality by adding new steps, calculations, configurations, and more.