Systems | Development | Analytics | API | Testing

Latest News

How to Design a Scalable Rate Limiting Algorithm

Rate limiting protects your APIs from inadvertent or malicious overuse by limiting how often each user can call the API. Without rate limiting, each user may make a request as often as they like, leading to “spikes” of requests that starve other consumers. Once enabled, rate limiting can only perform a fixed number of requests per second. A rate limiting algorithm helps automate the process. In the example chart, you can see how rate limiting blocks requests over time.

Kafka Total Cost of Ownership: What are you missing?

“We’ve seen two years’ worth of digital transformation in two months” said Microsoft’s Satya Nadella. Due to COVID-19, digital transformation roadmaps have been deleted, redrafted, doubled down and accelerated by up to a decade. Traditional companies are moving by osmosis towards streaming technologies such as Apache Kafka to kick off new digital services. But how much should it cost to experience 2030 in 2021?

Common Regulations that Data-Driven Entities Need to Know

For public and private entities, data collection is a way of life. That fact has led to the proliferation of common regulations to protect consumers and individuals from unacceptable use or storage of their private data. But it's not just data collection laws companies have to adhere to. There are many US-based and international statutes that put constraints on how they do business. What follows summarizes the most common regulations and how they can affect the work you do, day to day.

Cloudera Data Warehouse Demonstrates Best-in-Class Cloud-Native Price-Performance

Cloud data warehouses allow users to run analytic workloads with greater agility, better isolation and scale, and lower administrative overhead than ever before. With the ability to quickly provision on-demand and the lower fixed and administrative costs, the costs of operating a cloud data warehouse are driven mostly by the price-performance of the specific data warehouse platform.

Uncover Gold During an Economic Crisis: Five Steps to Monetizing Your Data

Because of the COVID-19 global pandemic, almost every industry is experiencing volatility, risks and changes to buying behavior. Nevertheless, in crisis often comes opportunity and a forcing factor for businesses to redefine themselves. Those looking to innovate after (or even during) this crisis should focus on two key concepts — data monetization and data modernization.

Looking at the uses of JMeter Timers

Timers in JMeter are incredibly important when it comes to the balance and pace of your performance tests, we are going to look at the Timers that ship with the standard JMeter installation in this Blog post but there are others that are available as a Plugin and hopefully this post will encourage you to investigate these further.

Add Java Agents to Existing Kubernetes and Helm Applications Instantly

In a recent blog post, one of my teammates, Josh, shared a few techniques for deploying Java agents in Kubernetes applications. We have been getting a lot of interest in the concepts we have shared and, per popular request, decided to raise the bar. Is it possible to add a Java agent without changing a single line in either the Dockerfile or the Kubernetes Manifest? Well, the answer is most definitely yes (!), and here’s how.

Loading complex CSV files into BigQuery using Google Sheets

BigQuery offers the ability to quickly import a CSV file, both from the web user interface and from the command line: Indeed, try to open this file up with BigQuery: and we get the errors like: This is because a row is spread across multiple lines, and so the starting quote on one line is never closed. This is not an easy problem to solve — lots of tools struggle with CSV files that have new lines inside cells. Google Sheets, on the other hand, has a much better CSV import mechanism.

Talend vs. MuleSoft vs. Xplenty: Which One Does ETL Better?

The key differences between Talend, MuleSoft, and Xplenty: Enterprise data volumes are increasing by 63 percent per month, according to a recent study. Twenty percent of organizations draw from 1,000 or more data sources. How do these companies extract and move all this data to a centralized destination for business analytics? As we know, Extract, Transform, and Load (ETL) streamlines this entire process. But smaller organizations lack the coding skills required for successful implementation.