Systems | Development | Analytics | API | Testing

Latest News

Cloud Data Management Guide: Solutions & Best Practices

Your data can quickly get out of control when you’re working with multiple cloud storage services and applications throughout your organization. Complex cloud ecosystems can make it difficult to know what data you have, how it’s being managed, whether it’s safe, and how to use it effectively. Cloud data management platforms can stop this frustrating scenario in its tracks.

Get to Know Your Retail Customer: Accelerating Customer Insight and Relevance

There are lessons to be learned from the brick and mortar or pure-play digital retailers that have been successful in the Covid-19 chaos. As the pandemic’s stress test of e-commerce, in-store insights, supply chain visibility, and fulfillment capabilities have revealed shortcomings, and long-lasting consumer experiences— it has also allowed many companies to pivot to very successful strategies built on enterprise data and the digitization efforts that accompany it.

Global View Distributed File System with Mount Points

Apache Hadoop Distributed File System (HDFS) is the most popular file system in the big data world. The Apache Hadoop File System interface has provided integration to many other popular storage systems like Apache Ozone, S3, Azure Data Lake Storage etc. Some HDFS users want to extend the HDFS Namenode capacity by configuring Federation of Namenodes. Other users prefer other alternative file systems like Apache Ozone or S3 due to their scaling benefit.

The practical benefits of augmented analytics

Augmented analytics uses emerging technologies like automation, artificial intelligence (AI), machine learning (ML) and natural language generation (NLG) to automate data manipulation, monitoring and analysis tasks and enhance data literacy. In our previous blog, we covered what augmented analytics actually is and what it really means for modern business intelligence.

What Is a Data Pipeline?

A data pipeline is a series of actions that combine data from multiple sources for analysis or visualization. In today’s business landscape, making smarter decisions faster is a critical competitive advantage. Companies desire their employees to make data-driven decisions, but harnessing timely insights from your company’s data can seem like a headache-inducing challenge.

What are ETL tools?

Thinking of building out an ETL process or refining your current one? Read more to learn about how ETL tools give you time to focus on building data models. ETL stands for extract-transform-load, and is commonly used when referring to the process of data integration. Extract refers to pulling data from a particular data source. Transforms are used to make that data into a processable format. Load is the final step to drop the data into the designated target.

Which is the Execution Order of Elements in Apache JMeter?

The execution order of the elements in an Apache JMeter performance test plan is very important, it helps understand the timeline of events and how your test will behave. Make sure that you understand every JMeter element before adding it to your test plan, this will help create faster performance tests and not jump into various errors.

Checklist For Mobile App Testing

Every mobile application, irrespective of its category, has a common goal of creating an impeccable user experience. App users want to see something new and innovative. And even though the great user experience is the most important, it only comes with quality. To ensure quality, mobile app testing is essential. In this blog, we will discuss a step-by-step mobile app testing checklist.

Introduction to Event Loop Utilization in Node.js

In the last year, I've spent many hours writing patches for libuv and Node to collect new metrics. The goal of this was to indirectly infer the state of the application without introducing measurable overhead. I've run a few hundred hours of benchmarks and collected over one million data points to make sure my analysis was correct. Eventually, I plan to write about all aspects of my research, but today we will focus on a metric that has already been added to Node.