Systems | Development | Analytics | API | Testing

Blog

The Rise of Ad Hoc and Citizen Integrators

In the past few years, there has been a shift in the data industry, leading to the emergence of a new category of data citizens: the ‘ad hoc’ or ‘citizen’ integrators. With these new personas adding to the (already long) list of data workers having access to corporate information, companies are needing to re-think the way they approach their data security and data governance strategies.

Making data-intensive processing efficient and portable with Apache Beam

The appearance of Hadoop and its related ecosystem was like a Cambrian explosion of open source tools and frameworks to process big amounts of data. But companies who invested early in big data found some challenges. For example, they needed engineers with expert knowledge not only on distributed systems and data processing but also on Java and the related JVM-based languages and tools.

Docker Tutorial: Get Going From Scratch

Docker is a platform for packaging, deploying, and running applications. Docker applications run in containers that can be used on any system: a developer’s laptop, systems on premises, or in the cloud. Containerization is a technology that’s been around for a long time, but it’s seen new life with Docker. It packages applications as images that contain everything needed to run them: code, runtime environment, libraries, and configuration.

Bugfender 2.0 - Behind the Scenes

As developers at Mobile Jazz, we know the headache of finding, reproducing, and fixing bugs in mobile apps. Several years ago, we got so fed up debugging mobile apps remotely that we started building a solution for ourselves. We did it by creating a way to remotely access the logging facilities of users’ devices. We built a makeshift server for application logs that allowed us to fix bugs across devices and continents.

Automation Made Easy: Building Jenkins Jobs with Jenkins DSL

Recently, I've seen a lot of articles and conversation around DevOps. DevOps is a set of practices that automates the processes between software development and IT teams, in order to enable be shorter development cycles, increased deployment frequency, and more dependable releases all in conjunction with business needs.

How to Develop a Data Processing Job Using Apache Beam

Are you familiar with Apache Beam? If not, don’t be ashamed, as one of the latest projects developed by the Apache Software Foundation and first released in June 2016, Apache Beam is still relatively new in the data processing world. As a matter of fact, it wasn’t until recently when I started to work closely with Apache Beam, that I loved to learn and learned to love everything about it.