Systems | Development | Analytics | API | Testing

Product Release

Introducing Confluent Cloud OpenSearch Sink Connector

Amazon OpenSearch is a popular fully managed analytics engine that makes it easier for customers to do interactive log analytics, real-time application monitoring, and semantic and keyword searches. It can also be used as a vector engine that helps organizations build and augment GenAI applications without managing infrastructure (we’ll talk about this in future blogs). Additionally, the service provides a reliable, scalable infrastructure designed to handle massive data volumes.

Web CI on Bitrise: Beta announcement

Mobile developers love using Bitrise, and we get it. But we also know your products and automation workflows extend beyond just mobile apps. That's why we're thrilled to announce Bitrise's expansion into web CI/CD! While in beta, this is your chance to help shape an industry-leading solution for your entire development process.

Insight With Eyesight: Qlik Introduces a New Era of Visualization

Our ability to tell stories is an art form as old as language itself. From ancient cave paintings to oral traditions passed through generations, the essence of stories has evolved alongside our communication methods. It began with visual tales etched on cave walls, transitioned into spoken narratives, and eventually found its way into written, printed, and typed forms.

Introducing Cloudera's AI Assistants

In the last couple of years, AI has launched itself to the forefront of technology initiatives across industries. In fact, Gartner predicts the AI software market will grow from $124 billion in 2022 to $297 billion in 2027. As a data platform company, Cloudera has two very clear priorities. First, we need to help customers get AI models based on trusted data into production faster than ever.

Introducing Choreo Copilot

We're excited to introduce Choreo Copilot (preview), which allows you to interact with Choreo. You can pose questions in natural language and Copilot will provide answers. Choreo Copilot enables you to grasp Choreo concepts, teaches you how to perform tasks in Choreo, and provides guidance when you encounter obstacles. Copilot is familiar with APIs in Choreo’s internal marketplace. Choreo already features an AI capability that enables API testing through natural language.

Solving Memory Leaks in Node.js has Never Been Easier, Introducing the Latest Version of N|Solid

We are thrilled to announce the release of a new feature in N|Solid that includes sample heap profiling and heap objects observability for main processes and worker threads. N|Solid is known for its Node.js performance and security observability and diagnostic tools and best-in-class low overhead has completed a new innovation to hunt memory leaks in development and production environments. Memory is a precious resource in computing.

Kong Mesh 2.7: Leveling Up Usability and Security

Kong Mesh 2.7 is here! And with it come a slew of new features to make our simple and powerful enterprise service mesh even easier to use and more powerful for your particular needs. In this blog post, we'll cover the new features and show how Kong Mesh 2.7 can deliver a superior service mesh experience for your organization.

Introducing Testim Copilot: Accelerate test creation through generative AI

We are thrilled to announce the arrival of Tricentis Testim Copilot, the first of many Tricentis Copilot solutions. Tricentis Copilot solutions are a collection of advanced generative AI capabilities available as add-ons across our products that help customers boost their efficiency throughout the entire testing lifecycle. With Testim Copilot, we’ve added new advanced generative AI capabilities to Testim, an already AI-driven test automation platform.

Snowflake Arctic: The Best LLM for Enterprise AI - Efficiently Intelligent, Truly Open

Building top-tier enterprise-grade intelligence using LLMs has traditionally been prohibitively expensive and resource-hungry, and often costs tens to hundreds of millions of dollars. As researchers, we have grappled with the constraints of efficiently training and inferencing LLMs for years.