Systems | Development | Analytics | API | Testing

Technology

Mastering StudioAssist: Your Essential Guide to Data-Driven Testing and Reusing AI-Generated Test Cases

As a software tester, you constantly look for ways to streamline your workflow and enhance your testing capabilities. StudioAssist, an intelligent feature in Katalon Studio, is designed to make your life easier by automating repetitive tasks and generating efficient test scripts. This blog will serve as your cookbook, providing you with step-by-step instructions on how to harness the full potential of StudioAssist.

How to bridge the gap between humans and AI

In this episode, hear Sadie St. Lawrence’s thoughts on how to effectively leverage Generative AI at work by asking the right questions, and how the technology can help you to expand on your divergent thinking. There’s so much more to the future of work with Generative AI now at its core. Sadie shares where we’re headed, and how we can bridge the gap between humans and AI.

Volumes Technical Preview: Blazing-fast NVMe SSD for Your Data

Ready? Day three of Koyeb launch week is on! When you deploy your apps on Koyeb, your data is on ephemeral disks. While this works great for stateless applications, this is challenging for stateful workloads like databases. Just in time to save the day, we are launching the technical preview of Volumes! You can now use Volumes to persist data between deployments, restarts, and even when services are paused. We're gradually onboarding users to ensure the best experience for everyone.

Empowering Enterprise Generative AI with Flexibility: Navigating the Model Landscape

The world of Generative AI (GenAI) is rapidly evolving, with a wide array of models available for businesses to leverage. These models can be broadly categorized into two types: closed-source (proprietary) and open-source models. Closed-source models, such as OpenAI’s GPT-4o, Anthropic’s Claude 3, or Google’s Gemini 1.5 Pro, are developed and maintained by private and public companies.

GPUs Public Preview: Run AI workloads on H100, A100, L40S, and more

Welcome to day two of Koyeb launch week. Today we're announcing not one, but two major pieces of news: Our lineup ranges from 20GB to 80GB of vRAM with A100 and H100 cards. You can now run high-precision calculations with FP64 instructions support and a gigantic 2TB/s of bandwidth on the H100. With prices ranging from $0.50/hr to $3.30/hr and always billed by the second, you'll be able to run training, fine-tuning, and inference workloads with a card adapted to your needs.

What is a Headless Data Architecture?

The headless data architecture. Is it a fad? Some marketecture? Or something real? In this video, Adam Bellemare takes you through the basics of the headless data architecture and why it’s beginning to emerge as its own respective pattern. Driven by the decoupling of data computation from storage, the headless data architecture provides the basis for a modular data ecosystem. Stream your data for near real-time low latency use cases, or convert it to an Iceberg table for analytical use cases.