Systems | Development | Analytics | API | Testing

Latest News

Introducing Tricentis Copilot solutions

We are thrilled to announce Tricentis Copilot solutions, a collection of advanced generative AI capabilities available across our products that help customers boost their efficiency throughout the entire testing lifecycle. With Tricentis Copilot solutions, you can autogenerate manual tests from requirements, optimize your portfolio, autogenerate custom code, and get meaningful insights.

Turbocharging Your Business with (Gen)AI

If you were to stop someone walking down the street and ask them how long artificial intelligence, or AI, has been a hot topic, they might say it’s something that’s emerged mostly in recent years. But AI has been around for a long time, with the term first being coined as long ago as 1955. Generative AI however is a different beast, and one that's largely responsible for moving the topic of AI to the tip of everyone’s tongues – from consumers to enterprises alike.

Testing generative AI systems and red teaming: An introductory guide

The topic of testing AI and ensuring its responsibility, safety, and security has never been more urgent. Controversy and incidents of AI misuse have increased 26-fold since 2021, highlighting growing concerns. As users quickly find out, AI tools are not infallible; they can make mistakes, display overconfidence, and lack critical questioning. The reality of the market is that AI is prone to error. This is exactly why testing AI is crucial. But how do we test AI?

Snowflake Launches the World's Best Practical Text-Embedding Model for Retrieval Use Cases

Today Snowflake is launching and open-sourcing with an Apache 2.0 license the Snowflake Arctic embed family of models. Based on the Massive Text Embedding Benchmark (MTEB) Retrieval Leaderboard, the largest Arctic embed model with only 334 million parameters is the only one to surpass average retrieval performance of 55.9, a feat only less practical to deploy models with over 1 billion parameters are able to achieve.

LLM Metrics: Key Metrics Explained

Organizations that monitor their LLMs will benefit from higher performing models at higher efficiency, while meeting ethical considerations like ensuring privacy and eliminating bias and toxicity. In this blog post, we bring the top LLM metrics we recommend measuring and when to use each one. In the end, we explain how to implement these metrics in your ML and gen AI pipelines.

Unleashing the Power of Digital Assurance in the Age of AI and Gen AI: Charting the Way Forward

In the pursuit of excellence, Quality Assurance (QA) has embarked on a profound journey of automation. Beginning with manual testing as its foundation, QA has progressed steadily through functional automation and smart automation, culminating in its embrace of Intelligent automation and Codeless automation. This evolution mirrors our transition from traditional waterfall models to agile methodologies.

Why RAG Has a Place in Your LLMOps

With the explosion of generative AI tools available for providing information, making recommendations, or creating images, LLMs have captured the public imagination. Although we cannot expect an LLM to have all the information we want, or sometimes even include inaccurate information, consumer enthusiasm for using generative AI tools continues to build.