Keywords AI

Changelog

New updates and improvements to Keywords AI.

October 15, 2024

🎈 LLM Caching with Images

Snapshot for version undefined

You can now cache your LLM requests with images on Keywords AI.

Processing images with LLMs is more expensive than text processing. To help manage costs:

Enable caching and include the image link in your API payload to cache the entire LLM request with an image.

October 12, 2024

⚖️ Evaluations 2.0

Snapshot for version undefined

We're thrilled to introduce LLM-as-a-judge Evaluation, now integrated with Relari and ragas. Evaluate performance across various metrics, including Hallucination, Answer relevancy, Customized evaluations, and more.

You can create evaluators to apply RAG-specific evaluations on data collected in Keywords AI.

Check out the docs here: Evaluations 2.0.

October 9, 2024

👩‍👩‍👦‍👦 New Organization Feature Update

Snapshot for version undefined
  • Create teams for your projects under one organization for easy switching.
  • Invite teammates to specific teams for better access control.

September 30, 2024

Prompts 2.0

Snapshot for version undefined

We've improved our Prompts page, creating a smoother process for deploying optimized prompts to production.

September 26, 2024

Introducing Lab & Testsets: Efficient LLM evaluation tools

Snapshot for version undefined

We've launched Lab & Testsets! 🔬

📡 Lab: A spreadsheet-style editor for running prompts and models across multiple test cases. Import testsets to easily test, evaluate and optimize your LLM outputs.

📊 Testsets: Easily manage and organize test cases. Import a CSV file and edit it like a Google Sheet.

September 18, 2024

Customize Retries on Keywords AI

Snapshot for version undefined

Retries feature: When an LLM call fails, our system detects the error and retries the request to prevent failover.

Customize your Retries settings!

  • Set number of retries
  • Define initial retry time in seconds

September 16, 2024

Manage your usage limit and rate limits.

Snapshot for version undefined

We've added a new Limits page where you can view your current usage and rate limits.

You can also customized your usage limits for every deployment.

  • Set monthly budget for every deployment
  • Set email alerts when you reach your budget

September 12, 2024

OpenAI o1 family now available on Keywords AI

Snapshot for version undefined

O1 family is designed to spend more time thinking before they respond. They can reason through complex tasks and solve harder problems than previous models in science, coding, and math.

September 11, 2024

BAML integration

Snapshot for version undefined

We've partnered with Boundary (YC W23)! Now you can build your LLM app using BAML and monitor it with Keywords AI (YC W24). Simply create a Keywords AI client to get started.

BAML is a templating language for writing typed LLM functions, treating prompts as functions. Learn how to integrate: BAML documentation.

September 7, 2024

User intent classification

Snapshot for version undefined

We participated in Mintlify's hackathon this weekend and launched an open-source project for user intent classification. This tool can be used to identify user intents in chatbots or other conversational AI projects.

For more details, please visit our Github repository.

August 29, 2024

Caches UI

Snapshot for version undefined

We've added Caches to our frontend, allowing you to see cache hit counts and the time and cost savings from using the cache.

For more information on Caches, please refer to our documentation.

August 24, 2024

LLM monitoring -> Multimodal LLM monitoring

Snapshot for version undefined

We're excited to announce that we've expanded to multimodal monitoring! Our unified Model API now supports over 200 LLMs, embedding models, and audio models — all monitored on a single platform.

Learn how to monitor your multimodal models by checking out our documentation.

August 16, 2024

PostHog integration

Snapshot for version undefined

You can bring LLM metrics from Keywords AI (YC W24) to your PostHog dashboard easily! Check out the tutorial here!

August 7, 2024

New LLM usage page

Snapshot for version undefined

We're introducing the new LLM Usage page! This page shows an overview of your LLM usage, including request numbers, LLM costs, and evaluation costs. It also breaks down your usage by month, helping you track patterns and improve AI performance.

July 27, 2024

LLM request caching

Snapshot for version undefined

Supercharge your LLM calls with caching. Our new Caches feature allows you to store and reuse LLM responses, eliminating redundant API calls.

This smart caching system optimizes your AI performance by delivering instant responses, reducing costs, and ensuring consistent, high-quality outputs.

July 24, 2024

Mistral Large 2 availability

Snapshot for version undefined

Integrated Mistral Large 2 into Keywords AI, try it out in our model playground.

July 23, 2024

Llama 3.1 family availability

Snapshot for version undefined

Integrated Llama 3.1 faimliy into Keywords AI, try it out in our model playground.

July 22, 2024

Introducing Threads

Snapshot for version undefined

We pushed our latest update to improve the observability of chatlogs. Now, you can group logs by thread ID to have a better understanding of the conversation flow.

July 18, 2024

GPT-4o mini availability

Snapshot for version undefined

Integrated GPT-4o mini into Keywords AI, try it out in our model playground.

July 16, 2024

Minute-level dashboard graphs

Snapshot for version undefined

We have put much effort into improving the performance of our dashboard. Now, you can see your LLM usage and performance at minute and hour levels. Its loading time is even 2x faster than before.

July 15, 2024

Introducing fallbacks

Snapshot for version undefined

Today, you could specify fallback models for your LLM deployments on our platform. If the primary model fails to respond, your fallback models will be used instead. This feature is especially useful for critical deployments where you can't afford any downtime.

July 10, 2024

Model load balancing

Snapshot for version undefined

We're thrilled to introduce our latest feature.

There are 2 ways to load balance your LLM requests. First, you can specify weights for the models you want to load balance in the code, ensuring that requests are distributed based on your desired percentages.

The second way is to easily add your credentials and set the weight for each one to distribute requests efficiently between deployments.

July 1, 2024

Introducing Datasets

Snapshot for version undefined

Today, we are introducing a new feature called Datasets. With Datasets, you can effortlessly save and export log data for various purposes, such as fine-tuning, synthetic data generation, and evaluation. Simply click the 'Create Dataset' button on the Logs page and adjust the filters as needed.

Keywords AI - the LLM observability platform.
Backed byCombinator