New pricing for the AI API

The cost per token of using LLMs has fallen substantially since we initially launched the AI API. Our initial pricing of one request per CCU was necessary at the time, but now with much more efficient models like Claude 3.5 Sonnet and GPT-4o available, we can offer a much lower price of 5 requests per CCU.

As of December 2024, the AI API is available for all paying Cube Cloud customers, including those on starter plans. You do not need to reach out to your AE or intercom to get it enabled; it is already enabled.

Here is a link to our AI API docs on how to use it: [https://cube.dev/docs/product/apis-integrations/ai-api]

The AI API needs to be enabled in your deployment in order to use it. Toggle to enable in your configuration flags:

Cube's AI API is already a powerful way to allow users to access your data with AI quickly, safely, and transparently. We've heard from customers that users often refer to values inside of a dimension instead of stating where the dimension equals a specific value. This usually means the request won't be successful, and the user won't get the data they're looking for.

We have released a feature called Value Search for use with the AI API, which addresses this problem. Value Search allows Cube Cloud customers to select specific dimensions to be indexed so their values can be referenced by a query. Our RAG system can then use relevant values to answer questions that contain these dimension values or similar terms. It will include the pair of the dimension name and the value to filter the dimension.

Here is an example of how you can make a dimension searchable in your Cube data model YAML:

- name: order_status
sql: order_status
type: string
meta:
ai:
searchable: true

Here is our documentation related to the feature: https://cube.dev/docs/product/apis-integrations/ai-api#value-search

BYOLLM

Committed Cube Cloud customers can now choose to bring their own LLM to use with the AI API and incur costs this way instead of paying one Cube Cloud CCU per 5 requests. The AI API is compatible with Claude 3.5 Sonnet, available from AWS Bedrock, GCP Vertex, Snowflake, and Anthropic, OR GPT4o, available from OpenAI or Microsoft Azure. This enables almost anyone to use the AI API with their own LLM, whether for cost, security, or both.

Here are instructions on how you can change the LLM provider for AI API use.

We are also offering free workshops for customers looking to put AI API into production until the end of April.

Launching new Slack app for Cube AI API

Many of our customers have requested this feature since we launched the AI API. They expressed interest in enabling their stakeholders to interact with their Cube semantic layer through Slack using AI, similar to our AI API. The AI API ensures that questions are answered rapidly, safely, and transparently.

You can install the Cube AI API Slack application easily in the settings page for AI API:

Once installed, the Slack app will introduce itself to your workspace individually, starting a conversation with each user. Users can then direct message the application with their data questions, which will be turned into threads where follow-up questions can be asked and feedback can be given. You can choose to receive a log of all questions asked and answered in a separate private channel for your engineering team to observe and learn from.

Want to learn more? Contact our sales team today.