Skip to main content

Embeddings

Convert text to vector representations for semantic search and RAG:
from cheapestinference import CheapestInference

client = CheapestInference()

embeddings = client.embeddings.create(
    model="BAAI/bge-large-en-v1.5",
    input=[
        "Artificial intelligence is transforming industries.",
        "Machine learning models require large datasets."
    ]
)

for embedding in embeddings.data:
    print(f"Vector dimension: {len(embedding.embedding)}")

Reranking

Improve search results with AI-powered reranking:
# Rerank search results
response = client.rerank.create(
    model="BAAI/bge-reranker-large",
    query="What is machine learning?",
    documents=[
        "Machine learning is a subset of AI...",
        "Python is a programming language...",
        "Neural networks are computing systems..."
    ]
)

for result in response.results:
    print(f"Document {result.index}: Score {result.score}")

Code Execution

Run Python code safely:
result = client.sandbox.execute(
    code="""
import pandas as pd
df = pd.DataFrame({'a': [1, 2, 3]})
print(df.sum())
    """,
    timeout=5
)

print(result.output)

Available Models

Embeddings

BGE, E5, and more

Reranking

Improve search relevance

Code Execution

Safe Python sandbox