← back
Baseten

Baseten

85

High-performance LLM inference platform with an OpenAI-compatible endpoint — swap your base URL to access DeepSeek, GLM, Kimi, GPT OSS, and more.

API KeyPay-as-you-goJSONBot Blocking: NoneNo CCopenai-compatibleinferencedeepseekstreamingtool-callingvision

Pros

  • + Drop-in OpenAI SDK compatibility — just change base_url and api_key
  • + Supports structured outputs, tool calling, streaming, and vision out of the box
  • + Access to high-performance open models (DeepSeek, GLM, Kimi) at competitive per-token pricing

Cons

  • - Rate limits not publicly documented — must contact support for increases
  • - No published OpenAPI spec, so agents can't auto-discover endpoints

Integration

example.py
from openai import OpenAIimport os
client = OpenAI(    base_url="https://inference.baseten.co/v1",    api_key=os.environ["BASETEN_API_KEY"])
response = client.chat.completions.create(    model="deepseek-ai/DeepSeek-V3.1",    messages=[{"role": "user", "content": "Hello"}])print(response.choices[0].message.content)

Details

Category
LLM
Pricing
Pay-as-you-go
Free Tier
Free credits on sign-up for experimentation
Rate Limit
Varies by account tier (req/min + token/min limits, 429 on exceed)
Verified
2026-03-31

Score Breakdown

Base +50
OpenAPI Spec (No) 0
Auth Complexity (API Key) +10
Response Format (JSON) +10
Bot Blocking (None) +10
No Credit Card (Yes) +5
Completely Free (No) 0
Total 85