Umans Code User Guide

Keep your agent working. All day.

Umans Code delivers a Claude Code-first agentic development experience with ultra-generous usage limits, powered by the best open-source models available. We currently serve Kimi K2.5 and MiniMax M2.5. We publish our model tests and reviews at blog.umans.ai.

Quick Start (Recommended)

The fastest way to get started is with the Umans CLI. It handles authentication and launches Claude Code with zero configuration.

macOS & Linux

bash
# Install the CLI (one-time)
curl -fsSL https://api.code.umans.ai/cli/install.sh | bash

# Launch Claude Code with Umans backend
umans claude

First run: The CLI opens your browser for authentication. Log in to your Umans account, and the CLI automatically receives your API key. Claude Code launches immediately with the Umans backend configured.

Subsequent runs: umans claude launches instantly using your saved credentials.

CLI Commands

bash
umans claude                          # Launch Claude Code (default: umans-coder)
umans claude --model umans-minimax-m2.5   # Use MiniMax M2.5 with vision/websearch handoffs
umans opencode                       # Launch OpenCode with Umans backend
umans status                         # Check authentication status
umans logout                         # Remove saved credentials
umans --help                         # Show all available commands

Manual Configuration (Alternative)

If the CLI does not work for your setup (Windows users, custom environments) or you prefer to configure tools manually, use these settings:

API Endpoint

SettingValue
Base URLhttps://api.code.umans.ai
Anthropic Endpointhttps://api.code.umans.ai/v1/messages
OpenAI Endpointhttps://api.code.umans.ai/v1/chat/completions
Model Nameumans-coder

Getting Your API Key

  1. Log in to app.umans.ai/billing
  2. Go to your Dashboard → API Keys
  3. Generate a new key (shown only once - copy it immediately)

Tool-Specific Setup

Claude Code Official Docs →

Using the CLI (Recommended):

bash
umans claude                    # Default: umans-coder (Kimi K2.5)
umans claude --model umans-minimax-m2.5   # MiniMax M2.5 with vision/websearch

Available Models:

ModelProviderCapabilitiesBest For
umans-coderKimi K2.5*Text, Vision, WebSearchDefault — we choose the best for you
umans-kimi-k2.5Kimi K2.5Text, Vision, WebSearchWhen you specifically want Kimi
umans-minimax-m2.5MiniMax M2.5Text, Vision (handoff), WebSearch (handoff)When you specifically want MiniMax

* Today, umans-coder routes to Kimi K2.5. This may change as we continuously evaluate models. See our model selection methodology at blog.umans.ai.

Manual configuration:

bash
export ANTHROPIC_BASE_URL=https://api.code.umans.ai
export ANTHROPIC_AUTH_TOKEN=sk-your-umans-api-key
claude --model umans-coder

OpenCode Official Docs →

Using the CLI (Recommended):

bash
umans opencode                          # Default: umans-coder
umans opencode --model umans-kimi-k2.5  # Use native Kimi K2.5

Manual configuration (add to ~/.opencode/config.json):

json
{
  "$schema": "https://opencode.ai/config.json",
  "provider": {
    "umans": {
      "npm": "@ai-sdk/anthropic",
      "name": "Umans coder",
      "options": {
        "baseURL": "https://api.code.umans.ai/v1",
        "apiKey": "sk-your-umans-api-key"
      },
      "models": {
        "umans-coder": {
          "name": "Umans coder"
        }
      }
    }
  }
}

Cursor IDE Official Docs →

  1. Open Cursor Settings → Models
  2. Enable Override OpenAI Base URL
  3. Set the base URL to: https://api.code.umans.ai/v1
  4. Paste your Umans API key in the API key field
  5. Add the custom model: umans-coder
  6. Select umans-coder in the model dropdown

Crush (Charm Bracelet) Official Docs →

Add to your Crush configuration ( ~/.config/crush/config.json ):

json
{
  "$schema": "https://charm.land/crush.json",
  "providers": {
    "umans": {
      "type": "anthropic",
      "base_url": "https://api.code.umans.ai",
      "api_key": "sk-your-umans-api-key",
      "models": [
        {
          "id": "umans-coder",
          "name": "Umans Coder",
          "default_max_tokens": 50000,
          "can_reason": true
        }
      ]
    }
  }
}

Any OpenAI-Compatible Tool

Umans Code exposes an OpenAI-compatible API. Configure any tool that supports custom OpenAI endpoints:

  • Base URL: https://api.code.umans.ai/v1
  • API Key: Your Umans API key (starts with sk-)
  • Model: umans-coder

API Reference

Anthropic-Compatible Endpoints

Umans Code implements the Anthropic Messages API.

POST /v1/messages

bash
curl -N -X POST https://api.code.umans.ai/v1/messages \
  -H "Content-Type: application/json" \
  -H "x-api-key: sk-your-umans-api-key" \
  -H "anthropic-version: 2023-06-01" \
  -d '{
    "model": "umans-coder",
    "messages": [{"role": "user", "content": "Hello!"}],
    "max_tokens": 4096,
    "stream": true
  }'

OpenAI-Compatible Endpoints

Umans Code also implements the OpenAI Chat Completions API.

POST /v1/chat/completions

bash
curl -N -X POST https://api.code.umans.ai/v1/chat/completions \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer sk-your-umans-api-key" \
  -d '{
    "model": "umans-coder",
    "messages": [{"role": "user", "content": "Hello!"}],
    "stream": true
  }'

Models

Our Philosophy: We believe in serving the best open-source models available. We continuously evaluate and filter models to ensure your agents stay productive all day—without the decision fatigue of choosing between dozens of options.

Available Models

ModelBaseBest ForTrade-off
umans-coderKimi K2.5*Default — we choose the best for youRoutes to our top pick (may change over time)
umans-kimi-k2.5Kimi K2.5When you specifically want KimiZero overhead, native multimodal
umans-minimax-m2.5MiniMax M2.5When you specifically want MiniMaxText model with Umans vision layer

* Today, umans-coder routes to Kimi K2.5. This may change as we continuously evaluate models. Read more at blog.umans.ai.

How to Choose

  • Use umans-coder (default) to let us choose the best model for you. We continuously evaluate and select what works best for most coding tasks. Today this is Kimi K2.5.
  • Use umans-kimi-k2.5 when you specifically want native Kimi K2.5. Overall best experience. It leads benchmarks on vision-heavy workflows and document understanding tasks.
  • Use umans-minimax-m2.5 when you specifically want MiniMax M2.5. Ideal for multi-language coding. It leads SWE-Bench Multilingual performance.

Benchmark Comparison

We believe in transparency. Select a benchmark below to see how our served models compare across different capabilities. Our models excel particularly in document understanding and multimodal tasks.

Multi-language software engineering (Java, JS, Go, C++, etc.).

February 2026
MiniMax M2.5 (umans-minimax-m2.5)
51.3%
Claude Opus 4.6
50.3%
Gemini 3 Pro
42.7%

🏆 State of the Art: On OmniDocBench, Kimi K2.5 achieves 88.8% — outperforming Gemini 3 Pro (88.5%), Claude Opus 4.5 (87.7%), and GPT-5.2 (85.7%) on document understanding and OCR tasks.

🚀 Best in Class Coding: MiniMax M2.5 leads on Multi-SWE-Bench with 51.3% (vs Claude Opus 4.6 at 50.3%) for multi-language coding (Java, Go, Rust, C++, etc.). Also achieves 80.2% on SWE-Bench Verified. Use umans-minimax-m2.5 for demanding software engineering tasks.

Sources: Moonshot AI, MiniMax, Anthropic, SWE-Bench Leaderboard, Qwen2.5-VL Technical Report

Note: Scores are from official model reports and independent evaluations. Different benchmarks test different capabilities—choose the one that matches your use case. Data available via /v1/models/info.

Model Information API

For programmatic access to current model information, including context windows, pricing, and capabilities:

bash
curl https://api.code.umans.ai/v1/models/info | jq

This public endpoint returns up-to-date information about all available models, their capabilities, and current pricing.

Troubleshooting

CLI Issues

"Command not found: umans"

  • Ensure ~/.local/bin or /usr/local/bin is in your PATH
  • Run source ~/.bashrc or source ~/.zshrc after installation

"Authentication failed"

  • Run umans logout to clear saved credentials
  • Run umans claude again to re-authenticate

Browser does not open

  • Copy the URL shown in the terminal and open it manually
  • The CLI displays a localhost callback URL - authentication will complete when you visit the URL

Connection Issues

"401 Unauthorized"

  • Your API key may be expired or revoked
  • Generate a new key in the Dashboard

"Rate limit exceeded"

  • You have hit your plan's usage limits
  • Check your usage in the Dashboard or upgrade your plan

Streaming interruptions

  • For long-running sessions, some networks may drop idle connections
  • Check your network stability or try a wired connection

Windows-Specific

The Umans CLI is not yet available for Windows. Use the manual configuration method with your preferred tool:

  1. Set environment variables in PowerShell:
    powershell
    $env:ANTHROPIC_BASE_URL="https://api.code.umans.ai"
    $env:ANTHROPIC_AUTH_TOKEN="sk-your-umans-api-key"
  2. Or configure directly in your tool's settings using the manual configuration values above

FAQ

What models does Umans Code use?

Umans Code serves the best open-source models available. We do the hard work of evaluating and selecting so you don't have to. Currently:

  • umans-coder — Our recommended default. We continuously evaluate and route to what works best (today: Kimi K2.5)
  • umans-kimi-k2.5 — Explicitly choose native Kimi K2.5 for vision-heavy workflows and document understanding
  • umans-minimax-m2.5 — Explicitly choose MiniMax M2.5 for multi-language coding and best-in-class SWE-Bench Multilingual performance

We publish our model evaluations and reviews at blog.umans.ai.

Can I use my own Claude Code license?

Yes. If you have a Claude Code subscription with Anthropic, you can use claude to run Claude Code with your Anthropic subscription. Use umans claude when you want to use Claude Code powered by Umans (best open-source model with unlimited tokens). Switch between them anytime.

Is my data secure?

Your code and conversations are processed through our infrastructure. We do not train on your data. Enterprise customers can opt for self-hosted deployments where all data remains within their infrastructure.

What happens if I hit my usage limit?

The API will return a rate limit error. You can monitor your usage in the Dashboard and upgrade your plan if needed. Limits reset according to your billing cycle.

Can I use the same API key for multiple machines?

Yes, but be aware that usage counts against your plan's limits across all usage. For team or multi-machine setups, consider the $50 plan with higher limits and parallel sessions.

Support

Need help?