Umans Code User Guide
Keep your agent working. All day.
Umans Code delivers a Claude Code-first agentic development experience with ultra-generous usage limits, powered by the best open-source models available. We currently serve Kimi K2.5 and MiniMax M2.5. We publish our model tests and reviews at blog.umans.ai.
Quick Start (Recommended)
The fastest way to get started is with the Umans CLI. It handles authentication and launches Claude Code with zero configuration.
macOS & Linux
# Install the CLI (one-time)
curl -fsSL https://api.code.umans.ai/cli/install.sh | bash
# Launch Claude Code with Umans backend
umans claudeVideo Demo: CLI with Claude Code walkthrough
First run: The CLI opens your browser for authentication. Log in to your Umans account, and the CLI automatically receives your API key. Claude Code launches immediately with the Umans backend configured.
Subsequent runs: umans claude launches instantly using your saved credentials.
CLI Commands
umans claude # Launch Claude Code (default: umans-coder)
umans claude --model umans-minimax-m2.5 # Use MiniMax M2.5 with vision/websearch handoffs
umans opencode # Launch OpenCode with Umans backend
umans status # Check authentication status
umans logout # Remove saved credentials
umans --help # Show all available commandsManual Configuration (Alternative)
If the CLI does not work for your setup (Windows users, custom environments) or you prefer to configure tools manually, use these settings:
API Endpoint
| Setting | Value |
|---|---|
| Base URL | https://api.code.umans.ai |
| Anthropic Endpoint | https://api.code.umans.ai/v1/messages |
| OpenAI Endpoint | https://api.code.umans.ai/v1/chat/completions |
| Model Name | umans-coder |
Getting Your API Key
- Log in to app.umans.ai/billing
- Go to your Dashboard → API Keys
- Generate a new key (shown only once - copy it immediately)
Tool-Specific Setup
Claude Code Official Docs →
Using the CLI (Recommended):
umans claude # Default: umans-coder (Kimi K2.5)
umans claude --model umans-minimax-m2.5 # MiniMax M2.5 with vision/websearchAvailable Models:
| Model | Provider | Capabilities | Best For |
|---|---|---|---|
| umans-coder | Kimi K2.5* | Text, Vision, WebSearch | Default — we choose the best for you |
| umans-kimi-k2.5 | Kimi K2.5 | Text, Vision, WebSearch | When you specifically want Kimi |
| umans-minimax-m2.5 | MiniMax M2.5 | Text, Vision (handoff), WebSearch (handoff) | When you specifically want MiniMax |
* Today, umans-coder routes to Kimi K2.5. This may change as we continuously evaluate models. See our model selection methodology at blog.umans.ai.
Manual configuration:
export ANTHROPIC_BASE_URL=https://api.code.umans.ai
export ANTHROPIC_AUTH_TOKEN=sk-your-umans-api-key
claude --model umans-coderOpenCode Official Docs →
Using the CLI (Recommended):
umans opencode # Default: umans-coder
umans opencode --model umans-kimi-k2.5 # Use native Kimi K2.5Manual configuration (add to ~/.opencode/config.json):
{
"$schema": "https://opencode.ai/config.json",
"provider": {
"umans": {
"npm": "@ai-sdk/anthropic",
"name": "Umans coder",
"options": {
"baseURL": "https://api.code.umans.ai/v1",
"apiKey": "sk-your-umans-api-key"
},
"models": {
"umans-coder": {
"name": "Umans coder"
}
}
}
}
}Cursor IDE Official Docs →
Video Demo: Setting up Cursor with Umans Code
- Open Cursor Settings → Models
- Enable Override OpenAI Base URL
- Set the base URL to:
https://api.code.umans.ai/v1 - Paste your Umans API key in the API key field
- Add the custom model:
umans-coder - Select
umans-coderin the model dropdown
Crush (Charm Bracelet) Official Docs →
Add to your Crush configuration ( ~/.config/crush/config.json ):
{
"$schema": "https://charm.land/crush.json",
"providers": {
"umans": {
"type": "anthropic",
"base_url": "https://api.code.umans.ai",
"api_key": "sk-your-umans-api-key",
"models": [
{
"id": "umans-coder",
"name": "Umans Coder",
"default_max_tokens": 50000,
"can_reason": true
}
]
}
}
}Any OpenAI-Compatible Tool
Umans Code exposes an OpenAI-compatible API. Configure any tool that supports custom OpenAI endpoints:
- Base URL:
https://api.code.umans.ai/v1 - API Key: Your Umans API key (starts with
sk-) - Model:
umans-coder
API Reference
Anthropic-Compatible Endpoints
Umans Code implements the Anthropic Messages API.
POST /v1/messages
curl -N -X POST https://api.code.umans.ai/v1/messages \
-H "Content-Type: application/json" \
-H "x-api-key: sk-your-umans-api-key" \
-H "anthropic-version: 2023-06-01" \
-d '{
"model": "umans-coder",
"messages": [{"role": "user", "content": "Hello!"}],
"max_tokens": 4096,
"stream": true
}'OpenAI-Compatible Endpoints
Umans Code also implements the OpenAI Chat Completions API.
POST /v1/chat/completions
curl -N -X POST https://api.code.umans.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer sk-your-umans-api-key" \
-d '{
"model": "umans-coder",
"messages": [{"role": "user", "content": "Hello!"}],
"stream": true
}'Models
Our Philosophy: We believe in serving the best open-source models available. We continuously evaluate and filter models to ensure your agents stay productive all day—without the decision fatigue of choosing between dozens of options.
Available Models
| Model | Base | Best For | Trade-off |
|---|---|---|---|
| umans-coder | Kimi K2.5* | Default — we choose the best for you | Routes to our top pick (may change over time) |
| umans-kimi-k2.5 | Kimi K2.5 | When you specifically want Kimi | Zero overhead, native multimodal |
| umans-minimax-m2.5 | MiniMax M2.5 | When you specifically want MiniMax | Text model with Umans vision layer |
* Today, umans-coder routes to Kimi K2.5. This may change as we continuously evaluate models. Read more at blog.umans.ai.
How to Choose
- Use
umans-coder(default) to let us choose the best model for you. We continuously evaluate and select what works best for most coding tasks. Today this is Kimi K2.5. - Use
umans-kimi-k2.5when you specifically want native Kimi K2.5. Overall best experience. It leads benchmarks on vision-heavy workflows and document understanding tasks. - Use
umans-minimax-m2.5when you specifically want MiniMax M2.5. Ideal for multi-language coding. It leads SWE-Bench Multilingual performance.
Benchmark Comparison
We believe in transparency. Select a benchmark below to see how our served models compare across different capabilities. Our models excel particularly in document understanding and multimodal tasks.
Multi-language software engineering (Java, JS, Go, C++, etc.).
February 2026🏆 State of the Art: On OmniDocBench, Kimi K2.5 achieves 88.8% — outperforming Gemini 3 Pro (88.5%), Claude Opus 4.5 (87.7%), and GPT-5.2 (85.7%) on document understanding and OCR tasks.
🚀 Best in Class Coding: MiniMax M2.5 leads on Multi-SWE-Bench with 51.3% (vs Claude Opus 4.6 at 50.3%) for multi-language coding (Java, Go, Rust, C++, etc.). Also achieves 80.2% on SWE-Bench Verified. Use umans-minimax-m2.5 for demanding software engineering tasks.
Sources: Moonshot AI, MiniMax, Anthropic, SWE-Bench Leaderboard, Qwen2.5-VL Technical Report
Note: Scores are from official model reports and independent evaluations. Different benchmarks test different capabilities—choose the one that matches your use case. Data available via /v1/models/info.
Model Information API
For programmatic access to current model information, including context windows, pricing, and capabilities:
curl https://api.code.umans.ai/v1/models/info | jqThis public endpoint returns up-to-date information about all available models, their capabilities, and current pricing.
Troubleshooting
CLI Issues
"Command not found: umans"
- Ensure
~/.local/binor/usr/local/binis in your PATH - Run
source ~/.bashrcorsource ~/.zshrcafter installation
"Authentication failed"
- Run
umans logoutto clear saved credentials - Run
umans claudeagain to re-authenticate
Browser does not open
- Copy the URL shown in the terminal and open it manually
- The CLI displays a localhost callback URL - authentication will complete when you visit the URL
Connection Issues
"401 Unauthorized"
- Your API key may be expired or revoked
- Generate a new key in the Dashboard
"Rate limit exceeded"
- You have hit your plan's usage limits
- Check your usage in the Dashboard or upgrade your plan
Streaming interruptions
- For long-running sessions, some networks may drop idle connections
- Check your network stability or try a wired connection
Windows-Specific
The Umans CLI is not yet available for Windows. Use the manual configuration method with your preferred tool:
- Set environment variables in PowerShell:powershell
$env:ANTHROPIC_BASE_URL="https://api.code.umans.ai" $env:ANTHROPIC_AUTH_TOKEN="sk-your-umans-api-key" - Or configure directly in your tool's settings using the manual configuration values above
FAQ
What models does Umans Code use?
Umans Code serves the best open-source models available. We do the hard work of evaluating and selecting so you don't have to. Currently:
- umans-coder — Our recommended default. We continuously evaluate and route to what works best (today: Kimi K2.5)
- umans-kimi-k2.5 — Explicitly choose native Kimi K2.5 for vision-heavy workflows and document understanding
- umans-minimax-m2.5 — Explicitly choose MiniMax M2.5 for multi-language coding and best-in-class SWE-Bench Multilingual performance
We publish our model evaluations and reviews at blog.umans.ai.
Can I use my own Claude Code license?
Yes. If you have a Claude Code subscription with Anthropic, you can use claude to run Claude Code with your Anthropic subscription. Use umans claude when you want to use Claude Code powered by Umans (best open-source model with unlimited tokens). Switch between them anytime.
Is my data secure?
Your code and conversations are processed through our infrastructure. We do not train on your data. Enterprise customers can opt for self-hosted deployments where all data remains within their infrastructure.
What happens if I hit my usage limit?
The API will return a rate limit error. You can monitor your usage in the Dashboard and upgrade your plan if needed. Limits reset according to your billing cycle.
Can I use the same API key for multiple machines?
Yes, but be aware that usage counts against your plan's limits across all usage. For team or multi-machine setups, consider the $50 plan with higher limits and parallel sessions.
Support
Need help?