Agent & Developer Docs
Everything you need to build AI agents and integrations on KibiBot — API reference, CLI, Kibi LLM Gateway, and OpenClaw setup.
Overview
KibiBot exposes a full Agent API so your AI agent or application can create tokens, check balances, and call frontier AI models — all with a single kb_... API key.
Create tokens, check quota, manage balances — REST endpoints at api.kibi.bot
Command-line tool for humans and CI pipelines — npm install -g @kibibot/cli
OpenAI-compatible AI model access billed to your Kibi Credits
# Agent API base URL
https://api.kibi.bot/agent/v1
# LLM Gateway base URL
https://llm.kibi.bot
# Auth header (all requests)
X-Api-Key: kb_...
Quick Start
Two ways to get started — pick what fits your workflow.
Before you start — get an API key
Go to kibi.bot/settings/api-keys → Create API Key → name it → copy the kb_... key.
Enable Kibi LLM Gateway if you need AI model access. Enable Agent Reload if you want the agent to top up credits from your trading wallet.
Best for AI agent users (OpenClaw, etc.). The agent learns KibiBot's full API automatically.
Copy and paste this into your agent's chat:
Then hand it your API key and start giving commands in plain language.
View SKILL.mdBest for humans, scripts, and CI pipelines.
npm install -g @kibibot/cli
kibi login # enter your API key
kibi whoami # verify connection
kibi token create # deploy your first tokenAll commands support --json for scripting.
CLI Reference
The @kibibot/cli package provides a kibi binary for humans and CI/CD pipelines. All commands support --json for machine-readable output.
Installation & AuthView on npm ↗
npm install -g @kibibot/cli
# Authenticate
kibi login # prompts for API key
kibi logout # clear stored key
kibi whoami # show current identityToken Commands
# Create a token (interactive prompts if flags omitted)
kibi token create \
--name "Moon Token" \
--symbol MOON \
--chain base \
--description "To the moon" \
--source https://x.com/user/status/123 \
--image-url https://example.com/moon.png
# Wait for deployment (default: polls until complete)
# Use --no-wait to return job ID immediately
# Get token info
kibi token info 0xABC123... --chain base
# Check job status
kibi status 12345
# List tokens you've created
kibi tokens createdBalance & Quota
# Show wallet balances across all chains
kibi balances
# Show daily token creation quota per chain
kibi quotaFee Earnings
# Summary across all chains
kibi fees
# Per-platform breakdown for a specific chain
kibi fees --chain bsc
kibi fees --chain base
kibi fees --chain solana
# Fee earnings for a specific token
kibi fees --chain bsc --platform flap --token 0x1234...
kibi fees --chain bsc --platform fourmeme --token 0x1234...
kibi fees --chain solana --platform pumpfun --token 5Yz8K...Kibi Credits & LLM
# Check Kibi Credit balance + agent reload config
kibi llm credits
# List available AI models
kibi llm models
# Reload Kibi Credits from trading wallet
kibi llm reload
# Emergency disable agent reload
kibi llm reload --disable
# Generate provider config for OpenClaw
kibi llm setup openclawOther Commands
# List all agent capabilities
kibi skills
# Config management
kibi config get # show all config
kibi config set apiKey kb_...
# Global flags (work on any command)
kibi <command> --json # machine-readable JSON output
kibi <command> --debug # verbose request/response logging
kibi --versionAgent API Reference
Base URL: https://api.kibi.bot/agent/v1 — Auth: X-Api-Key: kb_... header on all requests.
Identity
Token Creation
Quota
Balances
Fee Earnings API
Kibi LLM Gateway
The Kibi LLM Gateway is an OpenAI-compatible proxy billed to your Kibi Credits. Use the same kb_... API key — no separate account needed.
Base URL: https://llm.kibi.bot · Auth: X-Api-Key: kb_... or Authorization: Bearer kb_...
Supported Models
| Model ID | Provider | Context |
|---|---|---|
| Anthropic | ||
| claude-opus-4-6 | Anthropic | 1M |
| claude-opus-4-5 | Anthropic | 200k |
| claude-sonnet-4-6 | Anthropic | 1M |
| claude-sonnet-4-5 | Anthropic | 1M |
| claude-haiku-4-5 | Anthropic | 200k |
| OpenAI | ||
| gpt-5.4-pro | OpenAI | 1.05M |
| gpt-5.4 | OpenAI | 1.05M |
| gpt-5.4-mini | OpenAI | 400k |
| gpt-5.4-nano | OpenAI | 400k |
| gpt-5.2-pro | OpenAI | 400k |
| gpt-5.2 | OpenAI | 400k |
| gpt-5.2-codex | OpenAI | 400k |
| gpt-5.2-chat | OpenAI | 128k |
| gemini-3.1-pro | 1M | |
| gemini-3.1-flash-lite | 1M | |
| gemini-3-flash | 1M | |
| gemini-2.5-pro | 1M | |
| gemini-2.5-flash | 1M | |
| xAI | ||
| grok-4.1-fast | xAI | 2M |
| DeepSeek | ||
| deepseek-v3.2 | DeepSeek | 164k |
| Moonshot | ||
| kimi-k2.5 | Moonshot | 262k |
| Xiaomi | ||
| mimo-v2-pro | Xiaomi | 1M |
| mimo-v2-omni | Xiaomi | 262k |
| mimo-v2-flash | Xiaomi | 262k |
| ByteDance | ||
| seed-2.0-lite | ByteDance | 262k |
| seed-2.0-mini | ByteDance | 262k |
| Alibaba | ||
| qwen3-coder | Alibaba | 262k |
| qwen3.5-plus | Alibaba | 1M |
| qwen3.5-flash | Alibaba | 1M |
| MiniMax | ||
| minimax-m2.7 | MiniMax | 205k |
| minimax-m2.5 | MiniMax | 197k |
| Z.ai | ||
| glm-5-turbo | Z.ai | 203k |
| glm-5 | Z.ai | 80k |
OpenAI-compatible (all models)
curl https://llm.kibi.bot/v1/chat/completions \
-H "Content-Type: application/json" \
-H "X-Api-Key: kb_YOUR_KEY" \
-d '{
"model": "gpt-5.4",
"messages": [{"role": "user", "content": "Hello!"}],
"max_tokens": 256
}'Anthropic-compatible (Claude only)
curl https://llm.kibi.bot/v1/messages \
-H "Content-Type: application/json" \
-H "X-Api-Key: kb_YOUR_KEY" \
-d '{
"model": "claude-sonnet-4-6",
"messages": [{"role": "user", "content": "Hello!"}],
"max_tokens": 256
}'Other endpoints
GET /v1/models # list all models
GET /v1/models/openclaw # ready-to-paste OpenClaw config (no auth)Setup OpenClaw with Kibi LLM Gateway
Use Kibi as your OpenClaw agent's AI model provider — billed to your Kibi Credits instead of paying Anthropic/OpenAI directly. Also install the KibiBot skill so the agent can create tokens, check balances, and more.
Step 1 — Install the KibiBot skill
Ask your agent:
Or browse the skill source: github.com/KibiAgent/skills
Step 2 — Add provider to openclaw.json
💡 If you ran the install command in Step 1 with "apply its OpenClaw config", this step was done automatically. Verify in your config file.
File: ~/.openclaw/openclaw.json
This is the full config with all available models. You only need to include the models you want to use.
{
"models": {
"mode": "merge",
"providers": {
"kibi": {
"baseUrl": "https://llm.kibi.bot",
"apiKey": "kb_YOUR_KEY",
"api": "openai-completions",
"models": [
{ "id": "kibi-opus-4-6", "name": "Claude Opus 4.6", "api": "anthropic-messages", "contextWindow": 1000000, "maxTokens": 128000 },
{ "id": "kibi-opus-4-5", "name": "Claude Opus 4.5", "api": "anthropic-messages", "contextWindow": 200000, "maxTokens": 64000 },
{ "id": "kibi-sonnet-4-6", "name": "Claude Sonnet 4.6", "api": "anthropic-messages", "contextWindow": 1000000, "maxTokens": 128000 },
{ "id": "kibi-sonnet-4-5", "name": "Claude Sonnet 4.5", "api": "anthropic-messages", "contextWindow": 1000000, "maxTokens": 64000 },
{ "id": "kibi-haiku-4-5", "name": "Claude Haiku 4.5", "api": "anthropic-messages", "contextWindow": 200000, "maxTokens": 4096 },
{ "id": "kibi-gpt-5-4-pro", "name": "GPT 5.4 Pro", "contextWindow": 1050000, "maxTokens": 16384 },
{ "id": "kibi-gpt-5-4", "name": "GPT 5.4", "contextWindow": 1050000, "maxTokens": 16384 },
{ "id": "kibi-gpt-5-4-mini", "name": "GPT 5.4 Mini", "contextWindow": 400000, "maxTokens": 16384 },
{ "id": "kibi-gpt-5-4-nano", "name": "GPT 5.4 Nano", "contextWindow": 400000, "maxTokens": 16384 },
{ "id": "kibi-gpt-5-2-pro", "name": "GPT 5.2 Pro", "contextWindow": 400000, "maxTokens": 16384 },
{ "id": "kibi-gpt-5-2", "name": "GPT 5.2", "contextWindow": 400000, "maxTokens": 16384 },
{ "id": "kibi-gpt-5-2-codex", "name": "GPT 5.2 Codex", "contextWindow": 400000, "maxTokens": 16384 },
{ "id": "kibi-gpt-5-2-chat", "name": "GPT 5.2 Chat", "contextWindow": 128000, "maxTokens": 16384 },
{ "id": "kibi-gemini-3-1-pro", "name": "Gemini 3.1 Pro", "contextWindow": 1048576, "maxTokens": 16384 },
{ "id": "kibi-gemini-3-1-flash-lite", "name": "Gemini 3.1 Flash Lite", "contextWindow": 1048576, "maxTokens": 16384 },
{ "id": "kibi-gemini-3-flash", "name": "Gemini 3 Flash", "contextWindow": 1048576, "maxTokens": 16384 },
{ "id": "kibi-gemini-2-5-pro", "name": "Gemini 2.5 Pro", "contextWindow": 1048576, "maxTokens": 8192 },
{ "id": "kibi-gemini-2-5-flash", "name": "Gemini 2.5 Flash", "contextWindow": 1048576, "maxTokens": 8192 },
{ "id": "kibi-grok-4-1", "name": "Grok 4.1 Fast", "contextWindow": 2000000, "maxTokens": 16384 },
{ "id": "kibi-deepseek-v3-2", "name": "DeepSeek V3.2", "contextWindow": 164000, "maxTokens": 16384 },
{ "id": "kibi-kimi-k2-5", "name": "Kimi K2.5", "contextWindow": 262144, "maxTokens": 16384 },
{ "id": "kibi-mimo-v2-pro", "name": "MiMo-V2-Pro", "contextWindow": 1048576, "maxTokens": 16384 },
{ "id": "kibi-mimo-v2-omni", "name": "MiMo-V2-Omni", "contextWindow": 262144, "maxTokens": 16384 },
{ "id": "kibi-mimo-v2-flash", "name": "MiMo-V2-Flash", "contextWindow": 262144, "maxTokens": 16384 },
{ "id": "kibi-seed-2-0-lite", "name": "Seed 2.0 Lite", "contextWindow": 262144, "maxTokens": 16384 },
{ "id": "kibi-seed-2-0-mini", "name": "Seed 2.0 Mini", "contextWindow": 262144, "maxTokens": 16384 },
{ "id": "kibi-qwen-3-coder", "name": "Qwen3 Coder", "contextWindow": 262144, "maxTokens": 16384 },
{ "id": "kibi-qwen-3-5-plus", "name": "Qwen3.5 Plus", "contextWindow": 1000000, "maxTokens": 16384 },
{ "id": "kibi-qwen-3-5-flash", "name": "Qwen3.5 Flash", "contextWindow": 1000000, "maxTokens": 16384 },
{ "id": "kibi-minimax-m2-7", "name": "MiniMax M2.7", "contextWindow": 204800, "maxTokens": 16384 },
{ "id": "kibi-minimax-m2-5", "name": "MiniMax M2.5", "contextWindow": 196608, "maxTokens": 16384 },
{ "id": "kibi-glm-5-turbo", "name": "GLM 5 Turbo", "contextWindow": 202752, "maxTokens": 16384 },
{ "id": "kibi-glm-5", "name": "GLM 5", "contextWindow": 80000, "maxTokens": 16384 }
]
}
}
},
"agents": {
"defaults": {
"models": {
"kibi/kibi-opus-4-6": { "alias": "kibi-opus-4-6" },
"kibi/kibi-opus-4-5": { "alias": "kibi-opus-4-5" },
"kibi/kibi-sonnet-4-6": { "alias": "kibi-sonnet-4-6" },
"kibi/kibi-sonnet-4-5": { "alias": "kibi-sonnet-4-5" },
"kibi/kibi-haiku-4-5": { "alias": "kibi-haiku-4-5" },
"kibi/kibi-gpt-5-4-pro": { "alias": "kibi-gpt-5-4-pro" },
"kibi/kibi-gpt-5-4": { "alias": "kibi-gpt-5-4" },
"kibi/kibi-gpt-5-4-mini": { "alias": "kibi-gpt-5-4-mini" },
"kibi/kibi-gpt-5-4-nano": { "alias": "kibi-gpt-5-4-nano" },
"kibi/kibi-gpt-5-2-pro": { "alias": "kibi-gpt-5-2-pro" },
"kibi/kibi-gpt-5-2": { "alias": "kibi-gpt-5-2" },
"kibi/kibi-gpt-5-2-codex": { "alias": "kibi-gpt-5-2-codex" },
"kibi/kibi-gpt-5-2-chat": { "alias": "kibi-gpt-5-2-chat" },
"kibi/kibi-gemini-3-1-pro": { "alias": "kibi-gemini-3-1-pro" },
"kibi/kibi-gemini-3-1-flash-lite": { "alias": "kibi-gemini-3-1-flash-lite" },
"kibi/kibi-gemini-3-flash": { "alias": "kibi-gemini-3-flash" },
"kibi/kibi-gemini-2-5-pro": { "alias": "kibi-gemini-2-5-pro" },
"kibi/kibi-gemini-2-5-flash": { "alias": "kibi-gemini-2-5-flash" },
"kibi/kibi-grok-4-1": { "alias": "kibi-grok-4-1" },
"kibi/kibi-deepseek-v3-2": { "alias": "kibi-deepseek-v3-2" },
"kibi/kibi-kimi-k2-5": { "alias": "kibi-kimi-k2-5" },
"kibi/kibi-mimo-v2-pro": { "alias": "kibi-mimo-v2-pro" },
"kibi/kibi-mimo-v2-omni": { "alias": "kibi-mimo-v2-omni" },
"kibi/kibi-mimo-v2-flash": { "alias": "kibi-mimo-v2-flash" },
"kibi/kibi-seed-2-0-lite": { "alias": "kibi-seed-2-0-lite" },
"kibi/kibi-seed-2-0-mini": { "alias": "kibi-seed-2-0-mini" },
"kibi/kibi-qwen-3-coder": { "alias": "kibi-qwen-3-coder" },
"kibi/kibi-qwen-3-5-plus": { "alias": "kibi-qwen-3-5-plus" },
"kibi/kibi-qwen-3-5-flash": { "alias": "kibi-qwen-3-5-flash" },
"kibi/kibi-minimax-m2-7": { "alias": "kibi-minimax-m2-7" },
"kibi/kibi-minimax-m2-5": { "alias": "kibi-minimax-m2-5" },
"kibi/kibi-glm-5-turbo": { "alias": "kibi-glm-5-turbo" },
"kibi/kibi-glm-5": { "alias": "kibi-glm-5" }
}
}
}
}Or run kibi llm setup openclaw to generate this automatically.
Step 3 — Set as default model (optional)
{
"agents": {
"defaults": {
"model": { "primary": "kibi/kibi-sonnet-4-6" }
}
}
}Step 4 — Restart and verify
openclaw gateway restart
# Ask your agent:
# "what's my KibiBot Kibi Credit balance?"Reload Kibi Credits with Agent
Agent Reload lets your agent top up its own Kibi Credits directly from your trading wallet USDC/USDT — without any human interaction.
This is a manual trigger — your agent decides when to call it. There is no background auto-reload watcher.
Setup (one-time, human)
- Go to kibi.bot/credits → Agent Reload section → enable it, set amount and daily limit.
- Go to kibi.bot/settings/api-keys → Edit Permissions on your key → enable Kibi LLM Gateway and Agent Reload.
How it works
Example flow
# Check balance first
curl https://api.kibi.bot/agent/v1/balance/credits \
-H "X-Api-Key: kb_YOUR_KEY"
# If low, reload
curl -X POST https://api.kibi.bot/agent/v1/balance/credits/reload \
-H "X-Api-Key: kb_YOUR_KEY"
# Emergency stop (if agent misbehaves)
curl -X POST https://api.kibi.bot/agent/v1/balance/credits/reload/disable \
-H "X-Api-Key: kb_YOUR_KEY"Errors
| Parameter | Type | Required | Description |
|---|---|---|---|
| 403 | No | Reload not enabled for user or this API key | |
| 429 | No | Daily reload limit would be exceeded | |
| 400 | No | Insufficient USDC/USDT on all configured chains | |
| 500 | No | On-chain transaction failed |
Error Codes
All errors return JSON with a detail field describing what went wrong.
| Code | Meaning | Common cause |
|---|---|---|
| 401 | Unauthorized | Missing or invalid API key. Check kibi.bot/settings/api-keys. |
| 402 | Payment Required | Insufficient Kibi Credits (LLM) or trading wallet balance (token creation). |
| 403 | Forbidden | Feature not enabled for this key or user — check key permissions. |
| 404 | Not Found | Resource doesn't exist or belongs to another user. |
| 422 | Validation Error | Invalid request body — check required fields and value constraints. |
| 429 | Too Many Requests | Rate limited or daily cap exceeded. Wait and retry. |
| 500 | Internal Server Error | Server-side failure — usually safe to retry after a short delay. |