Kibi

Agent & Developer Docs

Everything you need to build AI agents and integrations on KibiBot — API reference, CLI, Kibi LLM Gateway, and OpenClaw setup.

Overview

KibiBot exposes a full Agent API so your AI agent or application can create tokens, check balances, and call frontier AI models — all with a single kb_... API key.

Agent API

Create tokens, check quota, manage balances — REST endpoints at api.kibi.bot

CLI

Command-line tool for humans and CI pipelines — npm install -g @kibibot/cli

LLM Gateway

OpenAI-compatible AI model access billed to your Kibi Credits

# Agent API base URL

https://api.kibi.bot/agent/v1

# LLM Gateway base URL

https://llm.kibi.bot

# Auth header (all requests)

X-Api-Key: kb_...

Quick Start

Two ways to get started — pick what fits your workflow.

Before you start — get an API key

Go to kibi.bot/settings/api-keys → Create API Key → name it → copy the kb_... key.

Enable Kibi LLM Gateway if you need AI model access. Enable Agent Reload if you want the agent to top up credits from your trading wallet.

Option A — Install Agent Skill

Best for AI agent users (OpenClaw, etc.). The agent learns KibiBot's full API automatically.

Copy and paste this into your agent's chat:

install the kibibot skill from https://github.com/KibiAgent/skills/blob/main/kibibot/SKILL.md

Then hand it your API key and start giving commands in plain language.

View SKILL.md
Option B — Install CLI
View on npm ↗

Best for humans, scripts, and CI pipelines.

bash
npm install -g @kibibot/cli
kibi login        # enter your API key
kibi whoami       # verify connection
kibi token create # deploy your first token

All commands support --json for scripting.

CLI Reference

The @kibibot/cli package provides a kibi binary for humans and CI/CD pipelines. All commands support --json for machine-readable output.

Installation & AuthView on npm ↗

bash
npm install -g @kibibot/cli

# Authenticate
kibi login                  # prompts for API key
kibi logout                 # clear stored key
kibi whoami                 # show current identity

Token Commands

bash
# Create a token (interactive prompts if flags omitted)
kibi token create \
  --name "Moon Token" \
  --symbol MOON \
  --chain base \
  --description "To the moon" \
  --source https://x.com/user/status/123 \
  --image-url https://example.com/moon.png

# Wait for deployment (default: polls until complete)
# Use --no-wait to return job ID immediately

# Get token info
kibi token info 0xABC123... --chain base

# Check job status
kibi status 12345

# List tokens you've created
kibi tokens created

Balance & Quota

bash
# Show wallet balances across all chains
kibi balances

# Show daily token creation quota per chain
kibi quota

Fee Earnings

bash
# Summary across all chains
kibi fees

# Per-platform breakdown for a specific chain
kibi fees --chain bsc
kibi fees --chain base
kibi fees --chain solana

# Fee earnings for a specific token
kibi fees --chain bsc --platform flap --token 0x1234...
kibi fees --chain bsc --platform fourmeme --token 0x1234...
kibi fees --chain solana --platform pumpfun --token 5Yz8K...

Kibi Credits & LLM

bash
# Check Kibi Credit balance + agent reload config
kibi llm credits

# List available AI models
kibi llm models

# Reload Kibi Credits from trading wallet
kibi llm reload

# Emergency disable agent reload
kibi llm reload --disable

# Generate provider config for OpenClaw
kibi llm setup openclaw

Other Commands

bash
# List all agent capabilities
kibi skills

# Config management
kibi config get              # show all config
kibi config set apiKey kb_...

# Global flags (work on any command)
kibi <command> --json        # machine-readable JSON output
kibi <command> --debug       # verbose request/response logging
kibi --version

Agent API Reference

Base URL: https://api.kibi.bot/agent/v1 — Auth: X-Api-Key: kb_... header on all requests.

Identity

Token Creation

Quota

Balances

Fee Earnings API

Kibi LLM Gateway

The Kibi LLM Gateway is an OpenAI-compatible proxy billed to your Kibi Credits. Use the same kb_... API key — no separate account needed.

Base URL: https://llm.kibi.bot · Auth: X-Api-Key: kb_... or Authorization: Bearer kb_...

Supported Models

Model IDProviderContext
Anthropic
claude-opus-4-6Anthropic1M
claude-opus-4-5Anthropic200k
claude-sonnet-4-6Anthropic1M
claude-sonnet-4-5Anthropic1M
claude-haiku-4-5Anthropic200k
OpenAI
gpt-5.4-proOpenAI1.05M
gpt-5.4OpenAI1.05M
gpt-5.4-miniOpenAI400k
gpt-5.4-nanoOpenAI400k
gpt-5.2-proOpenAI400k
gpt-5.2OpenAI400k
gpt-5.2-codexOpenAI400k
gpt-5.2-chatOpenAI128k
Google
gemini-3.1-proGoogle1M
gemini-3.1-flash-liteGoogle1M
gemini-3-flashGoogle1M
gemini-2.5-proGoogle1M
gemini-2.5-flashGoogle1M
xAI
grok-4.1-fastxAI2M
DeepSeek
deepseek-v3.2DeepSeek164k
Moonshot
kimi-k2.5Moonshot262k
Xiaomi
mimo-v2-proXiaomi1M
mimo-v2-omniXiaomi262k
mimo-v2-flashXiaomi262k
ByteDance
seed-2.0-liteByteDance262k
seed-2.0-miniByteDance262k
Alibaba
qwen3-coderAlibaba262k
qwen3.5-plusAlibaba1M
qwen3.5-flashAlibaba1M
MiniMax
minimax-m2.7MiniMax205k
minimax-m2.5MiniMax197k
Z.ai
glm-5-turboZ.ai203k
glm-5Z.ai80k

OpenAI-compatible (all models)

bash
curl https://llm.kibi.bot/v1/chat/completions \
  -H "Content-Type: application/json" \
  -H "X-Api-Key: kb_YOUR_KEY" \
  -d '{
    "model": "gpt-5.4",
    "messages": [{"role": "user", "content": "Hello!"}],
    "max_tokens": 256
  }'

Anthropic-compatible (Claude only)

bash
curl https://llm.kibi.bot/v1/messages \
  -H "Content-Type: application/json" \
  -H "X-Api-Key: kb_YOUR_KEY" \
  -d '{
    "model": "claude-sonnet-4-6",
    "messages": [{"role": "user", "content": "Hello!"}],
    "max_tokens": 256
  }'

Other endpoints

bash
GET  /v1/models            # list all models
GET  /v1/models/openclaw   # ready-to-paste OpenClaw config (no auth)

Setup OpenClaw with Kibi LLM Gateway

Use Kibi as your OpenClaw agent's AI model provider — billed to your Kibi Credits instead of paying Anthropic/OpenAI directly. Also install the KibiBot skill so the agent can create tokens, check balances, and more.

Step 1 — Install the KibiBot skill

Ask your agent:

install the kibibot skill from https://github.com/KibiAgent/skills/blob/main/kibibot/SKILL.md and apply its OpenClaw config

Or browse the skill source: github.com/KibiAgent/skills

Step 2 — Add provider to openclaw.json

💡 If you ran the install command in Step 1 with "apply its OpenClaw config", this step was done automatically. Verify in your config file.

File: ~/.openclaw/openclaw.json

This is the full config with all available models. You only need to include the models you want to use.

json
{
  "models": {
    "mode": "merge",
    "providers": {
      "kibi": {
        "baseUrl": "https://llm.kibi.bot",
        "apiKey": "kb_YOUR_KEY",
        "api": "openai-completions",
        "models": [
          { "id": "kibi-opus-4-6",              "name": "Claude Opus 4.6",          "api": "anthropic-messages", "contextWindow": 1000000,  "maxTokens": 128000 },
          { "id": "kibi-opus-4-5",              "name": "Claude Opus 4.5",          "api": "anthropic-messages", "contextWindow": 200000,   "maxTokens": 64000  },
          { "id": "kibi-sonnet-4-6",            "name": "Claude Sonnet 4.6",        "api": "anthropic-messages", "contextWindow": 1000000,  "maxTokens": 128000 },
          { "id": "kibi-sonnet-4-5",            "name": "Claude Sonnet 4.5",        "api": "anthropic-messages", "contextWindow": 1000000,  "maxTokens": 64000  },
          { "id": "kibi-haiku-4-5",             "name": "Claude Haiku 4.5",         "api": "anthropic-messages", "contextWindow": 200000,   "maxTokens": 4096   },
          { "id": "kibi-gpt-5-4-pro",           "name": "GPT 5.4 Pro",              "contextWindow": 1050000,    "maxTokens": 16384 },
          { "id": "kibi-gpt-5-4",               "name": "GPT 5.4",                  "contextWindow": 1050000,    "maxTokens": 16384 },
          { "id": "kibi-gpt-5-4-mini",          "name": "GPT 5.4 Mini",             "contextWindow": 400000,     "maxTokens": 16384 },
          { "id": "kibi-gpt-5-4-nano",          "name": "GPT 5.4 Nano",             "contextWindow": 400000,     "maxTokens": 16384 },
          { "id": "kibi-gpt-5-2-pro",           "name": "GPT 5.2 Pro",              "contextWindow": 400000,     "maxTokens": 16384 },
          { "id": "kibi-gpt-5-2",               "name": "GPT 5.2",                  "contextWindow": 400000,     "maxTokens": 16384 },
          { "id": "kibi-gpt-5-2-codex",         "name": "GPT 5.2 Codex",            "contextWindow": 400000,     "maxTokens": 16384 },
          { "id": "kibi-gpt-5-2-chat",          "name": "GPT 5.2 Chat",             "contextWindow": 128000,     "maxTokens": 16384 },
          { "id": "kibi-gemini-3-1-pro",        "name": "Gemini 3.1 Pro",           "contextWindow": 1048576,    "maxTokens": 16384 },
          { "id": "kibi-gemini-3-1-flash-lite", "name": "Gemini 3.1 Flash Lite",    "contextWindow": 1048576,    "maxTokens": 16384 },
          { "id": "kibi-gemini-3-flash",        "name": "Gemini 3 Flash",           "contextWindow": 1048576,    "maxTokens": 16384 },
          { "id": "kibi-gemini-2-5-pro",        "name": "Gemini 2.5 Pro",           "contextWindow": 1048576,    "maxTokens": 8192  },
          { "id": "kibi-gemini-2-5-flash",      "name": "Gemini 2.5 Flash",         "contextWindow": 1048576,    "maxTokens": 8192  },
          { "id": "kibi-grok-4-1",              "name": "Grok 4.1 Fast",            "contextWindow": 2000000,    "maxTokens": 16384 },
          { "id": "kibi-deepseek-v3-2",         "name": "DeepSeek V3.2",            "contextWindow": 164000,     "maxTokens": 16384 },
          { "id": "kibi-kimi-k2-5",             "name": "Kimi K2.5",                "contextWindow": 262144,     "maxTokens": 16384 },
          { "id": "kibi-mimo-v2-pro",           "name": "MiMo-V2-Pro",              "contextWindow": 1048576,    "maxTokens": 16384 },
          { "id": "kibi-mimo-v2-omni",          "name": "MiMo-V2-Omni",             "contextWindow": 262144,     "maxTokens": 16384 },
          { "id": "kibi-mimo-v2-flash",         "name": "MiMo-V2-Flash",            "contextWindow": 262144,     "maxTokens": 16384 },
          { "id": "kibi-seed-2-0-lite",         "name": "Seed 2.0 Lite",            "contextWindow": 262144,     "maxTokens": 16384 },
          { "id": "kibi-seed-2-0-mini",         "name": "Seed 2.0 Mini",            "contextWindow": 262144,     "maxTokens": 16384 },
          { "id": "kibi-qwen-3-coder",          "name": "Qwen3 Coder",              "contextWindow": 262144,     "maxTokens": 16384 },
          { "id": "kibi-qwen-3-5-plus",         "name": "Qwen3.5 Plus",             "contextWindow": 1000000,    "maxTokens": 16384 },
          { "id": "kibi-qwen-3-5-flash",        "name": "Qwen3.5 Flash",            "contextWindow": 1000000,    "maxTokens": 16384 },
          { "id": "kibi-minimax-m2-7",          "name": "MiniMax M2.7",             "contextWindow": 204800,     "maxTokens": 16384 },
          { "id": "kibi-minimax-m2-5",          "name": "MiniMax M2.5",             "contextWindow": 196608,     "maxTokens": 16384 },
          { "id": "kibi-glm-5-turbo",           "name": "GLM 5 Turbo",              "contextWindow": 202752,     "maxTokens": 16384 },
          { "id": "kibi-glm-5",                 "name": "GLM 5",                    "contextWindow": 80000,      "maxTokens": 16384 }
        ]
      }
    }
  },
  "agents": {
    "defaults": {
      "models": {
        "kibi/kibi-opus-4-6":              { "alias": "kibi-opus-4-6" },
        "kibi/kibi-opus-4-5":              { "alias": "kibi-opus-4-5" },
        "kibi/kibi-sonnet-4-6":            { "alias": "kibi-sonnet-4-6" },
        "kibi/kibi-sonnet-4-5":            { "alias": "kibi-sonnet-4-5" },
        "kibi/kibi-haiku-4-5":             { "alias": "kibi-haiku-4-5" },
        "kibi/kibi-gpt-5-4-pro":           { "alias": "kibi-gpt-5-4-pro" },
        "kibi/kibi-gpt-5-4":               { "alias": "kibi-gpt-5-4" },
        "kibi/kibi-gpt-5-4-mini":          { "alias": "kibi-gpt-5-4-mini" },
        "kibi/kibi-gpt-5-4-nano":          { "alias": "kibi-gpt-5-4-nano" },
        "kibi/kibi-gpt-5-2-pro":           { "alias": "kibi-gpt-5-2-pro" },
        "kibi/kibi-gpt-5-2":               { "alias": "kibi-gpt-5-2" },
        "kibi/kibi-gpt-5-2-codex":         { "alias": "kibi-gpt-5-2-codex" },
        "kibi/kibi-gpt-5-2-chat":          { "alias": "kibi-gpt-5-2-chat" },
        "kibi/kibi-gemini-3-1-pro":        { "alias": "kibi-gemini-3-1-pro" },
        "kibi/kibi-gemini-3-1-flash-lite": { "alias": "kibi-gemini-3-1-flash-lite" },
        "kibi/kibi-gemini-3-flash":        { "alias": "kibi-gemini-3-flash" },
        "kibi/kibi-gemini-2-5-pro":        { "alias": "kibi-gemini-2-5-pro" },
        "kibi/kibi-gemini-2-5-flash":      { "alias": "kibi-gemini-2-5-flash" },
        "kibi/kibi-grok-4-1":              { "alias": "kibi-grok-4-1" },
        "kibi/kibi-deepseek-v3-2":         { "alias": "kibi-deepseek-v3-2" },
        "kibi/kibi-kimi-k2-5":             { "alias": "kibi-kimi-k2-5" },
        "kibi/kibi-mimo-v2-pro":           { "alias": "kibi-mimo-v2-pro" },
        "kibi/kibi-mimo-v2-omni":          { "alias": "kibi-mimo-v2-omni" },
        "kibi/kibi-mimo-v2-flash":         { "alias": "kibi-mimo-v2-flash" },
        "kibi/kibi-seed-2-0-lite":         { "alias": "kibi-seed-2-0-lite" },
        "kibi/kibi-seed-2-0-mini":         { "alias": "kibi-seed-2-0-mini" },
        "kibi/kibi-qwen-3-coder":          { "alias": "kibi-qwen-3-coder" },
        "kibi/kibi-qwen-3-5-plus":         { "alias": "kibi-qwen-3-5-plus" },
        "kibi/kibi-qwen-3-5-flash":        { "alias": "kibi-qwen-3-5-flash" },
        "kibi/kibi-minimax-m2-7":          { "alias": "kibi-minimax-m2-7" },
        "kibi/kibi-minimax-m2-5":          { "alias": "kibi-minimax-m2-5" },
        "kibi/kibi-glm-5-turbo":           { "alias": "kibi-glm-5-turbo" },
        "kibi/kibi-glm-5":                 { "alias": "kibi-glm-5" }
      }
    }
  }
}

Or run kibi llm setup openclaw to generate this automatically.

Step 3 — Set as default model (optional)

json
{
  "agents": {
    "defaults": {
      "model": { "primary": "kibi/kibi-sonnet-4-6" }
    }
  }
}

Step 4 — Restart and verify

bash
openclaw gateway restart

# Ask your agent:
# "what's my KibiBot Kibi Credit balance?"

Reload Kibi Credits with Agent

Agent Reload lets your agent top up its own Kibi Credits directly from your trading wallet USDC/USDT — without any human interaction.

This is a manual trigger — your agent decides when to call it. There is no background auto-reload watcher.

Setup (one-time, human)

  1. Go to kibi.bot/credits → Agent Reload section → enable it, set amount and daily limit.
  2. Go to kibi.bot/settings/api-keys → Edit Permissions on your key → enable Kibi LLM Gateway and Agent Reload.

How it works

1Agent calls POST /agent/v1/balance/credits/reload
2Backend checks key permission + user-level config
3Checks daily reload limit hasn't been hit
4Finds first configured chain with sufficient USDC/USDT in trading wallet
5Executes on-chain transfer from trading wallet → KibiBot treasury
6Credits Kibi Credit balance immediately
7Returns tx hash + new balance

Example flow

bash
# Check balance first
curl https://api.kibi.bot/agent/v1/balance/credits \
  -H "X-Api-Key: kb_YOUR_KEY"

# If low, reload
curl -X POST https://api.kibi.bot/agent/v1/balance/credits/reload \
  -H "X-Api-Key: kb_YOUR_KEY"

# Emergency stop (if agent misbehaves)
curl -X POST https://api.kibi.bot/agent/v1/balance/credits/reload/disable \
  -H "X-Api-Key: kb_YOUR_KEY"

Errors

ParameterTypeRequiredDescription
403NoReload not enabled for user or this API key
429NoDaily reload limit would be exceeded
400NoInsufficient USDC/USDT on all configured chains
500NoOn-chain transaction failed

Error Codes

All errors return JSON with a detail field describing what went wrong.

CodeMeaningCommon cause
401UnauthorizedMissing or invalid API key. Check kibi.bot/settings/api-keys.
402Payment RequiredInsufficient Kibi Credits (LLM) or trading wallet balance (token creation).
403ForbiddenFeature not enabled for this key or user — check key permissions.
404Not FoundResource doesn't exist or belongs to another user.
422Validation ErrorInvalid request body — check required fields and value constraints.
429Too Many RequestsRate limited or daily cap exceeded. Wait and retry.
500Internal Server ErrorServer-side failure — usually safe to retry after a short delay.
Agent Docs | Kibi.bot (BNB)