Feature Description
- 🎨 Brand new UI interface (some interfaces are still pending updates)
- 🌍 Multi-language support (to be improved)
- 🎨 Added Midjourney-Proxy(Plus) API support
- 💰 Supports online top-up functionality, configurable in System Settings:
-
- Epay
-
- 🔍 Supports querying usage Quota by key:
- In conjunction with the project neko-api-key-tool, querying usage by key can be achieved.
- 📑 Pagination supports selecting the number of items displayed per page
- 🔄 SQLite database storage support, ready to use out of the box, lightweight and convenient
- 💵 Supports model billing by usage count, configurable in System Settings - Operation Settings
- ⚖️ Supports weighted random Channel selection
- 📈 Data Dashboard (Console)
- 🔒 Configurable models that a Token can call
- 🤖 Supports Telegram authorized login:
- System Settings - Configure Login & Registration - Allow login via Telegram
- Enter command /setdomain to @Botfather
- Select your bot, then enter http(s)://your_website_address/login
- The Telegram Bot Name is the bot username string without the @
- 🎵 Added Suno API API support
- 🔄 Supports Rerank models, currently compatible with Cohere and Jina, and can be integrated with Dify
- ⚡ OpenAI Realtime API - Supports OpenAI's Realtime API, supports Azure Channel
- Supports using the route /chat2link to enter the chat interface
- 🧠 Supports setting reasoning effort via model name suffix:
- OpenAI o-series models
-
- Add suffix
-highto set as high reasoning effort (e.g.:o3-mini-high)
- Add suffix
-
- Add suffix
-mediumto set as medium reasoning effort (e.g.:o3-mini-medium)
- Add suffix
-
- Add suffix
-lowto set as low reasoning effort (e.g.:o3-mini-low)
- Add suffix
-
- Claude thinking models
-
- Add suffix
-thinkingto enable thinking mode (e.g.:claude-3-7-sonnet-20250219-thinking)
- Add suffix
-
- OpenAI o-series models
- 🔄 Thinking to Content, supports setting the
thinking_to_contentoption inChannel - Edit - Channel Extra Settings, defaultfalse. When enabled, it will convert the thinking contentreasoning_contentinto a<think>tag and append it to the returned content. - 🔄 Model Rate Limiting, supports setting model rate limits in
System Settings - Rate Limit Settings, including total request count limit and successful request count limit. - 💰 Cache Billing Support, when enabled, billing can occur at a set Ratio upon cache hit:
- Set the Prompt Cache Ratio option in
System Settings - Operation Settings - Set the Prompt Cache Ratio in the Channel, range 0-1. For example, setting it to 0.5 means billing at 50% upon cache hit.
- Supported Channels:
-
- OpenAI
-
- Azure
-
- DeepSeek
-
- Claude
-
- Set the Prompt Cache Ratio option in
How is this guide?
Last updated on