Supports entering the chat interface via /chat2link route
🧠 Supports setting reasoning effort via model name suffix:
OpenAI o-series models
Add suffix -high for high reasoning effort (e.g.: o3-mini-high)
Add suffix -medium for medium reasoning effort (e.g.: o3-mini-medium)
Add suffix -low for low reasoning effort (e.g.: o3-mini-low)
Claude reasoning models
Add suffix -thinking to enable reasoning mode (e.g.: claude-3-7-sonnet-20250219-thinking)
🔄 Reasoning-to-content: supports setting thinking_to_content in Channel - Edit - Channel Extra Settings, default is false. When enabled, reasoning content reasoning_content will be converted to <think> tag and appended to the content in the response.
🔄 Model rate limiting: supports setting model rate limits in System Settings - Rate Limit Settings, supports setting total request limit and successful request limit
💰 Cache billing support: after enabling, you can bill according to the set ratio when cache hits occur:
Set prompt cache ratio option in System Settings - Operation Settings
Set prompt cache ratio in channels, range 0-1, e.g. 0.5 means 50% billing on cache hit