Skip to content

Model Configuration LLM

Configure AI model parameters and behavior.

Model Selection

yaml
channels:
  - name: main
    model: gpt-4o           # Default model
    models:                 # Available models
      - gpt-4o
      - gpt-4o-mini
      - gpt-4-turbo

Model Parameters

ParameterTypeRangeDescription
temperaturenumber0-2Creativity (higher = more random)
maxTokensnumber1-128000Max output tokens
topPnumber0-1Nucleus sampling
frequencyPenaltynumber-2 to 2Reduce repetition
presencePenaltynumber-2 to 2Encourage new topics
yaml
channels:
  - name: creative
    model: gpt-4o
    temperature: 1.2        # More creative
    maxTokens: 2048
    
  - name: precise
    model: gpt-4o
    temperature: 0.3        # More precise
    maxTokens: 4096

OpenAI

ModelContextBest For
gpt-4o128KGeneral, vision
gpt-4o-mini128KFast, cheap
o1 / o3-mini128KReasoning
gpt-4-turbo128KComplex tasks

Claude

ModelContextBest For
claude-sonnet-4200KLatest, best overall
claude-3-5-sonnet200KStrong performance
claude-3-5-haiku200KFast, cheap
claude-3-opus200KComplex tasks

Gemini

ModelContextBest For
gemini-2.5-flash1MLatest Flash
gemini-2.5-pro1MLatest Pro
gemini-2.0-flash1MFast, general
gemini-1.5-pro2MLong context

DeepSeek

ModelContextBest For
deepseek-chat64KV3, Chinese & coding
deepseek-reasoner64KR1 reasoning

Others

ModelProviderBest For
grok-3 / grok-3-minixAIGeneral, tools
mistral-large-latestMistralFlagship
llama-3.3-70b-versatileGroqUltra-fast inference
glm-4-plus / glm-4-flashZhipu AIChinese
qwen-max / qwen-plusQwenChinese

Per-Group Models

yaml
groups:
  123456789:
    channel: creative
    model: gpt-4o
  987654321:
    channel: precise
    model: gpt-4o-mini

Preset Override

yaml
# data/presets/coder.yaml
name: coder
model: gpt-4o
temperature: 0.2
maxTokens: 4096

Token Management

Token Usage

Monitor token usage in Web Panel → Statistics

yaml
context:
  maxMessages: 20        # Limit context length
  maxTokens: 8000        # Max context tokens

Next Steps

Released under the MIT License