File size: 1,648 Bytes
8f4d405
927854c
8f4d405
927854c
8f4d405
 
927854c
8f4d405
927854c
 
 
 
 
8f4d405
 
927854c
8f4d405
927854c
 
 
 
 
8f4d405
 
927854c
8f4d405
927854c
 
 
 
 
 
 
 
 
 
 
8f4d405
 
 
927854c
 
 
8f4d405
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
# models_config.py
# UPDATED: Novita AI API only - no local models
LLM_CONFIG = {
    "primary_provider": "novita_api",
    "models": {
        "reasoning_primary": {
            "model_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B:de-1a706eeafbf3ebc2",
            "task": "general_reasoning",
            "max_tokens": 4096,
            "temperature": 0.6,  # Recommended for DeepSeek-R1
            "top_p": 0.95,
            "force_reasoning_prefix": True,
            "is_chat_model": True
        },
        "classification_specialist": {
            "model_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B:de-1a706eeafbf3ebc2",
            "task": "intent_classification",
            "max_tokens": 512,
            "temperature": 0.5,  # Lower for consistency
            "top_p": 0.9,
            "force_reasoning_prefix": False,
            "is_chat_model": True
        },
        "safety_checker": {
            "model_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B:de-1a706eeafbf3ebc2",
            "task": "content_moderation",
            "max_tokens": 1024,
            "temperature": 0.5,
            "top_p": 0.9,
            "force_reasoning_prefix": False,
            "is_chat_model": True
        },
        "embedding_specialist": {
            "model_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B:de-1a706eeafbf3ebc2",
            "task": "embeddings",
            "note": "Embeddings via Novita API - may require special handling",
            "is_chat_model": True
        }
    },
    "routing_logic": {
        "strategy": "novita_api_only",
        "fallback_chain": [],
        "load_balancing": "single_endpoint"
    }
}