File size: 1,648 Bytes
8f4d405 927854c 8f4d405 927854c 8f4d405 927854c 8f4d405 927854c 8f4d405 927854c 8f4d405 927854c 8f4d405 927854c 8f4d405 927854c 8f4d405 927854c 8f4d405 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 |
# models_config.py
# UPDATED: Novita AI API only - no local models
LLM_CONFIG = {
"primary_provider": "novita_api",
"models": {
"reasoning_primary": {
"model_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B:de-1a706eeafbf3ebc2",
"task": "general_reasoning",
"max_tokens": 4096,
"temperature": 0.6, # Recommended for DeepSeek-R1
"top_p": 0.95,
"force_reasoning_prefix": True,
"is_chat_model": True
},
"classification_specialist": {
"model_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B:de-1a706eeafbf3ebc2",
"task": "intent_classification",
"max_tokens": 512,
"temperature": 0.5, # Lower for consistency
"top_p": 0.9,
"force_reasoning_prefix": False,
"is_chat_model": True
},
"safety_checker": {
"model_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B:de-1a706eeafbf3ebc2",
"task": "content_moderation",
"max_tokens": 1024,
"temperature": 0.5,
"top_p": 0.9,
"force_reasoning_prefix": False,
"is_chat_model": True
},
"embedding_specialist": {
"model_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B:de-1a706eeafbf3ebc2",
"task": "embeddings",
"note": "Embeddings via Novita API - may require special handling",
"is_chat_model": True
}
},
"routing_logic": {
"strategy": "novita_api_only",
"fallback_chain": [],
"load_balancing": "single_endpoint"
}
}
|