Research_AI_Assistant / models_config.py
JatsTheAIGen's picture
Initial commit V1
66dbebd
raw
history blame
1.39 kB
# models_config.py
LLM_CONFIG = {
"primary_provider": "huggingface",
"models": {
"reasoning_primary": {
"model_id": "mistralai/Mistral-7B-Instruct-v0.2",
"task": "general_reasoning",
"max_tokens": 2000,
"temperature": 0.7,
"cost_per_token": 0.000015,
"fallback": "meta-llama/Llama-2-7b-chat-hf"
},
"embedding_specialist": {
"model_id": "sentence-transformers/all-MiniLM-L6-v2",
"task": "embeddings",
"vector_dimensions": 384,
"purpose": "semantic_similarity",
"cost_advantage": "90%_cheaper_than_primary"
},
"classification_specialist": {
"model_id": "cardiffnlp/twitter-roberta-base-emotion",
"task": "intent_classification",
"max_length": 512,
"specialization": "fast_inference",
"latency_target": "<100ms"
},
"safety_checker": {
"model_id": "unitary/unbiased-toxic-roberta",
"task": "content_moderation",
"confidence_threshold": 0.85,
"purpose": "bias_detection"
}
},
"routing_logic": {
"strategy": "task_based_routing",
"fallback_chain": ["primary", "fallback", "degraded_mode"],
"load_balancing": "round_robin_with_health_check"
}
}