CatoG commited on
Commit
8e8603a
·
verified ·
1 Parent(s): 09d1413

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +42 -4
app.py CHANGED
@@ -20,11 +20,49 @@ import pandas as pd
20
 
21
  # You can add/remove models here
22
  MODEL_CHOICES = [
23
- "distilgpt2", # small, good default
24
- "gpt2", # a bit larger
25
- "sshleifer/tiny-gpt2", # very tiny toy model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  ]
27
- DEFAULT_MODEL = "distilgpt2"
28
 
29
  device = 0 if torch.cuda.is_available() else -1
30
 
 
20
 
21
  # You can add/remove models here
22
  MODEL_CHOICES = [
23
+ # Very small / light (good for CPU Spaces)
24
+ "distilgpt2",
25
+ "gpt2",
26
+ "sshleifer/tiny-gpt2",
27
+ "LiquidAI/LFM2-350M",
28
+ "google/gemma-3-270m-it",
29
+ "Qwen/Qwen2.5-0.5B-Instruct",
30
+ "mkurman/NeuroBLAST-V3-SYNTH-EC-150000",
31
+
32
+ # Small–medium (~1–2B) – still reasonable on CPU, just slower
33
+ "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
34
+ "google/gemma-3-1b-it",
35
+ "meta-llama/Llama-3.2-1B",
36
+ "litert-community/Gemma3-1B-IT",
37
+ "nvidia/Nemotron-Flash-1B",
38
+ "WeiboAI/VibeThinker-1.5B",
39
+ "Qwen/Qwen3-1.7B",
40
+
41
+ # Medium (~2–3B) – probably OK on beefier CPU / small GPU
42
+ "google/gemma-2-2b-it",
43
+ "thu-pacman/PCMind-2.1-Kaiyuan-2B",
44
+ "opendatalab/MinerU-HTML", # 0.8B but more specialised, still fine
45
+ "ministral/Ministral-3b-instruct",
46
+ "HuggingFaceTB/SmolLM3-3B",
47
+ "meta-llama/Llama-3.2-3B-Instruct",
48
+ "nvidia/Nemotron-Flash-3B-Instruct",
49
+ "Qwen/Qwen2.5-3B-Instruct",
50
+
51
+ # Heavier (4–8B) – you really want a GPU Space for these
52
+ "Qwen/Qwen3-4B",
53
+ "Qwen/Qwen3-4B-Thinking-2507",
54
+ "Qwen/Qwen3-4B-Instruct-2507",
55
+ "mistralai/Mistral-7B-Instruct-v0.2",
56
+ "allenai/Olmo-3-7B-Instruct",
57
+ "Qwen/Qwen2.5-7B-Instruct",
58
+ "meta-llama/Meta-Llama-3-8B-Instruct",
59
+ "meta-llama/Llama-3.1-8B",
60
+ "meta-llama/Llama-3.1-8B-Instruct",
61
+ "openbmb/MiniCPM4.1-8B",
62
+ "deepseek-ai/DeepSeek-R1-Distill-Llama-8B",
63
+ "rl-research/DR-Tulu-8B",
64
  ]
65
+ DEFAULT_MODEL = "Qwen/Qwen2.5-0.5B-Instruct" # or TinyLlama, or stick with distilgpt2
66
 
67
  device = 0 if torch.cuda.is_available() else -1
68