Spaces:
Running
Running
deepseek-ai models added
Browse files
app.py
CHANGED
@@ -17,7 +17,13 @@ llm_models = [
|
|
17 |
"meta-llama/Meta-Llama-3-8B-Instruct",
|
18 |
"mistralai/Mistral-7B-Instruct-v0.2",
|
19 |
"tiiuae/falcon-7b-instruct",
|
20 |
-
"deepseek-ai/DeepSeek-R1-
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
# "mistralai/Mixtral-8x22B-Instruct-v0.1", ## 281GB>10GB
|
22 |
# "NousResearch/Yarn-Mistral-7b-64k", ## 14GB>10GB
|
23 |
# "impira/layoutlm-document-qa", ## ERR
|
|
|
17 |
"meta-llama/Meta-Llama-3-8B-Instruct",
|
18 |
"mistralai/Mistral-7B-Instruct-v0.2",
|
19 |
"tiiuae/falcon-7b-instruct",
|
20 |
+
"deepseek-ai/DeepSeek-R1-Distill-Qwen-32B",
|
21 |
+
"deepseek-ai/deepseek-vl2",
|
22 |
+
"deepseek-ai/deepseek-vl2-small",
|
23 |
+
"deepseek-ai/deepseek-vl2-tiny",
|
24 |
+
"deepseek-ai/deepseek-llm-7b-chat",
|
25 |
+
"deepseek-ai/deepseek-math-7b-instruct",
|
26 |
+
# "deepseek-ai/DeepSeek-R1-Zero", ## 688GB > 10GB
|
27 |
# "mistralai/Mixtral-8x22B-Instruct-v0.1", ## 281GB>10GB
|
28 |
# "NousResearch/Yarn-Mistral-7b-64k", ## 14GB>10GB
|
29 |
# "impira/layoutlm-document-qa", ## ERR
|