-
-
-
-
-
-
Inference status
Active filters:
gptq
neuralmagic/Mistral-7B-Instruct-v0.3-GPTQ-4bit
Text Generation
•
Updated
•
1.24k
•
16
cookey39/Five_Phases_Mindset
Text Generation
•
Updated
•
61
•
1
Intel/Qwen2-1.5B-Instuct-int4-inc
Text Generation
•
Updated
•
25
•
1
Qwen/Qwen2-7B-Instruct-GPTQ-Int4
Text Generation
•
Updated
•
14.2k
•
23
neuralmagic/Meta-Llama-3-8B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
10.5k
•
3
neuralmagic/Meta-Llama-3-70B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
268
•
3
marcsun13/gemma-2-9b-it-GPTQ
Text Generation
•
Updated
•
1.07k
•
3
AI-MO/NuminaMath-7B-TIR-GPTQ
Text Generation
•
Updated
•
387
•
6
openerotica/WizardLM-2-8x22B-GPTQ-ERQ
Text Generation
•
Updated
•
23
•
1
pentagoniac/SEMIKONG-8b-GPTQ
Text Generation
•
Updated
•
433
•
22
model-scope/glm-4-9b-chat-GPTQ-Int8
Text Generation
•
Updated
•
203
•
2
ModelCloud/Mistral-Nemo-Instruct-2407-gptq-4bit
Text Generation
•
Updated
•
70.1k
•
4
rinna/llama-3-youko-8b-instruct-gptq
Text Generation
•
Updated
•
64
•
1
ModelCloud/gemma-2-27b-it-gptq-4bit
Text Generation
•
Updated
•
255
•
12
shuyuej/gemma-2-27b-it-GPTQ
Text Generation
•
Updated
•
814
•
3
shuyuej/Meta-Llama-3.1-8B-Instruct-GPTQ
Text Generation
•
Updated
•
126
•
1
hugging-quants/Meta-Llama-3.1-70B-Instruct-GPTQ-INT4
Text Generation
•
Updated
•
58.8k
•
22
hugging-quants/Meta-Llama-3.1-8B-Instruct-GPTQ-INT4
Text Generation
•
Updated
•
74k
•
21
TechxGenus/Mistral-Large-Instruct-2407-GPTQ
Text Generation
•
Updated
•
80
•
2
shuyuej/SFR-Embedding-2_R-GPTQ
Feature Extraction
•
Updated
•
82
•
1
shuyuej/bge-en-icl-GPTQ
Feature Extraction
•
Updated
•
29
•
2
shuyuej/gte-Qwen2-7B-instruct-GPTQ
Text Generation
•
Updated
•
73
•
2
neuralmagic/Meta-Llama-3.1-70B-Instruct-quantized.w4a16
Text Generation
•
Updated
•
117k
•
28
nejumi/calm3-22b-chat-GPTQ-Int4-calib-ja-1k
Text Generation
•
Updated
•
511
•
2
shuyuej/Mistral-Nemo-Instruct-2407-GPTQ-INT8
Updated
•
60
•
1
alpindale/Meta-Llama-3.1-70B-Instruct-GPTQ-INT8
Updated
•
42
•
2
team-hatakeyama-phase2/Tanuki-8x8B-dpo-v1.0-GPTQ-4bit
Text Generation
•
Updated
•
14.3k
•
3
Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int8
Image-Text-to-Text
•
Updated
•
7.14k
•
20
Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int4
Image-Text-to-Text
•
Updated
•
10.2k
•
18
Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int8
Image-Text-to-Text
•
Updated
•
3.28k
•
12