-
-
-
-
-
-
Inference Providers
Active filters: autoawq
hugging-quants/Meta-Llama-3.1-8B-Instruct-AWQ-INT4
Text Generation
• Updated
• 170k
• 88
Text Generation
• 7B • Updated
• 668
• 2
Text Generation
• 6B • Updated
• 2
kaitchup/Llama-3-8b-awq-4bit
Text Generation
• 8B • Updated
• 1
XavierSpycy/Meta-Llama-3-8B-Instruct-zh-10k
Text Generation
• Updated
• 11
XavierSpycy/Meta-Llama-3-8B-Instruct-zh-10k-GGUF
Text Generation
• Updated
• 19
XavierSpycy/Meta-Llama-3-8B-Instruct-zh-10k-GPTQ
Text Generation
• Updated
• 2
XavierSpycy/Meta-Llama-3-8B-Instruct-zh-10k-AWQ
Text Generation
• Updated
• 2
hugging-quants/Meta-Llama-3.1-405B-Instruct-AWQ-INT4
Text Generation
• 410B • Updated
• 1.3k
• 36
hugging-quants/Meta-Llama-3.1-70B-Instruct-AWQ-INT4
Text Generation
• Updated
• 107k
• 107
jburmeister/Meta-Llama-3.1-70B-Instruct-AWQ-INT4
Text Generation
• 71B • Updated
• 4
jburmeister/Meta-Llama-3.1-405B-Instruct-AWQ-INT4
Text Generation
• 410B • Updated
• 3
Kalei/Meta-Llama-3.1-70B-Instruct-AWQ-INT4-Custom
Text Generation
• 71B • Updated
• 1
UCLA-EMC/Meta-Llama-3.1-8B-AWQ-INT4
Text Generation
• 8B • Updated
• 5
UCLA-EMC/Meta-Llama-3.1-8B-Instruct-AWQ-INT4-32-2.17B
Text Generation
• 8B • Updated
• 31
• 1
reach-vb/Meta-Llama-3.1-8B-Instruct-AWQ-INT4-fix
Text Generation
• 8B • Updated
• 4
jburmeister/Meta-Llama-3.1-8B-Instruct-AWQ-INT4
Text Generation
• 8B • Updated
• 3
awilliamson/Meta-Llama-3.1-70B-Instruct-AWQ
Text Generation
• 71B • Updated
• 3
flowaicom/Flow-Judge-v0.1-AWQ
Text Generation
• 4B • Updated
• 4.39k
• 6
hugging-quants/Mixtral-8x7B-Instruct-v0.1-AWQ-INT4
Text Generation
• 47B • Updated
• 1.94k
hugging-quants/gemma-2-9b-it-AWQ-INT4
Text Generation
• 9B • Updated
• 1.55k
• 8
ibnzterrell/Nvidia-Llama-3.1-Nemotron-70B-Instruct-HF-AWQ-INT4
Text Generation
• 71B • Updated
• 402
• 6
NeuML/Llama-3.1_OpenScholar-8B-AWQ
Text Generation
• 8B • Updated
• 84
• 3
fbaldassarri/TinyLlama_TinyLlama_v1.1-autoawq-int4-gs128-asym
Text Generation
• 1B • Updated
• 1
fbaldassarri/TinyLlama_TinyLlama_v1.1-autoawq-int4-gs128-sym
Text Generation
• 1B • Updated
fbaldassarri/EleutherAI_pythia-14m-autoawq-int4-gs128-asym
Text Generation
• 14.1M • Updated
• 1
fbaldassarri/EleutherAI_pythia-14m-autoawq-int4-gs128-sym
Text Generation
• 14.1M • Updated
fbaldassarri/EleutherAI_pythia-31m-autoawq-int4-gs128-asym
Text Generation
• 30.5M • Updated
fbaldassarri/EleutherAI_pythia-31m-autoawq-int4-gs128-sym
Text Generation
• 30.5M • Updated
fbaldassarri/EleutherAI_pythia-70m-deduped-autoawq-int4-gs128-asym
Text Generation
• 70.4M • Updated