-
-
-
-
-
-
Inference Providers
Active filters:
torchao
pytorch/gemma-3-27b-it-AWQ-INT4
Image-Text-to-Text
•
Updated
•
185k
•
3
jerryzh168/llama3-int4wo-128
Updated
medmekk/Meta-Llama-3-8B-quantized-int8_weight_only
Updated
medmekk/Meta-Llama-3-8B-quantized-int8_dynamic_activation_int8_weight
Updated
medmekk/Meta-Llama-3-8B-quantized-int4_weight_only
Updated
medmekk/Meta-Llama-3-8B-quantized-int8_weight_only-2
Updated
medmekk/Meta-Llama-3-8B-quantized-int4_weight_only-2
Updated
medmekk/Meta-Llama-3-8B-torchao-int4_weight_only-gs-64
Updated
medmekk/Meta-Llama-3-8B-torchao-int4_weight_only-gs-32
medmekk/Meta-Llama-3-8B-torchao-int4_weight_only-gs_256
Updated
medmekk/Meta-Llama-3-8B-torchao-int8_weight_only
Updated
medmekk/Meta-Llama-3-8B-torchao-int8_dynamic_activation_int8_weight
medmekk/gpt2-torchao-int8_weight_only
Updated
medmekk/Llama-3.1-70B-torchao-int8_weight_only
Updated
medmekk/an_other_torchao_dynamic
Updated
marcsun13/Meta-Llama-3-8B-torchao-int8_weight_only
Updated
medmekk/Meta-Llama-3-8B-torchao-int4_weight_only-gs_128
medmekk/deepseek-coder-1.3b-base-torchao-int8_weight_only
medmekk/testing_repo_name
Updated