smart_toy
Mistral Large Language Models

Mixtral 8x22B (141.00B)

Parameters
141.00B
VRAM (FP16)
282.0GB
VRAM (INT4)
70.5GB
Context
65536

tune Quantization Options

Quantization VRAM Required Min GPU
FP16 (Half Precision) 282.0GB A100 / H100
INT8 (8-bit Integer) 141.0GB A100 / H100
Q4_K_M (GGUF 4-bit) 70.5GB A100 / H100
q3_k_m 56.4GB A100 / H100

Model Details

Family Mistral
Category Large Language Models
Parameters 141.00B
Context Length 65536