smart_toy
Gemma Large Language Models

Gemma 2 27B (27.00B)

Parameters
27.00B
VRAM (FP16)
54.0GB
VRAM (INT4)
13.5GB
Context
8192

tune Quantization Options

Quantization VRAM Required Min GPU
FP16 (Half Precision) 54.0GB A100 / H100
INT8 (8-bit Integer) 27.0GB A6000 / 2x 4090
Q4_K_M (GGUF 4-bit) 13.5GB RTX 4080
q3_k_m 10.8GB RTX 3060 / 4070

Model Details

Family Gemma
Category Large Language Models
Parameters 27.00B
Context Length 8192