smart_toy
Phi Large Language Models

Phi-3 Medium 14B (14.00B)

Parameters
14.00B
VRAM (FP16)
28.0GB
VRAM (INT4)
7.0GB
Context
128000

tune Quantization Options

Quantization VRAM Required Min GPU
FP16 (Half Precision) 28.0GB A6000 / 2x 4090
INT8 (8-bit Integer) 14.0GB RTX 4080
Q4_K_M (GGUF 4-bit) 7.0GB RTX 3070 / 4060
q3_k_m 5.6GB RTX 3070 / 4060

Model Details

Family Phi
Category Large Language Models
Parameters 14.00B
Context Length 128000