Q4_K_M
372 GBMin VRAM: 427.8 GB
Recommended VRAM: 483.6 GB
Min RAM: 558 GB
Context: 8K / 8K
Loading model details...
Fetching variants, compatibility details, and metadata.
Share GLM 5 with someone who is deciding what to run locally.
Social proof
0% of 984 scanned PCs run GLM 5 fully on GPU.
194 keep at least some work on GPU. Based on anonymous compatibility checks.
General-purpose local model brief
Best for
Consider alternatives if
Quantization tip: Benchmark at least two quantizations and validate with a task-specific eval set before production use.
New to local models? Smaller quantization variants are easier to run, while larger ones can improve quality at the cost of more memory.
Q4_K_M
372 GBMin VRAM: 427.8 GB
Recommended VRAM: 483.6 GB
Min RAM: 558 GB
Context: 8K / 8K
Q5_K_M
465 GBMin VRAM: 534.8 GB
Recommended VRAM: 604.5 GB
Min RAM: 698 GB
Context: 8K / 8K
Q8_0
744 GBMin VRAM: 855.6 GB
Recommended VRAM: 967.2 GB
Min RAM: 1116 GB
Context: 8K / 8K
FP16
1488 GBMin VRAM: 1711.2 GB
Recommended VRAM: 1934.4 GB
Min RAM: 2232 GB
Context: 8K / 8K
| Quantization | File Size | Min VRAM | Recommended VRAM | Min RAM | Context |
|---|---|---|---|---|---|
| Q4_K_M | 372 GB | 427.8 GB | 483.6 GB | 558 GB | 8K / 8K |
| Q5_K_M | 465 GB | 534.8 GB | 604.5 GB | 698 GB | 8K / 8K |
| Q8_0 | 744 GB | 855.6 GB | 967.2 GB | 1116 GB | 8K / 8K |
| FP16 | 1488 GB | 1711.2 GB | 1934.4 GB | 2232 GB | 8K / 8K |