Files
ik_llama.cpp/github-data/pull_requests/40 - Adding bf16 support to CUDA.md
2025-07-23 13:31:53 +02:00

19 lines
773 B
Markdown

### 🔀 [#40](https://github.com/ikawrakow/ik_llama.cpp/pull/40) - Adding bf16 support to CUDA
| **Author** | `ikawrakow` |
| :--- | :--- |
| **State** | ❌ **Closed** |
| **Created** | 2024-09-05 |
| **Updated** | 2024-09-14 |
---
#### Description
Haha, `llama.cpp` seems to not support `bf16` on CUDA?
This PR adds it. It works fine on my `RTX-4080`, but I have no idea if it will work on older GPUs (if I understood correctly it should, with reduced performance), ROCm, etc.
Performance is the same as `f16` for TG (TG-128 = 41.2 t/s for LLaMA-3.1-8B for both).
PP is lower but quite decent for prompt processing (PP-512(`bf16`) = 5250 t/s vs PP-512(`f16`) = 7250 t/s for LLaMA-3.1-8B). In any case, much better than running on the CPU for `bf16` models.