turboderp-org
An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs
Updated 2026-03-13 03:31:19 +00:00
A fast inference library for running LLMs locally on modern consumer-class GPUs
Updated 2026-03-04 13:12:19 +00:00
Web UI for ExLlamaV2
Updated 2025-02-05 22:50:35 +00:00