Commit Graph

  • 33a90b66b3 minor : warning fixes Georgi Gerganov 2023-06-17 20:24:11 +03:00
  • 4f9c43e3bd minor : warning fixes Georgi Gerganov 2023-06-17 20:24:11 +03:00
  • 6c85db305e Only one CUDA stream per device for async compute (#1898) Johannes Gäßler 2023-06-17 19:15:02 +02:00
  • 2c9380dd2f Only one CUDA stream per device for async compute (#1898) Johannes Gäßler 2023-06-17 19:15:02 +02:00
  • 6c1607e1dc llama : fix kv_cache n init (close #1903) Georgi Gerganov 2023-06-17 19:30:22 +03:00
  • 051e1b0e6a llama : fix kv_cache n init (close #1903) Georgi Gerganov 2023-06-17 19:30:22 +03:00
  • d005fe3f4f make : update for latest Arch (#1701) DaniAndTheWeb 2023-06-17 18:17:22 +02:00
  • 86c7571864 make : update for latest Arch (#1701) DaniAndTheWeb 2023-06-17 18:17:22 +02:00
  • 49fdf2d578 ggml : fix warnings under MSVC (#1908) Howard Su 2023-06-17 23:46:15 +08:00
  • 3d59ec5935 ggml : fix warnings under MSVC (#1908) Howard Su 2023-06-17 23:46:15 +08:00
  • 058ca1ef4d metal : add norm, cpy f16->f16, alibi kernels (#1823) Aaron Miller 2023-06-17 07:37:49 -07:00
  • 0711a5f6dc metal : add norm, cpy f16->f16, alibi kernels (#1823) Aaron Miller 2023-06-17 07:37:49 -07:00
  • fd7e289fca exposed modules so that they can be invoked by nix run github:ggerganov/llama.cpp#server etc (#1863) Faez Shakil 2023-06-17 17:13:05 +05:00
  • fc45a81bc6 exposed modules so that they can be invoked by nix run github:ggerganov/llama.cpp#server etc (#1863) Faez Shakil 2023-06-17 17:13:05 +05:00
  • 8ccf4e1c32 Server Example Refactor and Improvements (#1570) Randall Fitzgerald 2023-06-17 07:53:04 -04:00
  • 794db3e7b9 Server Example Refactor and Improvements (#1570) Randall Fitzgerald 2023-06-17 07:53:04 -04:00
  • c05873c91d hooks : setting up flake8 and pre-commit hooks (#1681) Jiří Podivín 2023-06-17 12:32:48 +02:00
  • 5ddf7ea1fb hooks : setting up flake8 and pre-commit hooks (#1681) Jiří Podivín 2023-06-17 12:32:48 +02:00
  • 0db3668c98 readme : alternative way to build for Android with CLBlast. (#1828) Gustavo Rocha Dias 2023-06-17 06:01:06 -03:00
  • bac19927c3 readme : alternative way to build for Android with CLBlast. (#1828) Gustavo Rocha Dias 2023-06-17 06:01:06 -03:00
  • 77bb001e67 Allow cmake to build ggml as a library (#1896) Kerfuffle 2023-06-17 01:49:42 -06:00
  • b4c6f46f17 Allow cmake to build ggml as a library (#1896) Kerfuffle 2023-06-17 01:49:42 -06:00
  • 7c55238339 train : get raw text instead of page with html (#1905) David Yang 2023-06-17 14:51:54 +08:00
  • 92f20d9942 train : get raw text instead of page with html (#1905) David Yang 2023-06-17 14:51:54 +08:00
  • d7df87f386 opencl : support k-quants (#1836) 0cc4m 2023-06-16 20:59:49 +02:00
  • d411968e99 opencl : support k-quants (#1836) 0cc4m 2023-06-16 20:59:49 +02:00
  • 49571b542f examples : add "simple" (#1840) SuperUserNameMan 2023-06-16 20:58:09 +02:00
  • b41b4cad6f examples : add "simple" (#1840) SuperUserNameMan 2023-06-16 20:58:09 +02:00
  • caeb4ff61b cmake : add auto detection of BLAS_INCLUDE_DIRS (#1886) Zenix 2023-06-17 03:53:04 +09:00
  • 13fe9d2d84 cmake : add auto detection of BLAS_INCLUDE_DIRS (#1886) Zenix 2023-06-17 03:53:04 +09:00
  • d66c754bea llama : fix embd when offloading non-repeating layers (#1891) Johannes Gäßler 2023-06-16 20:25:51 +02:00
  • ac3b886953 llama : fix embd when offloading non-repeating layers (#1891) Johannes Gäßler 2023-06-16 20:25:51 +02:00
  • 338a9e5938 Fixed possible macro redefinition (#1892) FrankHB 2023-06-17 02:25:01 +08:00
  • 5b9ccaf104 Fixed possible macro redefinition (#1892) FrankHB 2023-06-17 02:25:01 +08:00
  • 96f634dfed build : fix and ignore MSVC warnings (#1889) Borislav Stanimirov 2023-06-16 21:23:53 +03:00
  • 9cbf50c041 build : fix and ignore MSVC warnings (#1889) Borislav Stanimirov 2023-06-16 21:23:53 +03:00
  • 3ac04c5f96 CUDA : faster k-quant dot kernels (#1862) Kawrakow 2023-06-16 20:08:44 +03:00
  • 3d01122610 CUDA : faster k-quant dot kernels (#1862) Kawrakow 2023-06-16 20:08:44 +03:00
  • 190b17eb88 gitignore : add several entries specific to Visual Studio (#1888) Borislav Stanimirov 2023-06-16 09:58:11 +03:00
  • 602c748863 gitignore : add several entries specific to Visual Studio (#1888) Borislav Stanimirov 2023-06-16 09:58:11 +03:00
  • 6acf0d3549 Fixed CUDA runtime version check (#1879) Johannes Gäßler 2023-06-15 21:49:08 +02:00
  • a09f9195be Fixed CUDA runtime version check (#1879) Johannes Gäßler 2023-06-15 21:49:08 +02:00
  • e5798d67ac cmake : remove whitespaces Georgi Gerganov 2023-06-15 21:56:50 +03:00
  • bed9275617 cmake : remove whitespaces Georgi Gerganov 2023-06-15 21:56:50 +03:00
  • bc2a0a3d4a examples : add chat-vicuna.sh (#1854) yangli2 2023-06-15 11:05:53 -07:00
  • c36e81da62 examples : add chat-vicuna.sh (#1854) yangli2 2023-06-15 11:05:53 -07:00
  • bb406c5c8b cmake : set include path for OpenBlas (#1830) Igor Okulist 2023-06-15 12:51:26 -05:00
  • 3559433fec cmake : set include path for OpenBlas (#1830) Igor Okulist 2023-06-15 12:51:26 -05:00
  • d050d6ae67 swift : Package compile breaks due to ggml-metal.metal (#1831) Frederik Vogel 2023-06-16 02:47:04 +09:00
  • 69b34a0e80 swift : Package compile breaks due to ggml-metal.metal (#1831) Frederik Vogel 2023-06-16 02:47:04 +09:00
  • 741a6fed44 make : add train-text-from-scratch (#1850) daboe01 2023-06-15 19:42:48 +02:00
  • cf267d1c71 make : add train-text-from-scratch (#1850) daboe01 2023-06-15 19:42:48 +02:00
  • 2f41c7c27c readme : server compile flag (#1874) Srinivas Billa 2023-06-15 18:36:38 +01:00
  • 9dda13e5e1 readme : server compile flag (#1874) Srinivas Billa 2023-06-15 18:36:38 +01:00
  • 8fd9583b2c make : clean *.so files (#1857) sandyiscool 2023-06-15 23:06:06 +05:30
  • 37e257c48e make : clean *.so files (#1857) sandyiscool 2023-06-15 23:06:06 +05:30
  • 9af138537f Fix the validation of main device (#1872) Howard Su 2023-06-16 01:29:59 +08:00
  • 64cc19b4fe Fix the validation of main device (#1872) Howard Su 2023-06-16 01:29:59 +08:00
  • feaa009626 metal : parallel command buffer encoding (#1860) Georgi Gerganov 2023-06-15 20:29:48 +03:00
  • 4bfcc855ab metal : parallel command buffer encoding (#1860) Georgi Gerganov 2023-06-15 20:29:48 +03:00
  • da3f715abb Better error when using both LoRA + GPU layers (#1861) Johannes Gäßler 2023-06-15 19:06:46 +02:00
  • 6b8312e797 Better error when using both LoRA + GPU layers (#1861) Johannes Gäßler 2023-06-15 19:06:46 +02:00
  • 4557646a38 CUDA full GPU acceleration, KV cache in VRAM (#1827) Johannes Gäßler 2023-06-14 19:47:19 +02:00
  • 254a7a7a5f CUDA full GPU acceleration, KV cache in VRAM (#1827) Johannes Gäßler 2023-06-14 19:47:19 +02:00
  • 1133a63542 baby-llama : fix operator!= (#1821) 0xspringtime 2023-06-13 15:37:54 -04:00
  • 9254920265 baby-llama : fix operator!= (#1821) 0xspringtime 2023-06-13 15:37:54 -04:00
  • d49334df42 train : improved training-from-scratch example (#1652) xaedes 2023-06-13 21:04:40 +02:00
  • e32089b2c2 train : improved training-from-scratch example (#1652) xaedes 2023-06-13 21:04:40 +02:00
  • 499e22902e llama : do a warm-up eval at start for better timings (#1824) Georgi Gerganov 2023-06-13 20:20:07 +03:00
  • 2347e45e7b llama : do a warm-up eval at start for better timings (#1824) Georgi Gerganov 2023-06-13 20:20:07 +03:00
  • 5416059be8 Allow "quantizing" to f16 and f32 (#1787) Kerfuffle 2023-06-13 04:23:23 -06:00
  • 74d4cfa343 Allow "quantizing" to f16 and f32 (#1787) Kerfuffle 2023-06-13 04:23:23 -06:00
  • 44af4a2f75 Metal implementation for all k_quants (#1807) Kawrakow 2023-06-12 22:39:21 +03:00
  • 74a6d922f1 Metal implementation for all k_quants (#1807) Kawrakow 2023-06-12 22:39:21 +03:00
  • 1200071552 ci : run when changing only the CUDA sources (#1800) slaren 2023-06-12 19:12:47 +02:00
  • e4caa8da59 ci : run when changing only the CUDA sources (#1800) slaren 2023-06-12 19:12:47 +02:00
  • df687e822c Leverage mmap for offloading tensors to GPU (#1597) Howard Su 2023-06-12 20:44:16 +08:00
  • 58970a4c39 Leverage mmap for offloading tensors to GPU (#1597) Howard Su 2023-06-12 20:44:16 +08:00
  • 6c7fc1cc50 metal : fix failure to load model (#1817) Kawrakow 2023-06-12 14:31:36 +03:00
  • 8c0a10e64d metal : fix failure to load model (#1817) Kawrakow 2023-06-12 14:31:36 +03:00
  • 1ca9832378 Fix issue where interactive mode crashes when input exceeds ctx size (#1789) Kerfuffle 2023-06-11 08:19:17 -06:00
  • fa84c4b3e8 Fix issue where interactive mode crashes when input exceeds ctx size (#1789) Kerfuffle 2023-06-11 08:19:17 -06:00
  • 6b4bca53f7 Fixed WSL cuda's OOM error (#1594) Kyle Liang 2023-06-11 21:20:52 +08:00
  • 12b063f0ec Fixed WSL cuda's OOM error (#1594) Kyle Liang 2023-06-11 21:20:52 +08:00
  • eb97aa37ad Update SHA256SUMS with current hashes for models quantized using q4_0 (#1798) Ryan Landay 2023-06-11 17:38:53 +08:00
  • 31d2b5f4a4 Update SHA256SUMS with current hashes for models quantized using q4_0 (#1798) Ryan Landay 2023-06-11 17:38:53 +08:00
  • 2069d67d2b cmake : fix Metal build (close #1791) Georgi Gerganov 2023-06-10 22:56:53 +03:00
  • 4de0334f5c cmake : fix Metal build (close #1791) Georgi Gerganov 2023-06-10 22:56:53 +03:00
  • 378a3a814d k-quants : GCC12 compilation fix (#1792) Artyom Lebedev 2023-06-10 22:51:36 +03:00
  • 3f1223155a k-quants : GCC12 compilation fix (#1792) Artyom Lebedev 2023-06-10 22:51:36 +03:00
  • c6755e854f metal : fix issue with ggml-metal.metal path. Closes #1769 (#1782) Andrei 2023-06-10 10:47:34 -04:00
  • 303f5809f1 metal : fix issue with ggml-metal.metal path. Closes #1769 (#1782) Andrei 2023-06-10 10:47:34 -04:00
  • 5e19207f69 doc : fix wrong address of BLIS.md (#1772) Aisuko 2023-06-11 00:08:11 +10:00
  • 059e99066d doc : fix wrong address of BLIS.md (#1772) Aisuko 2023-06-11 00:08:11 +10:00
  • bbd64dc9df ggml : force no_alloc == false when creating opt tensors (close #1699) Georgi Gerganov 2023-06-10 12:06:45 +03:00
  • 17c10acfb4 ggml : force no_alloc == false when creating opt tensors (close #1699) Georgi Gerganov 2023-06-10 12:06:45 +03:00
  • a785c53962 metal : add Q4_1 implementation (#1785) Kawrakow 2023-06-10 11:28:11 +03:00
  • e9b66ee982 metal : add Q4_1 implementation (#1785) Kawrakow 2023-06-10 11:28:11 +03:00
  • efdb29c078 llama : support requantizing models instead of only allowing quantization from 16/32bit (#1691) Kerfuffle 2023-06-10 01:59:17 -06:00
  • 4f0154b0ba llama : support requantizing models instead of only allowing quantization from 16/32bit (#1691) Kerfuffle 2023-06-10 01:59:17 -06:00