index
:
llama.cpp.git
master
llama.cpp
user
about
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
Makefile
Age
Commit message (
Expand
)
Author
2023-07-14
make : fix combination of LLAMA_METAL and LLAMA_MPI (#2208)
James Reynolds
2023-07-10
mpi : add support for distributed inference via MPI (#2099)
Evan Miller
2023-07-07
docker : add support for CUDA in docker (#1461)
dylan
2023-07-05
Quantized dot products for CUDA mul mat vec (#2067)
Johannes Gäßler
2023-07-04
Allow old Make to build server. (#2098)
Henri Vasserman
2023-07-04
Update Makefile: clean simple (#2097)
ZhouYuChen
2023-06-28
llama : support input embeddings directly (#1910)
ningshanwutuobang
2023-06-26
k-quants : support for super-block size of 64 (#2001)
Kawrakow
2023-06-19
Convert vector to f16 for dequantize mul mat vec (#1913)
Johannes Gäßler
2023-06-18
metal : handle buffers larger than device's maxBufferLength (#1826)
Georgi Gerganov
2023-06-17
make : do not print help for simple example
Georgi Gerganov
2023-06-17
make : update for latest Arch (#1701)
DaniAndTheWeb
2023-06-17
Server Example Refactor and Improvements (#1570)
Randall Fitzgerald
2023-06-16
examples : add "simple" (#1840)
SuperUserNameMan
2023-06-16
CUDA : faster k-quant dot kernels (#1862)
Kawrakow
2023-06-15
make : add train-text-from-scratch (#1850)
daboe01
2023-06-15
make : clean *.so files (#1857)
sandyiscool
2023-06-13
Allow "quantizing" to f16 and f32 (#1787)
Kerfuffle
2023-06-10
make : add SSSE3 compilation use case (#1659)
rankaiyx
2023-06-07
k-quants : allow to optionally disable at compile time (#1734)
Georgi Gerganov
2023-06-06
ggml : fix builds, add ggml-quants-k.o (close #1712, close #1710)
Georgi Gerganov
2023-06-05
ggml : add SOTA 2,3,4,5,6 bit k-quantizations (#1684)
Kawrakow
2023-06-04
llama : Metal inference (#1642)
Georgi Gerganov
2023-05-28
LLAMA_DEBUG adds debug symbols (#1617)
Johannes Gäßler
2023-05-27
Include server in releases + other build system cleanups (#1610)
Kerfuffle
2023-05-26
cuda : performance optimizations (#1530)
Johannes Gäßler
2023-05-23
OpenCL Token Generation Acceleration (#1459)
0cc4m
2023-05-21
make : .PHONY clean (#1553)
Stefan Sydow
2023-05-20
feature : support blis and other blas implementation (#1536)
Zenix
2023-05-20
Revert "feature : add blis and other BLAS implementation support (#1502)"
Georgi Gerganov
2023-05-20
feature : add blis and other BLAS implementation support (#1502)
Zenix
2023-05-16
Add alternate include path for openblas (#1476)
sandyiscool
2023-05-13
make : fix PERF build with cuBLAS
Georgi Gerganov
2023-05-05
makefile: automatic Arch Linux detection (#1332)
DaniAndTheWeb
2023-05-05
Fix for OpenCL / clbast builds on macOS. (#1329)
Ionoclast Laboratories
2023-05-02
Call sh on build-info.sh (#1294)
DannyDaemonic
2023-05-01
Add git-based build information for better issue tracking (#1232)
DannyDaemonic
2023-04-30
build: add armv{6,7,8} support to cmake (#1251)
Pavol Rusnak
2023-04-30
Various fixes to mat_mul benchmark (#1253)
Stephan Walter
2023-04-29
ggml : adjust mul_mat_f16 work memory (#1226)
Georgi Gerganov
2023-04-29
build : fix reference to old llama_util.h
Georgi Gerganov
2023-04-29
cuBLAS: use host pinned memory and dequantize while copying (#1207)
slaren
2023-04-28
ggml : add CLBlast support (#1164)
0cc4m
2023-04-28
Add Manjaro CUDA include and lib dirs to Makefile (#1212)
Johannes Gäßler
2023-04-24
Fix cuda compilation (#1128)
slaren
2023-04-23
ggml : better PERF prints + support "LLAMA_PERF=1 make"
Georgi Gerganov
2023-04-22
ggml : fix AVX build + update to new Q8_0 format
Georgi Gerganov
2023-04-21
Improve cuBLAS performance by using a memory pool (#1094)
slaren
2023-04-20
Add Q4_3 support to cuBLAS (#1086)
slaren
2023-04-20
fix: LLAMA_CUBLAS=1 undefined reference 'shm_open' (#1080)
源文雨
[next]