index
:
llama.cpp.git
master
llama.cpp
user
about
summary
refs
log
tree
commit
diff
log msg
author
committer
range
Age
Commit message (
Expand
)
Author
2023-05-17
Remove unused n_parts parameter (#1509)
Stephan Walter
2023-05-17
benchmark-matmul: Print the average of the test results (#1490)
rankaiyx
2023-05-17
convert.py: Support models which are stored in a single pytorch_model.bin (#1...
Tom Jobbins
2023-05-16
~7% faster Q5_1 AVX2 code (#1477)
Ilya Kurdyukov
2023-05-16
define default model path once, sync path with readme (#1366)
András Salamon
2023-05-16
Add alternate include path for openblas (#1476)
sandyiscool
2023-05-15
fix get_num_physical_cores() (#1436)
zrm
2023-05-14
benchmark-matmul: fix clang-tidy issues, report results in GFLOPS (#1458)
slaren
2023-05-14
cuda : deduplicated dequantization code (#1453)
Johannes Gäßler
2023-05-14
ggml : alternative fix for race condition bug in non-inplace ggml_compute_for...
xaedes
2023-05-14
ggml : various fixes (#1450)
Georgi Gerganov
2023-05-14
ggml : add AVX support based on AVX2 code (#1430)
katsu560
2023-05-14
ggml : add GGML_QNT_VERSION to track quantization format changes
Georgi Gerganov
2023-05-13
cuda : fix convert function (#1412)
Georgi Gerganov
2023-05-13
make : fix PERF build with cuBLAS
Georgi Gerganov
2023-05-13
llama : fix unused warning
Georgi Gerganov
2023-05-13
ggml : multi-thread mul and diag_mask ops (#1428)
Georgi Gerganov
2023-05-13
ggml : GPU-accelerated token generation (#1412)
Johannes Gäßler
2023-05-13
ggml : implement backward pass for llama + small training-llama-from-scratch ...
xaedes
2023-05-13
ggml : sync alibi fix from ggml repo
Georgi Gerganov
2023-05-13
Adding SSE instructions to ggml_vec_dot_q4_0_q8_0 (#1413)
3ooabkhxtn
2023-05-13
llama : fix various warnings
Georgi Gerganov
2023-05-13
embedding : remove unused code (#1426)
Rinne
2023-05-13
readme : update Q4_0 perplexities
Georgi Gerganov
2023-05-13
llama : free ggml context in set / copy state data (close #1425)
Georgi Gerganov
2023-05-13
opencl : fix kernels for the new formats (#1422)
Henri Vasserman
2023-05-12
llama : fix --mtest option (close #1414)
Georgi Gerganov
2023-05-12
CLI args use - instead of _, backwards compatible (#1416)
Johannes Gäßler
2023-05-12
Add clang-tidy reviews to CI (#1407)
slaren
2023-05-12
readme : add C#/.NET bindings repo (#1409)
Rinne
2023-05-12
ggml : remove bit shuffling (#1405)
Georgi Gerganov
2023-05-11
prompts : model agnostic DAN (#1304)
CRD716
2023-05-10
main : add option to save full output to session (#1338)
Evan Jones
2023-05-09
Locale fix for Windows (#1379)
DannyDaemonic
2023-05-09
use pause asm insn in busyloop to run the CPU (13600K) 10 °C cooler (#1314)
Sami Farin
2023-05-08
Interface improvements and `--multiline-input` (previously `--author-mode`) (...
DannyDaemonic
2023-05-08
readme : add notice about upcoming breaking change
Georgi Gerganov
2023-05-08
readme : add TOC and Pygmalion instructions (#1359)
AlpinDale
2023-05-08
llama : fix hparams shadow (#1367)
Pavol Rusnak
2023-05-08
llama : require first token to be BOS (#1303)
Georgi Gerganov
2023-05-08
convert: add ability to convert safetensors files (#1276)
ubik2
2023-05-08
Documented CUDA reproducibility, added warning (#1346)
Johannes Gäßler
2023-05-07
CI: add Windows CLBlast and OpenBLAS builds (#1277)
Henri Vasserman
2023-05-06
ggml : Allow usage of CLBlast alongside Accelerate.framework (#1336)
swittk
2023-05-06
Remove default arguments from sampling functions (#1343)
Jed Fox
2023-05-05
makefile: automatic Arch Linux detection (#1332)
DaniAndTheWeb
2023-05-05
ci : add cublas to windows release (#1271)
Erik Scholz
2023-05-05
readme: add missing info (#1324)
Pavol Rusnak
2023-05-05
Fix for OpenCL / clbast builds on macOS. (#1329)
Ionoclast Laboratories
2023-05-05
Convert.py @staticmethod (#1327)
Benjamin Lecaillon
[prev]
[next]