index
:
llama.cpp.git
master
llama.cpp
user
about
summary
refs
log
tree
commit
diff
log msg
author
committer
range
Age
Commit message (
Expand
)
Author
2023-05-20
cuda : loading models directly into VRAM, norm calculation on GPU, broadcasti...
Johannes Gäßler
2023-05-20
Revert "feature : add blis and other BLAS implementation support (#1502)"
Georgi Gerganov
2023-05-20
feature : add blis and other BLAS implementation support (#1502)
Zenix
2023-05-20
llama : add llama_init_backend() API (close #1527)
Georgi Gerganov
2023-05-20
Fix for mingw (#1462)
DannyDaemonic
2023-05-20
llama : fix name shadowing and C4146 (#1526)
Maxime
2023-05-20
llama : fix compile warnings in llama_set_state_data()
Georgi Gerganov
2023-05-20
ggml : fix scalar implementation of Q4_1 dot
Georgi Gerganov
2023-05-19
ggml : use F16 instead of F32 in Q4_0, Q4_1, Q8_0 (#1508)
Georgi Gerganov
2023-05-19
tests : add missing header
Georgi Gerganov
2023-05-19
examples : add persistent chat (#1495)
Evan Jones
2023-05-19
main : make reverse prompt option act as a stop token in non-interactive mode...
Jason McCartney
2023-05-19
readme : adds WizardLM to the list of supported models (#1485)
David Kennedy
2023-05-19
minor : fix compile warnings
Georgi Gerganov
2023-05-18
make kv_f16 the default for api users (#1517)
Erik Scholz
2023-05-18
Fixes #1511 lambda issue for w64devkit (mingw) (#1513)
DannyDaemonic
2023-05-17
Remove unused n_parts parameter (#1509)
Stephan Walter
2023-05-17
benchmark-matmul: Print the average of the test results (#1490)
rankaiyx
2023-05-17
convert.py: Support models which are stored in a single pytorch_model.bin (#1...
Tom Jobbins
2023-05-16
~7% faster Q5_1 AVX2 code (#1477)
Ilya Kurdyukov
2023-05-16
define default model path once, sync path with readme (#1366)
András Salamon
2023-05-16
Add alternate include path for openblas (#1476)
sandyiscool
2023-05-15
fix get_num_physical_cores() (#1436)
zrm
2023-05-14
benchmark-matmul: fix clang-tidy issues, report results in GFLOPS (#1458)
slaren
2023-05-14
cuda : deduplicated dequantization code (#1453)
Johannes Gäßler
2023-05-14
ggml : alternative fix for race condition bug in non-inplace ggml_compute_for...
xaedes
2023-05-14
ggml : various fixes (#1450)
Georgi Gerganov
2023-05-14
ggml : add AVX support based on AVX2 code (#1430)
katsu560
2023-05-14
ggml : add GGML_QNT_VERSION to track quantization format changes
Georgi Gerganov
2023-05-13
cuda : fix convert function (#1412)
Georgi Gerganov
2023-05-13
make : fix PERF build with cuBLAS
Georgi Gerganov
2023-05-13
llama : fix unused warning
Georgi Gerganov
2023-05-13
ggml : multi-thread mul and diag_mask ops (#1428)
Georgi Gerganov
2023-05-13
ggml : GPU-accelerated token generation (#1412)
Johannes Gäßler
2023-05-13
ggml : implement backward pass for llama + small training-llama-from-scratch ...
xaedes
2023-05-13
ggml : sync alibi fix from ggml repo
Georgi Gerganov
2023-05-13
Adding SSE instructions to ggml_vec_dot_q4_0_q8_0 (#1413)
3ooabkhxtn
2023-05-13
llama : fix various warnings
Georgi Gerganov
2023-05-13
embedding : remove unused code (#1426)
Rinne
2023-05-13
readme : update Q4_0 perplexities
Georgi Gerganov
2023-05-13
llama : free ggml context in set / copy state data (close #1425)
Georgi Gerganov
2023-05-13
opencl : fix kernels for the new formats (#1422)
Henri Vasserman
2023-05-12
llama : fix --mtest option (close #1414)
Georgi Gerganov
2023-05-12
CLI args use - instead of _, backwards compatible (#1416)
Johannes Gäßler
2023-05-12
Add clang-tidy reviews to CI (#1407)
slaren
2023-05-12
readme : add C#/.NET bindings repo (#1409)
Rinne
2023-05-12
ggml : remove bit shuffling (#1405)
Georgi Gerganov
2023-05-11
prompts : model agnostic DAN (#1304)
CRD716
2023-05-10
main : add option to save full output to session (#1338)
Evan Jones
2023-05-09
Locale fix for Windows (#1379)
DannyDaemonic
[prev]
[next]