index
:
llama.cpp.git
master
llama.cpp
user
about
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
ggml.h
Age
Commit message (
Expand
)
Author
2023-04-17
Add LoRA support (#820)
slaren
2023-04-17
Speedup the AVX-512 implementation of ggml_vec_dot_q4_0() (#933)
Ivan Komarov
2023-04-15
ggml : add Q8_0 quantization for intermediate results (#951)
Georgi Gerganov
2023-04-14
Expose type name from ggml (#970)
Pavol Rusnak
2023-04-14
ggml : add unary and binary map operations (#874)
Kerfuffle
2023-04-13
ggml : add GGML_DEFAULT_N_THREADS
Georgi Gerganov
2023-04-11
Add enum llama_ftype, sync ggml_type to model files (#709)
Stephan Walter
2023-04-10
ggml : add ggml_cont() + optimize ggml_cpy() for contiguous dst
Georgi Gerganov
2023-04-10
Rewrite loading code to try to satisfy everyone:
comex
2023-04-08
Add quantize-stats command for testing quantization (#728)
unbounded
2023-04-05
ggml, llama : avoid heavy V transpose + improvements (#775)
Georgi Gerganov
2023-04-02
ggml : change ne to int64_t (#626)
Marian Cepok
2023-03-30
Ensure --mlock works properly with mmap() support
Justine Tunney
2023-03-30
Add mmap support for model files
Slaren
2023-03-28
ggml : introduce structs for the q4 data blocks (#356)
Stephan Walter
2023-03-24
Support calling mlock() on loaded model data on Linux and macOS (#453)
comex
2023-03-22
Deduplicate q4 quantization functions (#383)
Stephan Walter
2023-03-22
Introduce C-style API (#370)
Georgi Gerganov
2023-03-16
Add RMS norm and use it (#187)
hoangmit
2023-03-10
Initial release
Georgi Gerganov