index
:
llama.cpp.git
master
llama.cpp
user
about
summary
refs
log
tree
commit
diff
log msg
author
committer
range
Age
Commit message (
Expand
)
Author
2023-03-25
Cleanup STL headers + fix embedding examples + minor stuff
Georgi Gerganov
2023-03-25
Move chat scripts into "./examples"
Georgi Gerganov
2023-03-25
Add AVX2 implementation of dequantize_row_q4_1 (#505)
slaren
2023-03-25
Overhaul the examples structure
Georgi Gerganov
2023-03-25
Retire the ggml_mul_mat() branch for transposed src0 (#500)
Georgi Gerganov
2023-03-25
Disable prompt verbosity by default and add option to enable (#480)
Georgi Gerganov
2023-03-25
Add AVX2 implementation of dequantize_row_q4_0 (#467)
slaren
2023-03-25
Don't interefe with BLAS for large prompts by running only 1 thread
Georgi Gerganov
2023-03-25
Add longer DAN prompt for testing big batch numbers
Georgi Gerganov
2023-03-25
Add timings for the prompt evaluation (#478)
slaren
2023-03-25
Remove obsolete information from README
Georgi Gerganov
2023-03-25
Remove obsolete assert and fix compiler warning
Georgi Gerganov
2023-03-25
Fix nasty bug in ggml_compute_forward_mul_mat_f32() and reenable BLAS
Georgi Gerganov
2023-03-25
bounds checking for input prefix (#492)
anzz1
2023-03-25
feat: '--in-prefix STRING' option (#426)
anzz1
2023-03-25
Add support for file load progress reporting callbacks (#434)
Jed Fox
2023-03-25
Add missing struct annotation (#483)
Doomsdayrs
2023-03-25
Fix crash for 65B model with pre-allocated memory (#485)
Chris Kuehl
2023-03-24
Disable BLAS altogether - the bug is not just for qunatized mat mul
Georgi Gerganov
2023-03-24
Disable BLAS branch in mul_mat - seems there is a bug
Georgi Gerganov
2023-03-24
Immediately start processing the prompt before user input has been provided (...
Georgi Gerganov
2023-03-24
Reduce memory usage and allocate enough memory for largest context (#473)
Georgi Gerganov
2023-03-24
Temporary bump the memory buffer size - hopefully fix issues from 483bab2e
Georgi Gerganov
2023-03-24
Update README.md (#444)
Gary Mulder
2023-03-24
fix instruct mode (#445)
rabidcopy
2023-03-24
Properly free llama_context on failure
Georgi Gerganov
2023-03-24
additional optimizations for POWER9 (#454)
Cameron Kaiser
2023-03-24
Support calling mlock() on loaded model data on Linux and macOS (#453)
comex
2023-03-24
Add embedding mode with arg flag. Currently working (#282)
Luciano
2023-03-24
Add link to Roadmap discussion
Georgi Gerganov
2023-03-24
Revert "Fix memory allocation issues and seg faults"
Georgi Gerganov
2023-03-24
Fix memory allocation issues and seg faults
Georgi Gerganov
2023-03-23
Avoid the transposed X branch in the Z = X * Y matrix multiplication (#439)
Georgi Gerganov
2023-03-23
Fix quantize script not finding models in parent directory (#428)
Jed Fox
2023-03-23
Remove oboslete command from Docker script
Georgi Gerganov
2023-03-23
Obsolete
Georgi Gerganov
2023-03-23
Replace EOS with newline to prevent context/memory being flushed by EOS in in...
rabidcopy
2023-03-23
Fix GPTQ converter (#423)
Timmy Knight
2023-03-23
Generate library with CMake (#430)
nusu-github
2023-03-23
Command line args bounds checking (#424)
anzz1
2023-03-23
Fix Nix build
Ben Siraphob
2023-03-23
Revert "Delete SHA256SUMS for now" (#429)
Stephan Walter
2023-03-23
Fix Makefile echo escape codes (by removing them). (#418)
Kerfuffle
2023-03-23
Move model section from issue template to README.md (#421)
Gary Mulder
2023-03-23
Delete SHA256SUMS for now (#416)
anzz1
2023-03-23
Adjust repetition penalty ..
Georgi Gerganov
2023-03-23
Add link to recent podcast about whisper.cpp and llama.cpp
Georgi Gerganov
2023-03-23
CI: CMake: Separate build and test steps (#376)
anzz1
2023-03-23
Fix instruct mode broken by PR #354 (#409)
tjohnman
2023-03-22
Update issue template so people will use it (#404)
Gary Mulder
[prev]
[next]