index
:
llama.cpp.git
master
llama.cpp
user
about
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
ggml-opencl.cpp
Age
Commit message (
Expand
)
Author
2023-07-07
Fix opencl by wrap #if-else-endif with \n (#2086)
Howard Su
2023-07-04
[ggml] fix index for ne03 value in ggml_cl_mul_f32 (#2088)
Govlzkoy
2023-06-29
Porting the improved K-Quant CUDA kernels to OpenCL (#1966)
LostRuins
2023-06-17
ggml : fix warnings under MSVC (#1908)
Howard Su
2023-06-16
opencl : support k-quants (#1836)
0cc4m
2023-06-12
Leverage mmap for offloading tensors to GPU (#1597)
Howard Su
2023-06-09
OpenCL: Add release memory (#1741)
Robert Sung-wook Shin
2023-06-06
Multi GPU support, CUDA refactor, CUDA scratch buffer (#1703)
Johannes Gäßler
2023-06-06
Clblast fixes + enhancements to save VRAM and offload more layers (#1675)
LostRuins
2023-06-04
OpenCL: Fix duplication of layers in VRAM and RAM, add GPU mul kernel (#1653)
0cc4m
2023-05-28
opencl : no need to allocate cl_mem on heap (#1612)
Howard Su
2023-05-28
opencl : use strstr to check if fp16 supported (#1611)
Howard Su
2023-05-23
Fix handling of "invalid property" when creating OpenCL command queue (#1565)
Maarten ter Huurne
2023-05-23
OpenCL Token Generation Acceleration (#1459)
0cc4m