llama.cpp

Edit Package llamacpp
https://github.com/ggerganov/llama.cpp

The main goal of llama.cpp is to enable LLM inference with minimal setup and state-of-the-art performance on a wide variety of hardware - locally and in the cloud.

Refresh
Refresh
Source Files
Filename Size Changed
_service 0000000814 814 Bytes
_servicedata 0000000240 240 Bytes
llamacpp-4153.obscpio 0096390158 91.9 MB
llamacpp.changes 0000015877 15.5 KB
llamacpp.obsinfo 0000000096 96 Bytes
llamacpp.spec 0000005816 5.68 KB
Latest Revision
Eyad Issa's avatar Eyad Issa (VaiTon) committed (revision 33)
- Update to version 4153:
  * ci: Update oneAPI runtime dll packaging (#10428)
  * GitHub: ask for more info in issue templates (#10426)
  * CANN: Support Ascend310P to accelerate F32 and F16 Model (#10216)
  * cuda : optimize argmax (#10441)
  * llama : handle KV shift for recurrent models (#10402)
  * sync : ggml
  * ggml/sched : do not skip views in pre-assignments
  * ggml-opt: fix data corruption (ggml/1022)
  * vulkan: predicate max operation in soft_max shaders/soft_max (#10437)
  * cmake: add link dependencies to cmake find pkg (#10433)
  * llama : add .clang-format file (#10415)
  * vulkan: copy iq4_nl LUT into shared memory (#10409)
  * vulkan: further optimize mul_mat_vec using larger loads (#10387)
  * update rel to 4040 (#10395)
  * Fix missing file renames in Makefile due to changes in commit ae8de6d50a (#10413)
  * add cmake rvv support (#10411)
  * sync : ggml
  * metal : fox offset integer overflows in im2col (ggml/1015)
  * metal : add `GGML_UNARY_OP_ELU` kernel (ggml/1018)
  * cmake: force MSVC compiler charset to utf-8 (#9989)
  * Add required ggml-base and backend libs to cmake pkg (#10407)
  * cuda : fix CUDA_FLAGS not being applied (#10403)
  * llama : add check for KV cache shifts (#10401)
Comments 0
openSUSE Build Service is sponsored by