Releases: AmpereComputingAI/llama.cpp
Releases · AmpereComputingAI/llama.cpp
v1.2.2
Release notes:
- Fix llama-3 end of token issue
- Update server to support ollama (v0.1.33)
- llama.aio docker support server mode by default
SHA-256 hashes:
- 6c580006a8faf7b73a424b0020f1bda2684aa7e1796182f68bfa8b7fee08d991 llama_cpp_python-0.2.63-cp311-cp311-linux_aarch64.whl
- 1ffde8093abe18f638fb89273dd56664dd7ff6b8c82383099ea620d18ab562a7 llama_aio_v1.2.2_b769bc1.tar.gz