Researchers have developed a method called CompreSSM that compresses AI models during training, making them leaner and faster without sacrificing performance. This technique targets state-space models used in various applications, allowing for more efficient training processes.
Read originalThe latest version b8991 of llama.cpp has been released, featuring updates for various operating systems.
The latest update to llama-mmap improves compatibility with various platforms and model sizes. Key enhancements include support for 32-bit wasm and updates to gguf.cpp style.
