"Jan now runs faster on CUDA: Llama.cpp update and bug fix"

Jan now runs faster on CUDA. We updated llama.cpp with the latest improvements and fixed a CUDA backend bug upstream. Jan stays pinned to v6324 due to flash-attention changes. If you have auto-update on, you'll get this automatically - if not, we recommend turning it on.

To view or add a comment, sign in

Explore content categories