node-llama-cpp 3.0 is finally here. With node-llama-cpp, you can run large language models locally on your machine using the power of llama.cpp with a simple and easy-to-use API. It includes everything you need, from downloading models, to running them in the most optimized way for your hardware, and integrating them in your projects.