Llamafile continues pushing forward as the interesting Mozilla project to allow easily distributing and running AI large language models (LLMs) from a single file and in a cross-platform and cross-vendor hardware manner. Llamafile 0.9.3 is out today with more enhancements to this Mozilla Ocho project.
Llamafile 0.9.3 brings support for handling the Phi4 and Qwen3 models. Various changes based on Llama.cpp changes are needed for supporting the Phi4 and Qwen3 models within Llamafile packages.
The Llamafile 0.9.3 release also brings fixes for its newer LocalScore component. LocalScore is the new Mozilla Builders project for a local AI LLM benchmark. The LocalScore changes for Llamafile 0.9.3 incorporate some of my requests for enhancing the benchmark.
Plus this update brings some documentation updates and other minor enhancements. Downloads and more details on Llamafile 0.9.3 via GitHub. I’ll also be trying out some LocalScore AI benchmarks on different hardware now that the changes merged for this release should address some of my noted blockers.