Skip to content

Llamafile

Open source project for distributing and running LLMs with a single file that is capable of running on six operating systems.

Overview

Contributor
Justine Tunney
Platform
CLI
Github
Code
Stars
21.0K
Forks
1.1K
Technologies
C++, Metal

More Information

Llamafile turns LLMs into a single executable file. Whether you’re a developer or an end user, you simply choose the LLM you want to run, download its llamafile, and execute it. llamafile runs on six operating systems (Windows, macOS, Linux, OpenBSD, FreeBSD, and NetBSD), and generally requires no installation or configuration. It uses your fancy GPU, if you have one. Otherwise, it uses your CPU. It makes open LLMs usable on everyday consumer hardware, without any specialized knowledge or skill.

Our goal is to make open LLMs much more accessible to both developers and end users. We’re doing that by combining llama.cpp with Cosmopolitan Libc into one framework that collapses all the complexity of LLMs down to a single-file executable (called a “llamafile”) that runs locally on most computers, with no installation.

We believe that llamafile is a big step forward for access to open source AI. But there’s something even deeper going on here: llamafile is also driving what we at Mozilla call “local AI.”

Local AI is AI that runs on your own computer or device. Not in the cloud, or on someone else’s computer. Yours. This means it’s always available to you. You don’t need internet access to use a local AI. You can turn off your WiFi, and it will still work.

Contribute

Channel
#llamafile
Members
128
Join the Discord

Contributors

About Justine

Justine is and independent developer and the creator of Llamafile. She is an activist who started as a cybersecurity hacker and used that experience to improve the security of open-source software. She is also the creator of Redbean, an open-source web server that allows you to build local web apps in a single zip executable.

Other Projects