
#MeWriting I’m happy to announce the immediate availability of Mmojo Server for Windows. Mmojo Server is a Large Language Model (LLM) server that runs on your PC or laptop. It supports the industry standard OpenAI API, so you can connect AI applications to it. It works with an NVIDIA GPU if you have one available on your computer. It also works with your computer’s CPU, albeit a bit slower.
On Windows, Mmojo Server runs in a Windows Subsystem for Linux (WSL) sandbox. This let’s me ship you the fastest builds that are compatible with popular NVIDIA GPUs. It also help keep your Mmojo Server private to your computer. The linked instructions walk you through setting up WSL, downloading some models, downloading and installing the Mmojo Server software, and making it all run.
Install Mmojo Server for Windows — Instructions
The Mmojo Server software incorporates the popular llama.cpp LLM server software, and is fully open source to ensure that your data exchanged with Mmojo Server remains private. If you don’t trust me telling you that, you are welcome to inspect and build the source code! Mmojo Server is compatible with many .gguf models you can find on Hugging Face and other web sites.
Installing Mmojo Server is a fun do-it-yourself adventure! I’ve asked non-technical people to test my instructions with good results. Casual and occasional developers should have no problems.
Your PC or laptop should have a recent high-end Intel or AMD CPU, with at least 16 GB RAM and 100 GB available storage space. An NVIDIA GPU with 8GB or more of VRAM will make Mmojo Server faster.
If you need assistance via Zoom call and screen sharing, I offer a one-hour hands-on session, for (US) $100. It can be scheduled during extended west coast business hours. You will be working with me, the guy who made this thing work. Email me if interested.
-Brad
—
Brad Hutchings
brad@BradHutchings.com
