TL;DR
A WinForms chat client forllama.cpp
,whisper.cpp
and your questionable life choices.
Zero Electron. Zero telemetry. Zero regrets.
☑️ | What It Does |
---|---|
✅ | Launches in milliseconds—your GPU blinks and it’s already chatting. |
✅ | Edit / regenerate / yeet messages. |
✅ | Shows the model’s “thinking” stream (fun for prompt nerds). |
✅ | Drag-drop files → instant code blocks. |
✅ | One-click Hugging Face search & download. |
✅ | Built-in Google Search + webpage fetch (super-visible setup below). |
✅ | Optional speech I/O with whisper.cpp —talk smack to your computer. |
✅ | Tiny memory footprint—smaller than RGB keyboard driver. |
1) Grab an API key
https://console.cloud.google.com/apis/dashboard
→ new project → enable “Custom Search API” → copy the key.
2) Create a Search Engine ID
https://programmablesearchengine.google.com/controlpanel/overview
→ “Add” → “Search the entire web” → grab the cx ID.
3) Paste both values in Settings → Google Search Tool.
Congrats—~100 free queries per day. Abuse responsibly.
Chat Tab | Settings Tab | Models Tab | Hugging Face Tab |
---|---|---|---|
1. Install Visual Studio with C# and C++ development tools
2. Clone https://github.com/ggml-org/llama.cpp or download the source code of the latest release
3. Open "x64 Native Tools Command Prompt for VS <version>", cd to the llama.cpp folder and run "mkdir build && cd build"
4. `cmake .. -DGGML_NATIVE=OFF -DGGML_BACKEND_DL=ON -DGGML_AVX2=ON -DGGML_BMI2=ON -DGGML_CUDA=ON -DGGML_CUDA_F16=ON -DLLAMA_CURL=OFF -DLLAMA_ALL_WARNINGS=OFF -DLLAMA_BUILD_TESTS=OFF -DLLAMA_BUILD_TOOLS=OFF -DLLAMA_BUILD_EXAMPLES=OFF -DLLAMA_BUILD_SERVER=OFF`
5. Open build\llama.cpp.sln with visual studio and build the Release version
1. Clone https://github.com/CommanderLake/LMStud
2. Copy build\common\Release\common.lib, build\src\Release\llama.lib, build\ggml\src\Release\ggml.lib and ggml-base.lib to <LMStud solution folder>\lib\Release
3. From build\bin\Release copy all .dll files to <LMStud solution folder>\LM Stud\bin\x64\Release
4. If you wish to use the Debug version follow the last 2 steps but replace "Release" with "Debug"
5. If you want whisper.cpp for voice input the build steps are similar to llama.cpp https://github.com/ggml-org/whisper.cpp just copy just whisper.lib and whisper.dll files to the respective folders
1. Set up vcpkg, step "1 - Set up vcpkg": https://learn.microsoft.com/en-us/vcpkg/get_started/get-started
2. From a new visual studio command prompt...
3. vcpkg install SDL2:x64-windows-static
4. vcpkg install curl[openssl]:x64-windows-static
5. Open LM Stud.sln in Visual Studio
6. Adjust "Include Directories" in "VC++ Directories" in the Stud project to your llama.cpp, whisper.cpp and vcpkg locations
7. Build LM Stud
Section | Knobs & Dials |
---|---|
CPU Params / Batch | Generation threads, batch threads. |
Common | Context size, GPU layers, temperature, tokens to generate. |
Advanced | NUMA strategy, repeat penalty, top-k/p, batch size. |
Voice | Model picker, wake word, VAD, frequency threshold, GPU toggle. |
Tools | Enable Google Search (API key + cx), enable webpage fetch. |
- Fork it.
- Break it.
- PR it.
- Get internet points. (Memes optional but encouraged.)
MIT—short, sweet, almost readable.