# Linux/macOS
curl -L https://github.com/Michael-A-Kuykendall/shimmy/releases/latest/download/shimmy -o shimmy
chmod +x shimmy
# Windows
curl -L https://github.com/Michael-A-Kuykendall/shimmy/releases/latest/download/shimmy.exe -o shimmy.exePlace any .gguf file in one of these locations:
./models/your-model.gguf- Set
SHIMMY_BASE_GGUF=/path/to/your-model.gguf - Or just put it in
~/Downloads/- Shimmy will find it
Don't have a model? Try microsoft/Phi-3-mini-4k-instruct-gguf
./shimmy serveThat's it! Shimmy is now running on http://localhost:11435
VSCode Copilot:
// settings.json
{
"github.copilot.advanced": {
"serverUrl": "http://localhost:11435"
}
}Continue.dev:
{
"models": [{
"title": "Local Shimmy",
"provider": "openai",
"model": "your-model-name",
"apiBase": "http://localhost:11435/v1"
}]
}Cursor:
Set custom endpoint to http://localhost:11435
# List available models
./shimmy list
# Test generation
./shimmy generate --name your-model --prompt "Hello!" --max-tokens 10
# Or use curl
curl -X POST http://localhost:11435/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "your-model",
"messages": [{"role": "user", "content": "Hello!"}],
"max_tokens": 10
}'No models found?
- Make sure your
.gguffile is in./models/or setSHIMMY_BASE_GGUF - Run
./shimmy discoverto see what Shimmy can find
Port already in use?
./shimmy serve --bind 127.0.0.1:11436Need help?
- Open an issue
- Check existing discussions
Next: Check out integrations for more examples!