MLX Community Projects #654
Replies: 47 comments 18 replies
-
|
Text generation: mlx-tuning-fork |
Beta Was this translation helpful? Give feedback.
-
|
text generation: https://github.com/mzbac/mlx-moe-models |
Beta Was this translation helpful? Give feedback.
-
|
An implementation of Reinforcement Learning algorithms in MLX based in the Implementations from CleanRL. Still WIP because it’s missing a benchmark and some other minor things, but the implementations work correctly. |
Beta Was this translation helpful? Give feedback.
-
|
mlx-models. Currently supporting vision models by loading/converting from PyTorch checkpoints. Will later add support for text and audio models as well. |
Beta Was this translation helpful? Give feedback.
-
|
Hi I would love to add chat-with-mlx. It is a Chat UI + RAG Implementation on MLX. I wIll add more features later on (more advanced RAG pipeline + multimodal) |
Beta Was this translation helpful? Give feedback.
-
|
I have an example of training a simple language model using BitLinear instead of nn.Linear. It's a port of Karpathy's minGPT to MLX along with a custom implementation of a BitLinear module. https://github.com/adhulipa/mlx-mingpt I noticed this collection already has the far more meatier |
Beta Was this translation helpful? Give feedback.
-
|
Transformer Lab https://github.com/transformerlab/transformerlab-app is an LLM research platform that allows you to run, train, perform RAG, and evaluate LLMs through a GUI. |
Beta Was this translation helpful? Give feedback.
-
|
MLX RAG with GGUF Models: https://github.com/Jaykef/mlx-rag-gguf The code here builds on https://github.com/vegaluisjose/mlx-rag, it has been optimized to support RAG-based inferencing for .gguf models. I am using BAAI/bge-small-en for the embedding model, TinyLlama-1.1B-Chat-v1.0-GGUF as base model and the custom vector database script for indexing texts in a pdf file. Inference speeds can go up to ~413 tokens/sec for prompts and ~36 tokens/sec for generation on my 8G M2 Air. |
Beta Was this translation helpful? Give feedback.
-
|
@Jaykef Very cool, thanks for sharing |
Beta Was this translation helpful? Give feedback.
-
|
Vision: MLX3D A library for deep learning with 3D data using mlx. |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
|
JSON schema decoding (allowing function calling, including an OpenAI-compatible server with tools) using MLX: https://github.com/otriscon/llm-structured-output |
Beta Was this translation helpful? Give feedback.
-
|
Hello for text generation part, I'm happy to share with you that I've proposed and contributed to the integration of MLX with LibreChat.ai. So now you can use your local LLM powered by MLX through a fancy interface privately, enjoy! :D See danny-avila/LibreChat#2580 If in the future the community proposes an API servers supporting also multimodality, transcription, image generation for example, I will add them into LibreChat ;) It could be great also to have and LLM API supporting /models endpoint and multiple models simultaneously :D |
Beta Was this translation helpful? Give feedback.
-
|
Hello, mlx community, we are happy to share with you that we have contributed the first strong sub-4 bit LLM model zoo for MLX community.
The modern LLM families include Llama3/2, Phi-3, Mistral, 01-Yi, and Qwen. A mlx-style inference toolkit is also shared for the local web chatting.
We are an active team here, supporting the better low-bit community on the local platform. Enjoy! |
Beta Was this translation helpful? Give feedback.
-
|
mlx_micrograd - mlx port of Karpathy's micrograd - a tiny scalar-valued autograd engine with a small PyTorch-like neural network library on top. Installationpip install mlx_microgradExample usageExample showing a number of possible supported operations: from mlx_micrograd.engine import Value
a = Value(-4.0)
b = Value(2.0)
c = a + b
d = a * b + b**3
c += c + 1
c += 1 + c + (-a)
d += d * 2 + (b + a).relu()
d += 3 * d + (b - a).relu()
e = c - d
f = e**2
g = f / 2.0
g += 10.0 / f
print(f'{g.data}') # prints array(24.7041, dtype=float32), the outcome of this forward pass
g.backward()
print(f'{a.grad}') # prints array(138.834, dtype=float32), i.e. the numerical value of dg/da
print(f'{b.grad}') # prints array(645.577, dtype=float32), i.e. the numerical value of dg/db
|
Beta Was this translation helpful? Give feedback.
-
|
https://github.com/Trans-N-ai/swama Swama is a high-performance machine learning runtime written in pure Swift, designed specifically for macOS and built on Apple's MLX framework. It provides a powerful and easy-to-use solution for local LLM (Large Language Model) and VLM (Vision Language Model) inference. |
Beta Was this translation helpful? Give feedback.
-
|
Realized Toolio isn't listed.
|
Beta Was this translation helpful? Give feedback.
-
|
I built this https://github.com/arthurcolle/mlx.erl Still WIP |
Beta Was this translation helpful? Give feedback.
-
|
any benchmark nos of training on a reasonable datasets with MLX training? |
Beta Was this translation helpful? Give feedback.
-
|
Text generation: mlx-coconut Huge fan by the way! |
Beta Was this translation helpful? Give feedback.
-
|
BlossomTuneLLM-MLX combines mlx-lm with Flower to enable federated fine-tuning of SLMs (Small Language Models) on MacOS devices The project is the MLX-native evolution of an earlier codebase for FlowerTune LLM: How it works:
|
Beta Was this translation helpful? Give feedback.
-
|
Inferencer is an inference app that uses mlx-lm to expose the token entropy and probabilities to allow for controlling the output generated. |
Beta Was this translation helpful? Give feedback.
-
|
Would be great to have my mlx-lm-lora, and mlx-lm-lens packages in there too. |
Beta Was this translation helpful? Give feedback.
-
|
Hi! I'd like to share M-Courtyard — a macOS desktop app for fine-tuning LLMs It provides a full GUI workflow: Key features:
GitHub: https://github.com/Mcourtyard/m-courtyard [附上截图 EN 1 和 EN 3] Happy to hear any feedback! |
Beta Was this translation helpful? Give feedback.
-
|
It would be great to have the mlx-docs-l10n project maintained by the @localizethedocs organization. See the announcement post for more details. If we completes other MLX documentation localization projects in the future, maybe we could have a new category, Translation. For example:
|
Beta Was this translation helpful? Give feedback.
-
|
Hi all! Would be great to add these to the list: mlx-ruby - Ruby bindings for MLX with a nice ruby-esque DSL Some examples of MLX models exported through mlx-onnx running on WebGPU in the browser here. Cheers! |
Beta Was this translation helpful? Give feedback.
-
|
Published a benchmark & crisis recognition-focused MLX write-up from Calm Engineering on local fine-tuning of Phi-3.5/Qwen2.5-class models (3B–7B) on M3 Max (64GB): https://blog.calm.com/engineering/fine-tuning-slmllms-using-mlx What’s in it:
Thought this community might be interested in it. |
Beta Was this translation helpful? Give feedback.
-
|
mlx-code: Local Claude Code-style agent via mlx-lm It's a local Claude Code style coding agent built on Current features:
This is very early / first-pass code, basically a minimal proof-of-concept to see how far a simple local agent can go. |
Beta Was this translation helpful? Give feedback.
-
|
I've been maintaining Awesome MLX — a curated list of 80+ MLX community projects, organized by category (inference, training, audio, vision, Swift, etc.) with a Quick Start guide and model recommendations by RAM size. Anyone can add their project via a simple issue form. PRs welcome too! |
Beta Was this translation helpful? Give feedback.
-
|
MOLA: multi-LoRA inference server for MLX. One base model stays loaded, adapters route per request, no reloads. 8 adapters on Qwen3.5-9B, M5 Max 64GB: 732 tok/s same-adapter, 555 tok/s mixed at c64 (~24% overhead). Uses mx.gather_mm for batched multi-adapter decode. Still alpha, needs a small mlx-lm patch (script included). OpenAI-compatible API. |
Beta Was this translation helpful? Give feedback.


Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Let's collect some cool MLX integrations and community lead projects here for visibility!
If you have a project you would like to feature, leave a comment, and we will add it. If the project is build with MLX Swift, add it to the MLX Swift Community Project page.
Text Generation
Vision
Speech and Audio
Multi-modal
Misc
Educational
picoGPT.Beta Was this translation helpful? Give feedback.
All reactions