If you wish to be acknowledged for your contributions, please list your name with a short description of your contribution(s) below. For example:
- Jane Smith: Added the
fooexample.
MLX LM was developed with contributions from the following individuals:
- Shunta Saito: Added support for PLaMo models.
- Gökdeniz Gülmez: Added support for the following architectures:
OpenBMB's
MiniCPMandMiniCPM3, Kyutai'sHelium, State-Space'sMamba v1andMamba v2, Z.ai & THUKEG'sGLM,GLM4,GLM5 (GLM MoE DSA), Rednotedots.llm1, Baidu'sErnie4.5 MoE, inclusionAI'sBailing MoE e.g. Ling-family,Bailing MoE Linear e.g. Ling-Linear-family, Klear team - Kuaishou Technology'sKlear, AI21 Lab'sJambaIBM'sGranite MoE, Meituan'sLongCat, Nvidia'sNemotron H, Swiss-AI'sApertus, Nikity'sLille130m, Alibaba Qwen'sQwen3Next, Tele-AI'sTeleChat3, and Allenai'sOLMoEandOlmo 3; Helped add support for the following model architectures: Alibaba Qwen'sQwen3 & Qwen3MoE); Added support for the following training algorithms:Full Weight Fine-Tuning, and theMuonoptimizer; Added support for the following other features:Multiple Optimizers to choose for training, andreporting training metrics to WandB (Weights & Biases). - Prince Canuma: Helped add support for the following model architectures:
HuggingFace's
Starcoder2, Cohere'sCohere (1 and 2), Alibaba Qwen'sQwen (2, 3 and MoE), Microsoft'sPhi (3 and 3.5 MoE),BitNet1.58, Meta'sLlama (3 and 4), MinimaxAI'sMiniMax, MoonshotAI'sKimi-Linear, LiquidAI'sLFM2andLFM2 MoE, Google DeepMind'sGemma 3, TII'sFalcon H1and InterLM'sInternLM 2.5. - Ivan Fioravanti: Added support for the following architectures:
ServiceNow-AI's
Apriel 1.5, Tencent'sHunyuan Dense V1andHunyuan MoE V1. - Tarjei Mandt: Added support for the following architectures:
Step 3.5 Flash, MoonshotAI'sKimi K2.5, Upstage'sSolar Open, LG AI Research'sK-Exaone MoE, Meituan'sLongCat Flash LiteHelped add support for the following model architectures: Z.ai & THUKEG'sGLM5 (GLM MoE DSA)