Open-source language models often underperform on African languages and demand high computational resourcesβbarriers to real-world use in the African context. To make language AI truly inclusive, we need models that are smaller, smarter, and optimized for resource-constrained environments.
The Lelapa AI Buzuzu-Mavi Challenge tasked participants with compressing Lelapa AIβs InkubaLMβan open-source small language model (SLM)βwhile maintaining or improving performance for two key African languages: Swahili and Hausa.
This repository presents our Bronze Medal-winning solution. π₯
β
Compress InkubaLM to reduce size and inference cost
β
Retain or improve model accuracy on core NLP tasks
β
Ensure usability on low-resource devices and CPUs
β
Focus on Swahili and Hausa performance
The model was evaluated across three NLP tasks:
- π£οΈ Sentiment Analysis
- π§ Natural Language Inference (AfriXNLI β true/false reasoning)
- π Machine Translation (English β Swahili & Hausa)
Performance could be improved by either:
- Increasing task accuracy,
- Reducing model size,
- Or both.
π§ Quantization β Reduced precision (8-bit & 4-bit) for faster, leaner models
βοΈ Pruning β Removed redundant parameters
π Language-Specific Fine-tuning β Custom fine-tuning on Swahili and Hausa datasets
This work moves us closer to a future where African languages have equal representation in the AI ecosystem. Smaller, smarter models enable:
- β Faster NLP on standard CPUs
- β Offline language tools
- β Scalable deployment in education, agriculture, health, and customer service