Skip to content

Pull requests: EvolvingLMMs-Lab/lmms-eval

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Add trust_remote_code param for huggingface model.
#1280 opened Mar 30, 2026 by sablin39 Loading…
1 of 7 tasks
feat: add MMBench static evaluation mode (no OpenAI API needed)
#1276 opened Mar 26, 2026 by Luodian Loading…
3 tasks
fix: improve evaluation logic across 10+ existing benchmarks
#1274 opened Mar 26, 2026 by Luodian Loading…
3 tasks
feat: add COVER and WM-aBench video understanding benchmarks
#1273 opened Mar 26, 2026 by Luodian Loading…
4 tasks
feat: add VBench video generation evaluation benchmark
#1271 opened Mar 26, 2026 by Luodian Loading…
3 tasks
feat: add MiniMax as LLM judge provider
#1263 opened Mar 22, 2026 by octo-patch Loading…
3 tasks done
ProTip! no:milestone will show everything without a milestone.