Skip to content

knoveleng/mod-evaluate

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Mixture of Distributions - Evaluation

This repository contains the code and evaluation scripts for the paper MoD: A Distribution-Based Approach for Merging Large Language Models.
The paper has been accepted in Neurips 2024's LLM Merging Competition.

Comparision

Setup

1. Clone Repository and Install Requirements

Start by cloning the repository and setting up a virtual environment:

git clone https://github.com/knovel-eng/mod-evaluate.git
cd mod-evaluate
python -m venv venv
source venv/bin/activate
pip install --upgrade pip
pip install -r requirements.txt

2. Set Up Experiments

To prepare the experiments, clone the necessary dependencies and set up additional modules:

# Clone Qwen2.5-Math and copy shell scripts for evaluation
git clone https://github.com/QwenLM/Qwen2.5-Math.git
cp -r sh/ Qwen2.5-Math/evaluation/

# Clone the main MoD repository
cd Qwen2.5-Math/
git clone https://github.com/knovel-eng/mod.git

# Navigate to the latex2sympy folder and install dependencies
cd evaluation/latex2sympy
pip install -e .

# Install remaining dependencies
cd ..
pip install -r requirements.txt 
pip install vllm==0.5.1 --no-build-isolation
pip install transformers

Running Experiments

To run evaluations for different model configurations, execute the following shell scripts:

sh sh/run_eval_qwen2_1.5_math.sh
sh sh/run_eval_qwen2_7_math.sh

Cite our Work

@misc{dang2024moddistributionbasedapproachmerging,
      title={MoD: A Distribution-Based Approach for Merging Large Language Models}, 
      author={Quy-Anh Dang and Chris Ngo},
      year={2024},
      eprint={2411.00406},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2411.00406}, 
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •  

Languages