Skip to content

Efficient and Effective Weight-Ensembling Mixture of Experts for Multi-Task Model Merging. Arxiv, 2024.

License

Notifications You must be signed in to change notification settings

EnnengYang/Efficient-WEMoE

Repository files navigation

Code for "Efficient and Effective Weight-Ensembling Mixture of Experts for Multi-Task Model Merging"


Installation

This project relies on FusionBench-v0.1.6. Please refer to it to configure the base environment.

git clone https://github.com/EnnengYang/Efficient-WEMoE
cd Efficient-WEMoE

pip install -e . # install the package in editable mode

Note

Our code is also integrated into FusionBench-V0.2.2. Refer to https://github.com/tanganke/fusion_bench for more information.


Run Experiment

  • Multi-task performance when merging CLIP-ViT-B/32 or CLIP-ViT-B/16 or CLIP-ViT-L/14 models on all eight tasks
bash examples/sparse_we_moe/grid_search_sparse_we_moe_ratio.sh
  • Generalization results on two unseen tasks when merging ViT-B/32 models on six tasks
bash examples/sparse_we_moe/generalization_vit_b32.sh
  • Ablations of the test data distribution on ViT-B/32 or CLIP-ViT-B/16
bash examples/sparse_we_moe/roubustness.sh

Note: The results of E-WEMoE's experiment can be found in './results/sparse_we_moe/'.


Acknowledgement

Our implementation references the code below, thanks to them.

About

Efficient and Effective Weight-Ensembling Mixture of Experts for Multi-Task Model Merging. Arxiv, 2024.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published