Official code of Attention-MoA: Enhancing Mixture-of-Agents via Inter-Agent Semantic Attention and Deep Residual Synthesis
The paper could be accessed at: Attention-MoA Paper Link.
-
Large-Scale Configuration:
This setup utilizes SOTA large language models: Claude-4.5-Sonnet, Gemini-2.5-Pro, GPT-4.1, Qwen-Max, and DeepSeek-V3.1.
-
Small-Scale Configuration:
This setup on smaller, efficient models: Mistral-Small-3.2-24B-Instruct-2506, Qwen3-32B, gemma-3-12b-it, Llama-4-Scout-17B-16E-Instruct and gpt-oss-20b
We evaluate Attention-MoA on three benchmarks: AlpacaEval 2.0, MT-Bench, and FLASK.
intall requirements
pip3 install -r requirements.txt
cd alpaca_eval
pip3 install -e .
cd FastChat
pip3 install -e ".[model_worker,llm_judge]"
cd ..
export openai api key and base url
export OPENAI_API_KEY="your_api_key"
export OPENAI_API_BASE="your_base_url"bash eval_alpaca.sh
bash eval_mtbench.sh
bash eval_flask.sh
We acknowledge and recommend other excellent works in the Mixture-of-Agents series:
If you find this work useful for your research, please cite our paper:
@article{wen2026attentionmoa,
title={Attention-MoA: Enhancing Mixture-of-Agents via Inter-Agent Semantic Attention and Deep Residual Synthesis},
author={Wen, Jianyu and Wei, Yang and Yu, Xiongxi and Xiao, Changxuan and Zeng, Ke},
journal={arXiv preprint arXiv:2601.16596},
year={2026}
}



