Skip to main content

Spring Deadline: Sunday, March 1 @ 11:59pm PT. Click here to apply.

Back to Research
Accepted to NAACL SRW 2025

Introducing MAPO: Momentum-Aided Gradient Descent Prompt Optimization

Anthony Cui, Pranav Nandyalam, Ethan Cheung, Aiden Lei

Abstract

Momentum-Aided Prompt Optimization (MAPO) enhances the efficiency and efficacy of prompt optimization for Large Language Models (LLMs). Building on ProTeGi, MAPO uses positive natural language "gradients" and a momentum-based extension to refine prompts effectively. By tracking gradient history, MAPO avoids local minima and oscillations. It also utilizes beam search and an Upper Confidence Bound (UCB) algorithm for balanced candidate expansion and selection. Benchmark testing shows that MAPO achieves faster convergence time with fewer API calls and higher F1 scores than ProTeGi, proving it as a robust and scalable solution for automated prompt engineering in LLMs.

Citation

Anthony Cui, Pranav Nandyalam, Ethan Cheung, Aiden Lei. "Introducing MAPO: Momentum-Aided Gradient Descent Prompt Optimization". Accepted to NAACL SRW 2025.

Details

Conference
Accepted to NAACL SRW 2025
Authors
4 authors

Publish Your Research

Join Algoverse and work with world-class mentors to publish at top AI conferences.

Start Your Application