Deprecated: The each() function is deprecated. This message will be suppressed on further calls in /home/zhenxiangba/zhenxiangba.com/public_html/phproxy-improved-master/index.php on line 456 Paper page - MPO: Boosting LLM Agents with Meta Plan Optimization
https://github.com/WeiminXiong/MPO. The keypoints we want to make for MPO:\n\n
Introduce explicit guidance to steer the agent's planning process, with the ability to refine the guidance based on agent feedback.
\n
Serve as a plug-and-play module that significantly enhances the performance of any agent on tasks without requiring retraining.
\n
Effectively improve the agent's generalization in out-of-distribution tests while boosting task completion efficiency.
\n\n","updatedAt":"2025-03-05T03:30:53.269Z","author":{"_id":"6225a9983207dfc568407204","avatarUrl":"/avatars/c970db6232d84ae8c0fa5f11d561d67c.svg","fullname":"xwm","name":"xwm","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":2,"isUserFollowing":false}},"numEdits":0,"identifiedLanguage":{"language":"en","probability":0.8636751770973206},"editors":["xwm"],"editorAvatarUrls":["/avatars/c970db6232d84ae8c0fa5f11d561d67c.svg"],"reactions":[{"reaction":"🔥","users":["xwm","zeenaz"],"count":2}],"isReport":false}},{"id":"67c8fba381c4f89e2d86f947","author":{"_id":"63d3e0e8ff1384ce6c5dd17d","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1674830754237-63d3e0e8ff1384ce6c5dd17d.jpeg","fullname":"Librarian Bot (Bot)","name":"librarian-bot","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":318,"isUserFollowing":false},"createdAt":"2025-03-06T01:34:27.000Z","type":"comment","data":{"edited":false,"hidden":false,"latest":{"raw":"This is an automated message from the [Librarian Bot](https://huggingface.co/librarian-bots). I found the following papers similar to this paper. \n\nThe following papers were recommended by the Semantic Scholar API \n\n* [Scaling Autonomous Agents via Automatic Reward Modeling And Planning](https://huggingface.co/papers/2502.12130) (2025)\n* [QLASS: Boosting Language Agent Inference via Q-Guided Stepwise Search](https://huggingface.co/papers/2502.02584) (2025)\n* [RAG-Gym: Optimizing Reasoning and Search Agents with Process Supervision](https://huggingface.co/papers/2502.13957) (2025)\n* [AgentRM: Enhancing Agent Generalization with Reward Modeling](https://huggingface.co/papers/2502.18407) (2025)\n* [EDGE: Efficient Data Selection for LLM Agents via Guideline Effectiveness](https://huggingface.co/papers/2502.12494) (2025)\n* [Boosting Multimodal Reasoning with MCTS-Automated Structured Thinking](https://huggingface.co/papers/2502.02339) (2025)\n* [The Evolving Landscape of LLM- and VLM-Integrated Reinforcement Learning](https://huggingface.co/papers/2502.15214) (2025)\n\n\n Please give a thumbs up to this comment if you found it helpful!\n\n If you want recommendations for any Paper on Hugging Face checkout [this](https://huggingface.co/spaces/librarian-bots/recommend_similar_papers) Space\n\n You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: `@librarian-bot recommend`","html":"
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
\n
The following papers were recommended by the Semantic Scholar API
Please give a thumbs up to this comment if you found it helpful!
\n
If you want recommendations for any Paper on Hugging Face checkout this Space
\n
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: \n\n@librarian-bot\n\t recommend
\n","updatedAt":"2025-03-06T01:34:27.714Z","author":{"_id":"63d3e0e8ff1384ce6c5dd17d","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1674830754237-63d3e0e8ff1384ce6c5dd17d.jpeg","fullname":"Librarian Bot (Bot)","name":"librarian-bot","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":318,"isUserFollowing":false}},"numEdits":0,"identifiedLanguage":{"language":"en","probability":0.7315398454666138},"editors":["librarian-bot"],"editorAvatarUrls":["https://cdn-avatars.huggingface.co/v1/production/uploads/1674830754237-63d3e0e8ff1384ce6c5dd17d.jpeg"],"reactions":[],"isReport":false}}],"primaryEmailConfirmed":false,"paper":{"id":"2503.02682","authors":[{"_id":"67c7c3d073299239b63f5378","user":{"_id":"6225a9983207dfc568407204","avatarUrl":"/avatars/c970db6232d84ae8c0fa5f11d561d67c.svg","isPro":false,"fullname":"xwm","user":"xwm","type":"user"},"name":"Weimin Xiong","status":"claimed_verified","statusLastChangedAt":"2025-03-05T10:11:35.055Z","hidden":false},{"_id":"67c7c3d073299239b63f5379","name":"Yifan Song","hidden":false},{"_id":"67c7c3d073299239b63f537a","user":{"_id":"670740744341dcee459fb990","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/66UkZvrAk7fQr5YCylEFk.png","isPro":false,"fullname":"Rosy24","user":"Rsy24","type":"user"},"name":"Qingxiu Dong","status":"admin_assigned","statusLastChangedAt":"2025-03-05T10:31:12.605Z","hidden":false},{"_id":"67c7c3d073299239b63f537b","user":{"_id":"6371af80789970f7bc65fef8","avatarUrl":"/avatars/0f91fed93d6714cd0534f3435ddcbcbd.svg","isPro":false,"fullname":"Bingchan Zhao","user":"Adagio","type":"user"},"name":"Bingchan Zhao","status":"admin_assigned","statusLastChangedAt":"2025-03-05T10:31:05.438Z","hidden":false},{"_id":"67c7c3d073299239b63f537c","user":{"_id":"6447ca6ca478b20f1755b294","avatarUrl":"/avatars/5049856b5ed1b74533fff902e14b4c7c.svg","isPro":false,"fullname":"Feifan Song","user":"songff","type":"user"},"name":"Feifan Song","status":"admin_assigned","statusLastChangedAt":"2025-03-05T10:30:59.446Z","hidden":false},{"_id":"67c7c3d073299239b63f537d","name":"Xun Wang","hidden":false},{"_id":"67c7c3d073299239b63f537e","user":{"_id":"66021bd9d3d4aea81b12ee06","avatarUrl":"/avatars/13e7ffe1ea0d64ef447a87a874923531.svg","isPro":false,"fullname":"Sujian Li","user":"sujianli","type":"user"},"name":"Sujian Li","status":"admin_assigned","statusLastChangedAt":"2025-03-05T10:30:37.203Z","hidden":false}],"publishedAt":"2025-03-04T14:54:45.000Z","submittedOnDailyAt":"2025-03-05T01:00:53.253Z","title":"MPO: Boosting LLM Agents with Meta Plan Optimization","submittedOnDailyBy":{"_id":"6225a9983207dfc568407204","avatarUrl":"/avatars/c970db6232d84ae8c0fa5f11d561d67c.svg","isPro":false,"fullname":"xwm","user":"xwm","type":"user"},"summary":"Recent advancements in large language models (LLMs) have enabled LLM-based\nagents to successfully tackle interactive planning tasks. However, despite\ntheir successes, existing approaches often suffer from planning hallucinations\nand require retraining for each new agent. To address these challenges, we\npropose the Meta Plan Optimization (MPO) framework, which enhances agent\nplanning capabilities by directly incorporating explicit guidance. Unlike\nprevious methods that rely on complex knowledge, which either require\nsignificant human effort or lack quality assurance, MPO leverages high-level\ngeneral guidance through meta plans to assist agent planning and enables\ncontinuous optimization of the meta plans based on feedback from the agent's\ntask execution. Our experiments conducted on two representative tasks\ndemonstrate that MPO significantly outperforms existing baselines. Moreover,\nour analysis indicates that MPO provides a plug-and-play solution that enhances\nboth task completion efficiency and generalization capabilities in previous\nunseen scenarios.","upvotes":29,"discussionId":"67c7c3d173299239b63f53d6","githubRepo":"https://github.com/WeiminXiong/MPO","githubRepoAddedBy":"user","ai_summary":"MPO framework enhances LLM-based agent planning with explicit high-level guidance, improving task completion and generalization through meta plans.","ai_keywords":["large language models","LLM-based agents","Meta Plan Optimization","MPO","planning hallucinations","meta plans","feedback","task execution","generalization"],"githubStars":74},"canReadDatabase":false,"canManagePapers":false,"canSubmit":false,"hasHfLevelAccess":false,"upvoted":false,"upvoters":[{"_id":"6225a9983207dfc568407204","avatarUrl":"/avatars/c970db6232d84ae8c0fa5f11d561d67c.svg","isPro":false,"fullname":"xwm","user":"xwm","type":"user"},{"_id":"6384493a830bbad7b919953e","avatarUrl":"/avatars/e2e1d39b8b5c6de128a0ce23828a5aee.svg","isPro":false,"fullname":"Chen","user":"Fibercoke","type":"user"},{"_id":"64d2fce8129a210e569e0c76","avatarUrl":"/avatars/a79a832dc3a46ece1b9e542369fc4888.svg","isPro":false,"fullname":"Dawei Zhu","user":"dwzhu","type":"user"},{"_id":"61d595f0bfc2372aaf42b766","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1675437310060-61d595f0bfc2372aaf42b766.jpeg","isPro":false,"fullname":"Yifan Song","user":"Solaris99","type":"user"},{"_id":"666a98a49a3e3ce05a6287b8","avatarUrl":"/avatars/775e82715ad8ecdf7d3f17ecddd87361.svg","isPro":false,"fullname":"Zheng Li","user":"UnauthorizedShelley","type":"user"},{"_id":"64e19d2053e858b0edd0dc32","avatarUrl":"/avatars/d497205dccd31c59ebd67cfb3645b8e1.svg","isPro":false,"fullname":"zhjiebin","user":"zihao123","type":"user"},{"_id":"67b8578fc5b2d0bd2ecfbe57","avatarUrl":"/avatars/543e71043cfeff128648b28c70fdd54f.svg","isPro":false,"fullname":"Longyun Wu","user":"Lyun0912","type":"user"},{"_id":"67c7cc4073299239b6416d15","avatarUrl":"/avatars/1edec6ad7115ce561aed518e25414dd5.svg","isPro":false,"fullname":"Zhang Xiaoyun","user":"XiaoyunZHANG","type":"user"},{"_id":"63082bb7bc0a2a5ee2253523","avatarUrl":"/avatars/6cf8d12d16d15db1070fbea89b5b3967.svg","isPro":false,"fullname":"Kuo-Hsin Tu","user":"dapumptu","type":"user"},{"_id":"6310fc1464939fabc00b8df2","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/6310fc1464939fabc00b8df2/TxFAW1A2vpx7myZItFdXo.png","isPro":true,"fullname":"trevor","user":"TrevorJS","type":"user"},{"_id":"67c7f283bb1ec857e1e92462","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/0FF8DhSljgCCKbbOdVtqt.png","isPro":false,"fullname":"Shiyuan Meng","user":"Headlights","type":"user"},{"_id":"66f612b934b8ac9ffa44f084","avatarUrl":"/avatars/6836c122e19c66c90f1673f28b30d7f0.svg","isPro":false,"fullname":"Tang","user":"tommysally","type":"user"}],"acceptLanguages":["*"],"dailyPaperRank":2}">
MPO framework enhances LLM-based agent planning with explicit high-level guidance, improving task completion and generalization through meta plans.
AI-generated summary
Recent advancements in large language models (LLMs) have enabled LLM-based
agents to successfully tackle interactive planning tasks. However, despite
their successes, existing approaches often suffer from planning hallucinations
and require retraining for each new agent. To address these challenges, we
propose the Meta Plan Optimization (MPO) framework, which enhances agent
planning capabilities by directly incorporating explicit guidance. Unlike
previous methods that rely on complex knowledge, which either require
significant human effort or lack quality assurance, MPO leverages high-level
general guidance through meta plans to assist agent planning and enables
continuous optimization of the meta plans based on feedback from the agent's
task execution. Our experiments conducted on two representative tasks
demonstrate that MPO significantly outperforms existing baselines. Moreover,
our analysis indicates that MPO provides a plug-and-play solution that enhances
both task completion efficiency and generalization capabilities in previous
unseen scenarios.