Skip to content

QingyangZhang/EMPO

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

EMPO: Fully Unsupervised LLM Reasoning Incentivization

🤗 HF Models and Datasets Collection | 📑 Arxiv Preprint

For any questions, feel free to open an issue or directly contact to Qingyang Zhang, happy to help and discuss!

If you find this repo helpful, please consider to star🌟 this repo for support our work 🙏🙏

Table of Contents

Overview

EMPO (Entropy Minimized Policy Optimization) does not require any supervised information for incentivizing reasoning capabilities (i.e., neither verifiable reasoning traces, problems with golden answers, nor additional pre-trained reward models). By continuously minimizing the predictive entropy of LLMs on unlabeled user queries, EMPO enables self-supervised RL for reasoning capabilities.

News

  • [2025-04-08] We introduce EMPO, which makes the first attempt on fully unsupervised LLM reasoning incentivization. Check out our arxiv preprint (first released at 2025.04.08): https://arxiv.org/abs/2504.05812
  • [2025-04-30] We release the training and evaluation code for both mathematical reasoning and free-form natural reasoning tasks.
  • [2025-06-10] EMPO was accepted by ICML 2025 Test-time Adaption Workshop. See you in Vancouver!

Repository Structure

This repository contains two self-contained implementations of EMPO:

  • trl: Based on Hugging Face’s trl, a cutting-edge library designed for post-training foundation models.

   ↳ Built on commit v0.14-release

  • verl: Based on VERL, a high-performance RL training library designed for LLMs.

   ↳ Built on commit v0.4x

Both are licensed under Apache 2.0 and include their respective LICENSE and NOTICE files.

TRL Quick Start

Developed upon trl 0.14.0. See trl for details.

cd trl
pip install -r requirements.txt
sh empo-1.5B-NM-COT-20K.sh

Noted that trl 0.14.0 is already a relatively outdated training framework. We will choose verl for further development for efficiency and compatibility.

Verl Quick Start

Developed upon TTRL, with necessary modification to upgrade to the latest verl==0.4.0. See verl for details.

cd verl
USE_MEGATRON=0 bash scripts/install_vllm_sglang_mcore.sh
pip install --no-deps -e .
sh examples/ttrl/empo-math.sh

Evaluation:

The evaluation scripts from the Online-DPO-R1, please refer to the original codebase for more details.

For mathematical tasks:

cd eval_math
sh test.sh

As suggested by Spurious Rewards and Incorrect Baseline, we adopt the same test prompt to both pre-RL Qwen Base models and RL-trained models. Besides, we add Random+Format Reward Baseline for more comprehensive comparison. You can also modify the code here to investigate the influence of different test prompt.

Model Supervision MATH Minerva Math Olympiad Bench AIME24 AMC23 Avg.
1.5B model
Qwen2.5-Math None 52.2 10.7 25.2 10.0 42.5 28.1
Qwen2.5-Math-Instruct ${q, r, a}$ 73.8 30.9 38.7 6.7 52.5 40.5
Qwen2.5-Math w/SFT ${q, r, a}$ 61.8 26.1 27.1 3.3 37.5 31.2
Qwen2.5-Math w/Rand Format ${q, a}$ 65.0 26.1 30.7 10.0 55.0 37.4
Qwen2.5-Math w/GRPO ${q, a}$ 75.2 32.0 33.6 16.7 52.5 42.0
Qwen2.5-Math w/EMPO ${q}$ 73.0 32.4 36.6 13.3 55.0 42.1
7B model
Qwen2.5-Math None 64.8 15.1 26.7 6.7 40.0 30.7
Qwen2.5-Math Instruct ${q, r, a}$ 82.8 43.8 41.2 16.7 62.5 49.4
Qwen2.5-Math w/SFT ${q, r, a}$ 72.2 34.6 33.2 10.0 45.0 39.0
Qwen2.5-Math w/Rand Format ${q, a}$ 73.0 26.5 37.0 26.7 52.5 43.1
Qwen2.5-Math w/ODPO ${q, a}$ 76.8 30.9 37.9 26.7 62.5 47.0
Qwen2.5-Math w/GRPO ${q, a}$ 77.8 39.7 39.1 20.0 57.5 46.8
Qwen2.5-Math w/EMPO ${q}$ 78.0 40.4 37.3 20.0 65.0 48.1

Acknowledgement

This repo is built upon Semantic Entropy, Open-R1, Online-DPO-R1, and TTRL. We thank all these researchers for generously sharing their insights, model weights, data, and codes.

Related Works

There are many awesome works related to this paper that you may also interested with:

More papers are listed in Awesome Reinforcement Learning with Internal Reward Paper list.

Citation

If you find this work helpful, please consider to star🌟 this repo. Thanks for your support!

@article{zhang2025right,
  title={Right Question is Already Half the Answer: Fully Unsupervised LLM Reasoning Incentivization},
  author={Zhang, Qingyang and Wu, Haitao and Zhang, Changqing and Zhao, Peilin and Bian, Yatao},
  journal={arXiv preprint arXiv:2504.05812},
  year={2025}
}

About

EMPO, A Fully Unsupervised RLVR Method

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 5

Languages

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy