This is the official github repository for "ReFeed: Multi-dimensional Summarization Refinement with Reflective Reasoning on Feedback"
ReFeed rethinks summary refinement by showing that reflective reasoning over multi-dimensional feedback is key to achieving balanced, robust improvements.
Our contributions:
- π§ Introduces reflective reasoning for multi-dimensional refinement, enabling models to resolve trade-offs, mitigate order bias, and filter noisy feedback simultaneously
- ποΈ Releases SumFeed-CoT, a large-scale Long-CoT dataset that distills high-quality reflective reasoning from large reasoning models into lightweight models
- π Demonstrates strong empirical gains and robustness, achieving consistent improvements across faithfulness, completeness, and conciseness while remaining resilient to feedback quality and ordering
Our ReFeed model is available on Hugging Face π€:
| Model | Backbone | Link |
|---|---|---|
| ReFeed-8B | Llama-3.1-8B-Instruct | π€ |
We release our datasets through Hugging Face π€:
| Dataset | Description | Link |
|---|---|---|
| SumFeed-CoT | Training set for ReFeed (7713 samples) | π€ |
-
Environment Setup
- We recommend following the environment setup in the open-r1 documentation on huggingface
- Ensure all dependencies are properly installed and configured.
-
Data Preparation
- Obtain the SumFeed-CoT dataset from π€ Hugging Face.
-
Configuration
- Use our provided configuration file.
For training, use the following command:
sh ./script/sft.shFor inference, use our provided prompt.
- π ReFeed achieves the best overall performance across faithfulness, completeness, and conciseness, outperforming all previous refinement methods that optimize a single dimension.
- βοΈ Reflective reasoning enables balanced improvements, effectively mitigating trade-offs that arise when optimizing multiple dimensions simultaneously.
- π Strong robustness to feedback order and noise, showing minimal performance variance under shuffled or low-quality feedback settings.
- π§© Efficient distillation: a lightweight 8B model matches teacher-level refinement quality while significantly reducing inference cost.
Please consider citation if our paper is useful in your research.
@inproceedings{yun2025refeed,
title={ReFeed: Multi-dimensional Summarization Refinement with Reflective Reasoning on Feedback},
author={Taewon Yun and Jihwan Oh and Hyangsuk Min and Yuho Lee and Jihwan Bang and Jason Cai and Hwanjun Song},
booktitle={Second Conference on Language Modeling},
year={2025},
url={https://openreview.net/forum?id=6BGDGKZN7q}
}This research was supported by KISTI, and by the NRF. For GPU infrastructure, our work was supported by the IITP grant funded by MSIT. This work was supported by Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (RS-2025-02653113, High-Performance Research AI Computing Infrastructure Support at the 2 PFLOPS Scale)

