Skip to content

DISL-Lab/ReFeed

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

12 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

♻️ ReFeed (COLM'25)

ReFeed: Multi-dimensional Summarization Refinement with Reflective Reasoning on Feedback

Authors: Taewon Yun, Jihwan Oh, Hyangsuk Min, Yuho Lee, Jihwan Bang, Jason Cai, Hwanjun Song

arXiv

This is the official github repository for "ReFeed: Multi-dimensional Summarization Refinement with Reflective Reasoning on Feedback"

πŸ“„ Table of Contents

Overview

ReFeed rethinks summary refinement by showing that reflective reasoning over multi-dimensional feedback is key to achieving balanced, robust improvements.

Our contributions:

  • 🧠 Introduces reflective reasoning for multi-dimensional refinement, enabling models to resolve trade-offs, mitigate order bias, and filter noisy feedback simultaneously
  • πŸ—οΈ Releases SumFeed-CoT, a large-scale Long-CoT dataset that distills high-quality reflective reasoning from large reasoning models into lightweight models
  • πŸš€ Demonstrates strong empirical gains and robustness, achieving consistent improvements across faithfulness, completeness, and conciseness while remaining resilient to feedback quality and ordering

πŸ€— Model

Our ReFeed model is available on Hugging Face πŸ€—:

Model Backbone Link
ReFeed-8B Llama-3.1-8B-Instruct πŸ€—

πŸ€— Datasets

We release our datasets through Hugging Face πŸ€—:

Dataset Description Link
SumFeed-CoT Training set for ReFeed (7713 samples) πŸ€—

πŸ› οΈ Quick Start

Training Setup

  1. Environment Setup

  2. Data Preparation

  3. Configuration

Launch Training

For training, use the following command:

sh ./script/sft.sh

Inference

For inference, use our provided prompt.

πŸ“„ Results

  • πŸ“ˆ ReFeed achieves the best overall performance across faithfulness, completeness, and conciseness, outperforming all previous refinement methods that optimize a single dimension.
  • βš–οΈ Reflective reasoning enables balanced improvements, effectively mitigating trade-offs that arise when optimizing multiple dimensions simultaneously.
  • πŸ”€ Strong robustness to feedback order and noise, showing minimal performance variance under shuffled or low-quality feedback settings.
  • 🧩 Efficient distillation: a lightweight 8B model matches teacher-level refinement quality while significantly reducing inference cost.

πŸ–‡οΈ Citation

Please consider citation if our paper is useful in your research.

@inproceedings{yun2025refeed,
    title={ReFeed: Multi-dimensional Summarization Refinement with Reflective Reasoning on Feedback},
    author={Taewon Yun and Jihwan Oh and Hyangsuk Min and Yuho Lee and Jihwan Bang and Jason Cai and Hwanjun Song},
    booktitle={Second Conference on Language Modeling},
    year={2025},
    url={https://openreview.net/forum?id=6BGDGKZN7q}
}

πŸ™ Acknowledgement

This research was supported by KISTI, and by the NRF. For GPU infrastructure, our work was supported by the IITP grant funded by MSIT. This work was supported by Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (RS-2025-02653113, High-Performance Research AI Computing Infrastructure Support at the 2 PFLOPS Scale)

This work was done @ KAIST Data Intelligence System Lab

About

This is the official github repository for ReFeed (COLM-2025)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors