Skip to content

Official inference implementation of the paper "DON'T SETTLE TOO EARLY: SELF-REFLECTIVE REMASKING FOR DIFFUSION LANGUAGE MODELS". [ICLR 2026]

Notifications You must be signed in to change notification settings

maple-research-lab/RemeDi

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

5 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

RemeDi: Remasking-enabled Diffusion Language Model

weixin RemeDiΒ 

Static BadgeΒ  Static BadgeΒ 

Official inference implementation of the paper "DON'T SETTLE TOO EARLY: SELF-REFLECTIVE REMASKING FOR DIFFUSION LANGUAGE MODELS".

🚧 Update Progress

[2026-1-28]

  • Model weights for RemeDi have been uploaded to HuggingFace (see resource links below).
  • Inference code is released.

πŸ”¬ Method Overview

RemeDi lets every token be revised at every diffusion step. Instead of fixing in an early guess, the model evaluates the quality of each token and can remask low-confidence positions, allowing later steps to resample them with richer contextβ€”built-in self-correction.

RemeDi extends the original model with a dual-stream transformer:

  • Token Prediction Stream (TPS) predicts masked tokens as usual.

  • Unmasking Policy Stream (UPS) outputs per-token confidence scores, deciding which tokens to unmask or remask.

At each denoising step, tokens with low confidence can be remasked and resampled, enabling iterative refinement. For the training and RL algorithms, see the Methods section of the paper.

RemeDi architecture and performance radar

πŸ“ˆ Key Results

RemeDi performance table

πŸ“‚ Repository Structure

β”œβ”€β”€ inference.py     # inference scripts
β”œβ”€β”€ remedi/          # networks configs 
└── README.md

πŸš€ Inference

To run inference, execute: python inference.py

πŸ“₯ Citation

@article{huang2025don,
  title={Don't Settle Too Early: Self-Reflective Remasking for Diffusion Language Models},
  author={Huang, Zemin and Wang, Yuhang and Chen, Zhiyang and Qi, Guo-Jun},
  journal={arXiv preprint arXiv:2509.23653},
  year={2025}
}

About

Official inference implementation of the paper "DON'T SETTLE TOO EARLY: SELF-REFLECTIVE REMASKING FOR DIFFUSION LANGUAGE MODELS". [ICLR 2026]

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •  

Languages