Logo
Beyond Hard Masks: Progressive Token Evolution for Diffusion Language Models

1Zhejiang University    2National University of Singapore    3Zhejiang University of Technology   
*Equal contribution Corresponding author


Abstract

Diffusion Language Models (DLMs) offer a promising alternative for language modeling by enabling parallel decoding through iterative refinement. However, most DLMs rely on hard binary masking and discrete token assignments, which hinder the revision of early decisions and underutilize intermediate probabilistic representations. In this paper, we propose EvoToken-DLM, a novel diffusion-based language modeling approach that replaces hard binary masks with evolving soft token distributions. EvoToken-DLM enables a progressive transition from masked states to discrete outputs, supporting revisable decoding. To effectively support this evolution, we introduce continuous trajectory supervision, which aligns training objectives with iterative probabilistic updates. Extensive experiments across multiple benchmarks show that EvoToken-DLM consistently achieves superior performance, outperforming strong diffusion-based and masked DLM baselines.

Overview of EvoToken-DLM
Comparison between MDLMs and EvoToken-DLM. (a) Standard MDLMs employ only two token states, alternating between <mask> and discrete decoded tokens, leading to abrupt mask-to-token transitions. (b) EvoToken-DLM introduces soft tokens represented by probability distributions and four token states, enabling tokens to evolve progressively through iterative refinement. The top-right panel illustrates a quantitative comparison between the two approaches under the same settings based on LLaDA-Instruct-8B.

Motivation

Most MDLMs rely on hard binary masking with discrete token assignments. Once a token is decoded, it is treated as final and excluded from further refinement, resulting in an abrupt transition from uncertainty to determinism. This irreversibility limits the model's ability to revise early decisions and undermines the iterative refinement paradigm of diffusion-based language modeling. In addition, although MDLMs compute token distributions for all positions at each refinement step, only a small subset of positions are updated, while the remaining probabilistic information is discarded.

Illustration of inefficient utilization of predictions in masked diffusion language models
Inefficient utilization of predictions in masked diffusion language models, where distributions are computed for all positions but only a subset are used for decoding. [M1, M2, ..., Mn] denote the initial mask tokens following prompt P, and disti represents the predicted probability distribution for the i-th token in the generation sequence. In this example, the total sequence of 542 tokens consists of 30 prompt tokens and 512 generated tokens, while only two positions are updated per step.

Methodology


Progressive Inference with EvoToken-DLM

We formally define the progressive inference procedure of EvoToken-DLM as follows. Given a prompt P, the objective is to generate a response of length N. The output is partitioned into M = N/B discrete blocks, each of size B. The sequence X is constructed by concatenating the prompt P with N tokens, denoted as X = (P, x1, x2, ..., xN), where each token xi is characterized by a pair (ei, zi), comprising continuous embeddings ei and a token state zi. Initially, all target positions are initialized as mask tokens, where zi = [MASK] for all i ∈ {1, ..., N}, and the corresponding embedding sequence is represented as E = (eP, e<mask>1, ..., e<mask>N). During the evolution process, each token xi transitions through a state space consisting of four distinct stages: $$ \texttt{[MASK]},\ \mathrm{Soft}([\texttt{MASK}] \cup \mathcal{V}),\ \mathrm{Soft}(\mathcal{V}),\ \texttt{[Decode]}, $$ where \(\mathcal{V}\) is the vocabulary.

Progressive step-wise token update with blockwise decoding in EvoToken-DLM.
Progressive step-wise token update with blockwise decoding in EvoToken-DLM.

Token Prediction. At each inference step, we input embeddings E into the model to obtain a predicted distribution \(\{p_i^c\}_{c=1}^{|\mathcal{V}|}\) over the vocabulary for each position i. We retain the top-K probabilities and renormalize them to obtain \(\{\hat{p}_i^c\}_{c=1}^{K}\), along with their corresponding tokens \(\{\hat{v}_i^c\}_{c=1}^{K} \subseteq \mathcal{V}\). Soft embeddings are then computed as: $$ \begin{aligned} e_i^{\text{dist}} &= \sum_{c=1}^{K} \hat{p}_i^c \cdot e^{\hat{v}_i^c}, \\ e_i^{\text{dist+M}} &= \alpha \, e_i^{<\text{mask}>} + (1-\alpha) \, e_i^{\text{dist}}, \end{aligned} $$ where \(\alpha \in [0,1]\) controls the mixing ratio of the mask embedding.

Embedding Assignment by Token State. For token xi, its newly generated embeddings at the current step is assigned based on its current state: $$ e_i = \begin{cases} e_i^{<\text{mask}>}, & z_i = \texttt{[MASK]} \\ e_i^{\text{dist+M}}, & z_i = \mathrm{Soft}([\texttt{MASK}] \cup \mathcal{V}) \\ e_i^{\text{dist}}, & z_i = \mathrm{Soft}(\mathcal{V}) \\ e^{v_i}, & z_i = \texttt{[Decode]} \end{cases} $$ where vi is selected as the token in the vocabulary with the highest confidence among all historical predictions made after xi enters the \(\mathrm{Soft}(\mathcal{V})\) state.

Step-wise Token Update. By default, tokens in the [MASK] state transition to \(\mathrm{Soft}([\texttt{MASK}] \cup \mathcal{V})\), whereas tokens already in the \(\mathrm{Soft}([\texttt{MASK}] \cup \mathcal{V})\), \(\mathrm{Soft}(\mathcal{V})\), or [Decode] states retain their current state. At each step, a subset of tokens currently in the [MASK] or \(\mathrm{Soft}([\texttt{MASK}] \cup \mathcal{V})\) states in the current block is selected to transition to the \(\mathrm{Soft}(\mathcal{V})\) state. Let S denote the set of these selected tokens. The complete update rule is formalized as: $$ z_i \gets \begin{cases} \mathrm{Soft}([\texttt{MASK}] \cup \mathcal{V}), & z_i \in \{\mathrm{Soft}([\texttt{MASK}] \cup \mathcal{V}), \\ & \quad \quad \texttt{[MASK]}\} \text{ and } x_i \notin S \\ \mathrm{Soft}(\mathcal{V}), & x_i \in S \text{ or } z_i = \mathrm{Soft}(\mathcal{V}) \\ \texttt{[Decode]}, & z_i = \texttt{[Decode]} \end{cases} $$

Blockwise Decoding. Let \(\mathcal{B}\) denote the set of tokens in the current block. Once all tokens in \(\mathcal{B}\) reach the \(\mathrm{Soft}(\mathcal{V})\) state, they are simultaneously converted to the [Decode] state: $$ \begin{aligned} z_i &\gets \texttt{[Decode]}, \quad \forall x_i \in \mathcal{B}, \\ &\text{if all tokens } x_j \in \mathcal{B} \text{ are in the } \mathrm{Soft}(\mathcal{V}) \text{ state}. \end{aligned} $$ Combining step-wise token update with blockwise decoding, EvoToken-DLM allows each token to gradually refine its representation from [MASK] to final [Decode] through progressive token evolution.

Continuous Trajectory Supervision

Unlike conventional masked diffusion frameworks, EvoToken-DLM employs a progressive evolution mechanism. In this approach, the current states and embeddings of the tokens are conditioned on the cumulative history of the preceding refinements. This temporal dependency renders standard single-step denoising objectives inapplicable, necessitating a specialized training paradigm that models the trajectory of token evolution. We propose continuous trajectory supervision, a training strategy that aligns model optimization with iterative probabilistic token refinement along the diffusion trajectory \(\mathcal{T}\). This approach ensures consistency from training to inference.

Continuous trajectory supervision
Continuous trajectory supervision by performing Δτ consecutive refinement steps during training and applying supervision at each step, aligning the training objective with the inference process.

Initialization and Masking Strategy. Given a sequence comprising a prompt and a target response, we sample a contiguous segment of length L from the response as the current training block. To align with the blockwise inference procedure, tokens preceding this block are set to the ground truth, while tokens after this block are replaced with [MASK]. Within the selected block, we randomly mask a subset of tokens to initialize the state X(0).

Trajectory Unrolling. Starting from X(0), we simulate \(\Delta \tau\) consecutive refinement steps to sample the trajectory: $$ X^{(i)}, \; \mathcal{L}^{(i)} = \mathrm{Model}(X^{(i-1)}), \quad \forall i = 1, \dots, \Delta \tau, $$ where each forward pass produces probability distributions, updated continuous embeddings, and updated token states according to the progressive inference rules described in Section above.

Cumulative Trajectory Loss. At each step i, we compute a supervised loss \(\mathcal{L}^{(i)}\) against the ground-truth tokens within the current block. Rather than backpropagating only through the final step, we perform a backward pass for every forward step: $$ \nabla_\theta \mathcal{L}^{(i)}, \quad i = 1, \dots, \Delta \tau. $$ By explicitly simulating the progressive refinement during training, continuous trajectory supervision aligns the learning objective with the inference behavior of EvoToken-DLM.


Experimental Results


Results

We employ LLaDA-Instruct-8B as our primary backbone for fine-tuning. For fine-tuning, we utilize the S1K dataset and train the pretrained model for a default duration of 10k steps using continuous trajectory supervision. Evaluations are performed across several mathematical and reasoning benchmarks, including Countdown, GSM8K, MATH500, and SVAMP. For the proposed refinement mechanism, we perform a grid search for the hyperparameter α within the candidate set {0.5, 0.6, 0.7, 0.8, 0.9} and report the performance associated with the optimal α for each setting. We compare EvoToken-DLM against the original LLaDA-Instruct-8B and the FT-baseline across multiple reasoning benchmarks. EvoToken-DLM predominantly surpasses both baselines, exhibiting substantial performance leaps under varying configurations. Specifically, at NFE/Gen Len = 1, our method yields average accuracy gains of 17.45% on Countdown, 3.08% on GSM8K, 2.06% on MATH500, and 3.23% on SVAMP compared to the original model. These results underscore the superiority of our soft token evolution framework in enhancing reasoning capabilities and generation quality.

Performance comparison results
Performance comparison on the Countdown, GSM8K, MATH500 and SVAMP datasets across various generation lengths and NFEs based on LLaDA-Instruct-8B. EvoToken-DLM is initialized from LLaDA-Instruct-8B and fine-tuned for 10k steps using continuous trajectory supervision. Comparisons are conducted against both the baseline model and the sft-baseline.

Qualitative Visualization

We provide a qualitative visualization of the inference process. By tracing the evolution of a selected subsequence, we observe how initial uncertain tokens progressively converge into precise and coherent results. This visualization confirms that EvoToken-DLM effectively implements a progressive refinement mechanism, allowing the model to iteratively calibrate its predictions within the diffusion framework.

Qualitative visualization of inference process
An illustrative example of EvoToken-DLM during inference, showing intermediate refinement states for a selected subsequence across successive steps. The block size is set to 12, and the refinement process for the first 16 output tokens is visualized. For each position, only the top K=3 most probable tokens are retained.

Citation

@article{zhong2026beyond,
  title={Beyond Hard Masks: Progressive Token Evolution for Diffusion Language Models},
  author={Zhong, Linhao and Wu, Linyu and Fang, Bozhen and Feng, Tianjian and Jing, Chenchen and Wang, Wen and Zhang, Jiaheng and Chen, Hao and Shen, Chunhua},
  journal={arXiv preprint arXiv:2601.07351},
  year={2026}
}