Abstract
MemDLM addresses the train-inference mismatch in diffusion language models by incorporating a bi-level optimization framework with parametric memory that enhances both training efficiency and inference performance through adaptive denoising processes.
Diffusion Language Models (DLMs) offer attractive advantages over Auto-Regressive (AR) models, such as full-attention parallel decoding and flexible generation. However, they suffer from a notable train-inference mismatch: DLMs are trained with a static, single-step masked prediction objective, but deployed through a multi-step progressive denoising trajectory. We propose MemDLM (Memory-Enhanced DLM), which narrows this gap by embedding a simulated denoising process into training via Bi-level Optimization. An inner loop updates a set of fast weights, forming a Parametric Memory that captures the local trajectory experience of each sample, while an outer loop updates the base model conditioned on this memory. By offloading memorization pressure from token representations to parameters, MemDLM yields faster convergence and lower training loss. Moreover, the inner loop can be re-enabled at inference time as an adaptation step, yielding additional gains on long-context understanding. We find that, when activated at inference time, this Parametric Memory acts as an emergent in-weight retrieval mechanism, helping MemDLM further reduce token-level attention bottlenecks on challenging Needle-in-a-Haystack retrieval tasks. Code: https://github.com/JarvisPei/MemDLM.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- MetaState: Persistent Working Memory for Discrete Diffusion Language Models (2026)
- GradMem: Learning to Write Context into Memory with Test-Time Gradient Descent (2026)
- Dynamic Long Context Reasoning over Compressed Memory via End-to-End Reinforcement Learning (2026)
- Residual Context Diffusion Language Models (2026)
- DyLLM: Efficient Diffusion LLM Inference via Saliency-based Token Selection and Partial Attention (2026)
- VDLM: Variable Diffusion LMs via Robust Latent-to-Text Rendering (2026)
- Training Large Reasoning Models Efficiently via Progressive Thought Encoding (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Models citing this paper 2
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper