Skip to content

ErikZ719/CoTA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Context Tokens are Anchors: Understanding the Repeat Curse in dMLLMs from an Information Flow Perspective

arXiv

Context Tokens are Anchors: Understanding the Repeat Curse in dMLLMs from an Information Flow Perspective
ICLR 2026
Qiyan Zhao, Xiaofeng Zhang*, Shuochen Chang, Qianyu Chen, Xiaosong Yuan, Xuhang Chen, Luoqi Liu, Jiajun Zhang, Xu-Yao Zhang, Da-Han Wang

🛠️ Install

git clone https://github.com/ErikZ719/CoTA
cd CoTA/train
bash init_env.sh
python generate_demo.py

The motivation of this paper is

image

The patterns of repeat tokens under kv-cache are as follows:

image

The layer wise of LLaDA-V e are as follows:

image

After CoTA, repeat tokens became normal tokens, and the attention map are:

image

Citation

@inproceedings{zhao2026context,
  title={Context Tokens are Anchors: Understanding the Repeat Curse in dMLLMs from an Information Flow Perspective},
  author={Zhao, Qiyan and Zhang, Xiaofeng and Chang, Shuochen and Chen, Qianyu and Yuan, Xiaosong and Chen, Xuhang and Liu, Luoqi and Zhang, Jiajun and Zhang, Xu-Yao and Wang, Da-Han},
  booktitle={The Fourteenth International Conference on Learning Representations (ICLR)},
  year={2026}
}

Acknowledgement

This repo is built on LLaDA-V (models) and dLLM-Cache . Many thanks for their efforts. The use of our code should also follow the original licenses.

About

[ICLR 26] Context Tokens are Anchors: Understanding the Repeat Curse in dMLLMs from an Information Flow Perspective

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors