Kavli Affiliate: Jing Wang | First 5 Authors: Jing Wang, Ao Ma, Jiasong Feng, Dawei Leng, Yuhui Yin | Summary: The global self-attention mechanism in diffusion transformers involves redundant computation due to the sparse and redundant nature of visual information, and the attention map of tokens within a spatial window shows significant similarity. To address […]
Continue.. Qihoo-T2X: An Efficient Proxy-Tokenized Diffusion Transformer for Text-to-Any-Task