Transformer++: a long sequence modeling method based on direction-aware dual attention and multi-head sampling

Ruiqin Wang, Qishun Ji, Zhenzhen Sheng, Yang Qi. Transformer++: a long sequence modeling method based on direction-aware dual attention and multi-head sampling. Appl. Intell., 55(17):1103, November 2025. [doi]

Authors

Ruiqin Wang

This author has not been identified. Look up 'Ruiqin Wang' in Google

Qishun Ji

This author has not been identified. Look up 'Qishun Ji' in Google

Zhenzhen Sheng

This author has not been identified. Look up 'Zhenzhen Sheng' in Google

Yang Qi

This author has not been identified. Look up 'Yang Qi' in Google