no code implementations • 29 Sep 2021 • Jen-Tzung Chien, Yu-Han Huang
To strengthen the sequential learning representation, this paper presents a new disentangled mask attention in transformer where the redundant features are reduced and the semantic information is enriched.