no code implementations • 20 Dec 2019 • Ta-Chun Su, Guan-Ying Chen
Therefore, we created a new approach, ET-USB, that incorporates users' sequential and nonsequential features; we apply the powerful Transformer encoder, a self-attention network model, to capture the information underlying user behavior sequences.
no code implementations • ICLR 2020 • Ta-Chun Su, Hsiang-Chih Cheng
However, in many downstream tasks, information between layers is ignored by BERT for fine-tuning.