1 code implementation • 30 Apr 2024 • Shisen Yue, Siyuan Song, Xinyuan Cheng, Hai Hu
While all models generate largely fluent and self-consistent text, their explanations score low on reasonability except for GPT-4, suggesting that most LLMs cannot produce satisfactory explanations of the implicatures in the conversation.
1 code implementation • 3 Feb 2024 • Byung-Doh Oh, Shisen Yue, William Schuler
Additionally, training dynamics reveal that during later training steps, all model variants learn to predict rare words and that larger model variants do so more accurately, which explains the detrimental effect of both training data amount and model size on fit to reading times.
2 code implementations • 16 Apr 2023 • Yikang Liu, Ziyin Zhang, Wanyang Zhang, Shisen Yue, Xiaojing Zhao, Xinyuan Cheng, Yiwen Zhang, Hai Hu
To address these challenges in English language teaching, we first present ArguGPT, a balanced corpus of 4, 038 argumentative essays generated by 7 GPT models in response to essay prompts from three sources: (1) in-class or homework exercises, (2) TOEFL and (3) GRE writing tasks.