The NLP Task Effectiveness of Long-Range Transformers
Offered By: Center for Language & Speech Processing(CLSP), JHU via YouTube
Course Description
Overview
Explore the effectiveness of long-range Transformer variants in natural language processing tasks through this insightful conference talk from EACL 2023. Delve into a comprehensive study comparing seven Transformer model variants across five challenging NLP tasks and seven datasets. Gain valuable insights into the advantages and previously unrecognized drawbacks of modified attention mechanisms in long-range Transformers. Discover how these models perform in content selection and query-guided decoding, while also learning about their limitations in attending to distant tokens and accumulating approximation errors. Understand the impact of pretraining and hyperparameter settings on model performance, and explore various methods for investigating attention behaviors beyond traditional metric scores. Enhance your knowledge of state-of-the-art NLP models and their real-world applications in this 12-minute presentation by researchers from the Center for Language & Speech Processing (CLSP) at Johns Hopkins University.
Syllabus
The NLP Task Effectiveness of Long-Range Transformers - EACL 2023
Taught by
Center for Language & Speech Processing(CLSP), JHU
Related Courses
Linear CircuitsGeorgia Institute of Technology via Coursera مقدمة في هندسة الطاقة والقوى
King Abdulaziz University via Rwaq (رواق) Magnetic Materials and Devices
Massachusetts Institute of Technology via edX Linear Circuits 2: AC Analysis
Georgia Institute of Technology via Coursera Transmisión de energía eléctrica
Tecnológico de Monterrey via edX