YoVDO

Text Dependencies: Information Cascades, Translations and Multi-Input Attention - 2018

Offered By: Center for Language & Speech Processing(CLSP), JHU via YouTube

Tags

Computational Linguistics Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore text dependencies, information cascades, and multi-input attention in this comprehensive lecture by David Smith from the Center for Language & Speech Processing at Johns Hopkins University. Delve into methods for inferring and exploiting dependency structures in texts, including a new directed spanning tree model for information cascades and a contrastive training procedure. Learn about extracting parallel passages from multilingual corpora using polylingual topic models to improve translation systems. Discover techniques for detecting and correcting multiple transcriptions of the same passage in noisy OCR data using multi-input attention models. Gain insights into applications of natural language processing in information retrieval, social sciences, and humanities, as well as computational linguistic models of structure learning and historical change.

Syllabus

Text Dependencies: Information Cascades, Translations and Multi-Input Attention - David Smith 2018


Taught by

Center for Language & Speech Processing(CLSP), JHU

Related Courses

Miracles of Human Language: An Introduction to Linguistics
Leiden University via Coursera
Language and Mind
Indian Institute of Technology Madras via Swayam
Text Analytics with Python
University of Canterbury via edX
Playing With Language
TED-Ed via YouTube
Computational Language: A New Kind of Science
World Science U