YoVDO

Learning the Structure of EHR with Graph Convolutional Transformer - Edward Choi

Offered By: Stanford University via YouTube

Tags

Machine Learning Courses Health Care Courses Electronic Health Records Courses Self-Attention Mechanisms Courses

Course Description

Overview

Explore the potential of graph convolutional transformers in learning the structure of electronic health records (EHR) in this comprehensive lecture by Edward Choi. Delve into the graphical nature of EHR data stored in relational databases and discover how combining graph convolution with self-attention can enhance supervised prediction tasks. Gain insights into recent developments in multi-modal learning using Transformers and understand the application of interpretable deep learning methods for longitudinal electronic health records. Learn about hierarchical representations, regularization techniques, visit complexity, diagnosis code prediction, and the self-attention mechanism. Examine case studies, qualitative examples, and meaningful structures derived from learning the graphical structure of EHR data.

Syllabus

Introduction
What is EHR
Case Study
Structure of EHR
Hierarchical representation
Comparison
Regularization
Visit Complexity
Diagnosis Code Prediction
Selfattention mechanism
QKB operation
Learning the graphical structure
Learning meaningful structures
Qualitative examples
Conclusion


Taught by

Stanford MedAI

Tags

Related Courses

Axial-DeepLab - Stand-Alone Axial-Attention for Panoptic Segmentation
Yannic Kilcher via YouTube
Linformer - Self-Attention with Linear Complexity
Yannic Kilcher via YouTube
Synthesizer - Rethinking Self-Attention in Transformer Models
Yannic Kilcher via YouTube
The Narrated Transformer Language Model
Jay Alammar via YouTube
Knowledge Is Embedded in Language Neural Networks but Can They Reason?
Simons Institute via YouTube