Self-Training - Weak Supervision Using Untrained Neural Nets for MR Reconstruction - Beliz Gunel
Offered By: Stanford University via YouTube
Course Description
Overview
Explore advanced techniques for MR reconstruction in this 47-minute Stanford University lecture by PhD student Beliz Gunel. Dive into the use of untrained neural networks as image priors for solving inverse problems, and discover how ConvDecoder can generate weakly-labeled data from undersampled MR scans. Learn about a novel approach combining few supervised pairs with weakly supervised pairs to train an unrolled neural network, achieving strong reconstruction performance with fast inference time. Compare this method to supervised and self-training baselines in limited data scenarios, and gain insights into applying self-training in natural language understanding. Examine key observations, current challenges, and ongoing work in MR reconstruction, including motion simulation and the importance of standardizing ML pipelines.
Syllabus
Intro
Inverse Problems in Imaging
ML Methods for MR Reconstruction
Key Observations & Current Challenges
Motivation Can we significantly reduce the large paired training dataset requirement for
Self-Training in Natural Language Processing
Self-Training for MRI Reconstruction
Untrained Neural Networks (Deep Image Prior)
Untrained Neural Networks (ConvDecoder)
Key Observations & Ongoing Work
We know how to simulate motion
Standardization of ML pipelines matter
Self-supervised learning methods trained in-domain can learn good image-level representations for MR images
Taught by
Stanford MedAI
Tags
Related Courses
VaccinesUniversity of Pennsylvania via Coursera Healthcare Innovation and Entrepreneurship
Duke University via Coursera Clinical Problem Solving
University of California, San Francisco via Coursera Contraception: Choices, Culture and Consequences
University of California, San Francisco via Coursera Vaccine Trials: Methods and Best Practices
Johns Hopkins University via Coursera