Learning to Control Safety-Critical Systems
Offered By: Simons Institute via YouTube
Course Description
Overview
Explore a comprehensive lecture on the challenges and potential of integrating AI tools into safety-critical systems. Delve into the work of Adam Wierman from the California Institute of Technology as he addresses the crucial question of providing guarantees for black-box AI tools in critical applications. Examine the structure of constraints in sequential decision-making and discover various projects aimed at developing robust, localizable tools that combine model-free and model-based approaches. Learn about the efforts to create AI tools with formal guarantees on performance, stability, safety, and sample complexity for use in data centers, electricity grids, transportation, and other vital sectors. Gain insights into overcoming the limitations of machine-learned algorithms, such as lack of worst-case performance guarantees and difficulties in distributed, networked settings. Understand the importance of addressing distribution shift and global information availability in local controllers to ensure reliable AI implementation in safety-critical networked systems.
Syllabus
Learning to Control Safety-Critical Systems
Taught by
Simons Institute
Related Courses
Safety and Robustness for Deep Learning with Provable Guarantees - Marta Kwiatkowska - OxfordAlan Turing Institute via YouTube Applicable and Achievable Formal Verification
USENIX via YouTube Hypervisor-less Virtio for Real-time and Safety
Linux Foundation via YouTube Collaborate on Linux for Use in Safety-Critical Systems
Linux Foundation via YouTube Cybersecurity Risk Assessment for Safety-Critical Systems
Black Hat via YouTube