LLM Security: Practical Protection for AI Developers
Offered By: Databricks via YouTube
Course Description
Overview
Explore practical strategies for securing Large Language Models (LLMs) in AI development during this 29-minute conference talk. Delve into the security risks associated with utilizing open-source LLMs, particularly when handling proprietary data through fine-tuning or retrieval-augmented generation (RAG). Examine real-world examples of top LLM security risks and learn about emerging standards from OWASP, NIST, and MITRE. Discover how a validation framework can empower developers to innovate while safeguarding against indirect prompt injection, prompt extraction, data poisoning, and supply chain risks. Gain insights from Yaron Singer, CEO & Co-Founder of Robust Intelligence, on deploying LLMs securely without hindering innovation.
Syllabus
LLM Security: Practical Protection for AI Developers
Taught by
Databricks
Related Courses
AI CTF Solutions - DEFCon31 Hackathon and Kaggle CompetitionRob Mulla via YouTube Indirect Prompt Injections in the Wild - Real World Exploits and Mitigations
Ekoparty Security Conference via YouTube Hacking Neural Networks - Introduction and Current Techniques
media.ccc.de via YouTube The Curious Case of the Rogue SOAR - Vulnerabilities and Exploits in Security Automation
nullcon via YouTube Mastering Large Language Model Evaluations - Techniques for Ensuring Generative AI Reliability
Data Science Dojo via YouTube