LLM Security 101 - Risks, Attacks, and Mitigation Strategies
Offered By: Trelis Research via YouTube
Course Description
Overview
Syllabus
LLM Security Risks
Video Overview
Resources and Scripts
Installation and Server Setup
Jailbreak attacks to avoid Safety Guardrails
Detecting jailbreak attacks
Llama Guard and its prompt template
Llama Prompt Guard
Testing Jailbreak Detection
Testing for false positives with Llama Guard
Off-topic Requests
Prompt Injection Attacks Container escape, File access / deletion, DoS
1. Detecting Injection Attacks with a Custom Guard
Preventing Injection Attacks via User Authentication
37 Using Prepared Statements to avoid SQL Injection Attacks
Response Sanitisation to avoid Injection Attacks
Malicious Code Attacks
Building a custom classifier for malicious code
Using Codeshield to detect malicious code
Malicious Code Detection Performance
Effect of Guards/shields on Response Time / Latency
Final Tips
Resources
Taught by
Trelis Research
Related Courses
AI CTF Solutions - DEFCon31 Hackathon and Kaggle CompetitionRob Mulla via YouTube Indirect Prompt Injections in the Wild - Real World Exploits and Mitigations
Ekoparty Security Conference via YouTube Hacking Neural Networks - Introduction and Current Techniques
media.ccc.de via YouTube The Curious Case of the Rogue SOAR - Vulnerabilities and Exploits in Security Automation
nullcon via YouTube Mastering Large Language Model Evaluations - Techniques for Ensuring Generative AI Reliability
Data Science Dojo via YouTube