What You Need to Know Before
You Start
Starts 6 July 2025 01:17
Ends 6 July 2025
Adversarial Training for LLMs' Safety Robustness
Simons Institute
2777 Courses
1 hour 1 minute
Optional upgrade avallable
Not Specified
Progress at your own speed
Free Video
Optional upgrade avallable
Overview
Discover cutting-edge techniques designed to bolster the safety and robustness of Large Language Models (LLMs) by engaging in adversarial training methods. Led by the esteemed researcher Gauthier Gidel from IVADO-Mila, this session is a must-watch for those interested in the intersection of artificial intelligence and computer science.
Enhance your understanding and skills in LLMs safety robustness today.
Available through YouTube, this course is categorized under Artificial Intelligence and Computer Science Courses, providing invaluable insights for learners and professionals seeking to deepen their expertise in AI safety measures.
Syllabus
- Introduction to Adversarial Training
- Fundamentals of Large Language Models (LLMs)
- Understanding Adversarial Attacks
- Adversarial Training Techniques
- Improving Safety Robustness in LLMs
- Practical Implementation of Adversarial Training
- Case Studies and Applications
- Challenges and Future Directions
- Wrap-up and Discussion
- Additional Resources
Subjects
Computer Science