What You Need to Know Before
You Start
Starts 22 June 2025 02:36
Ends 22 June 2025
00
days
00
hours
00
minutes
00
seconds
59 minutes
Optional upgrade avallable
Not Specified
Progress at your own speed
Free Video
Optional upgrade avallable
Overview
Explore the safety challenges in AI, focusing on out-of-distribution issues and safety guarantees for large language models with Aditi Raghunathan.
Syllabus
- Introduction to AI Safety
- Out-of-Distribution (OOD) Issues
- Theoretical Foundations
- Large Language Models (LLMs)
- Safety Guarantees in AI
- Techniques for Enhancing Safety
- Case Studies
- Ethical Considerations
- Practical Workshops
- Future Directions and Open Research Challenges
- Course Wrap-up
Overview of AI safety concerns
Importance of addressing safety in AI systems
Definition and examples of OOD
Impact of OOD on AI system performance
Strategies for detecting OOD data
Statistical and probabilistic foundations of OOD
Robustness in AI models
Evaluation metrics for OOD scenarios
Introduction to large language models
Common use cases and applications
Limitations and failure modes
Definition and examples of safety guarantees
Approaches for ensuring safety in AI models
Verification and validation techniques
Robust training methods
Adversarial training and defenses
Model interpretability and trustworthiness
Analysis of real-world AI failures
Lessons learned and safety improvements
Ethical implications of AI safety
Balancing performance with safety
Hands-on exercises with open-source tools
Simulations of OOD scenarios and safety assessments
Emerging trends in AI safety
Key areas for further research and development
Review and discussion of key concepts
Final thoughts on the future of AI safety and OOD challenges
Subjects
Computer Science