AIAP: An Overview of Technical AI Alignment with Rohin Shah (Part 2)
Listen to Episode Here Show Notes The space of AI alignment research is highly dynamic, and it's often
View episode
Listen to Episode Here Show Notes The space of AI alignment research is highly dynamic, and it's often
View episode
Listen to Episode Here Show Notes The space of AI alignment research is highly dynamic, and it's often
View episode
Listen to Episode Here Show Notes "To make AI systems broadly useful for challenging real-world tasks, we need them
View episode
Listen to Episode Here Show Notes "How do we combine concepts to form thoughts? How can the same thought
View episode
Listen to Episode Here Show Notes What motivates cooperative inverse reinforcement learning? What can we gain from recontextualizing our safety
View episode
Listen to Episode Here Show Notes What role does inverse reinforcement learning (IRL) have to play in AI alignment? What
View episode
Listen to Episode Here Show Notes Are there such things as moral facts? If so, how might we be able
View episode
Listen to Episode Here Show Notes How are we to make progress on AI alignment given moral uncertainty? What are
View episode
Listen to Episode Here Show Notes What role does metaethics play in AI alignment and safety? How might paths to
View episode
Listen to Episode Here Show Notes What role does cyber security play in AI alignment and safety? What is AI
View episode
Listen to Episode Here Show Notes In a classic taxonomy of risks developed by Nick Bostrom (seen below), existential risks
View episode
Listen to Episode Here Show Notes Inverse Reinforcement Learning and Inferring Human Preferences is the first podcast in the new
View episodeNo matter your level of experience or seniority, there is something you can do to help us ensure the future of life is positive.