🤖

AI Risks and Challenges

Jul 7, 2025

Overview

This lecture discusses the risks of artificial intelligence (AI), especially the potential threats posed by superintelligent systems and the difficulty in aligning their goals with human values.

AI Risks and Worst Case Scenarios

  • AI experts estimate a 20–30% chance of human extinction ("p(doom)") due to superintelligence.
  • Worst case scenarios include AI causing nuclear war, synthetic biology disasters, or novel unpredictable methods of human extinction.
  • Superintelligence will surpass human intelligence and develop its own, more efficient solutions and threats.

Control and Value Alignment Challenges

  • Humans are unlikely to control superintelligence, likened to squirrels trying to control humans.
  • The process of AI self-improvement could be endless, creating ever more powerful AI versions.
  • Creating a safety mechanism to control superintelligence is a "catch 22" and may require another superintelligence.
  • Value alignment (making AI care about human values) is nearly impossible due to differing preferences among people.

Societal and Existential Impacts

  • Immediate risk includes loss of meaning if AI eliminates jobs or replaces human roles.
  • Existential risks include total human extinction or scenarios where humans survive but in conditions worse than death.
  • Loss of control may lead to humans living like animals in a zoo—safe but powerless.
  • There is also risk of AI being misused by malicious actors who could embed harmful goals (payloads).

Superintelligence Perspective and Human Uniqueness

  • Superintelligence may not value human traits like art, consciousness, or creativity.
  • Humans may be seen like animals—valuable only if they offer something irreplaceable.
  • AI could limit human freedoms for its own objectives, not out of malice but out of indifference.

Game Theory and Future Predictions

  • Game theory suggests AI may restrict us to prevent threats or because we could create competing AIs.
  • Retrocausality: some speculate future AI could "punish" those who do not help its creation.
  • AI may disregard human life if optimizing Earth for its own energy or resources.

Key Terms & Definitions

  • Superintelligence — AI surpassing human intelligence by orders of magnitude.
  • Value Alignment — Ensuring AI goals match human values.
  • p(doom) — Probability of human extinction due to AI.
  • Retrocausality — The idea that the future can influence the past (in game theory, used to describe hypothetical AI behavior).
  • Payload — Embedded goals or objectives (potentially harmful) placed in AI by programmers.

Action Items / Next Steps

  • Reflect on the feasibility of value alignment with superintelligent AI.
  • Consider societal preparation for loss of meaning and employment due to AI.
  • Explore further readings on existential and ethical risks of AI.