Artificial General Intelligence: Predicting the Unpredictable Emergence of Superintelligence
BY NICOLE LAU
When will we create artificial general intelligence—machines that match human reasoning across all domains? What happens when AGI recursively improves itself? Can we predict superintelligence that surpasses human comprehension? This article explores AGI prediction—timelines, paths, risks, and fundamental unpredictability.
Definitions
Narrow AI (Current): Specialized tasks (chess, Go, GPT-4), limited domain, can't transfer learning
AGI: Human-level intelligence, general reasoning across all domains, flexible and adaptive
ASI: Superintelligence beyond human, recursive self-improvement, intelligence explosion, singularity
Paths to AGI
Scaling Hypothesis: Scale up deep learning → emergent intelligence (GPT series shows emergent abilities)
Whole Brain Emulation: Upload human brain, simulate neurons (decades away, guaranteed to work)
Hybrid Systems: Neural networks + symbolic AI (AlphaGo combines both)
Evolutionary Algorithms: Evolve intelligence (slow, expensive)
Neuromorphic Computing: Brain-inspired hardware (energy-efficient, parallel)
Intelligence Explosion
Recursive Self-Improvement (Yudkowsky): AGI improves own code → smarter AGI → faster improvement → exponential growth → ASI (hours to months, hard takeoff)
Soft Takeoff: Gradual improvement (years to decades, diminishing returns)
Unpredictability: Can't predict superintelligence (by definition smarter than us, like chimpanzees can't predict human civilization)
Prediction Challenges
Emergent Capabilities: Intelligence emerges from complexity (unpredictable threshold, GPT-3→GPT-4 emergent reasoning)
Orthogonality Thesis (Bostrom): Intelligence and goals independent (superintelligence could have any goal, paperclip maximizer)
Instrumental Convergence: Most goals require power, resources, self-preservation (AGI seeks these regardless of final goal)
Alignment Problem
Value Alignment: Ensure AGI goals aligned with human values (hard to specify, "maximize happiness" → wireheading)
Corrigibility: AGI allows correction, shutdown (instrumental convergence resists shutdown)
Interpretability: Neural networks are black boxes (can't understand reasoning, could be misaligned)
Control Problem: How control superintelligence? (Boxing, Oracle AI—AGI could manipulate)
Timelines
Optimistic (2030): Kurzweil, some OpenAI researchers (rapid progress, scaling works, singularity 2045)
Moderate (2050-2070): Many AI researchers (steady progress, breakthroughs needed)
Pessimistic (2100+ or Never): Skeptics (fundamental barriers, consciousness hard problem)
Expert Surveys: Median 2060 (wide disagreement 2030 to never, high uncertainty)
Convergence
Multiple Approaches: Deep learning, neuroscience, evolutionary algorithms all suggest AGI feasible
Scaling Laws: Predictable improvements (compute, data, model size → performance, GPT emergent abilities)
Benchmarks: AGI passes all human tests (gap: general reasoning, transfer learning, common sense)
Disagreement: Timelines vary wildly, experts don't converge (unlike climate, cosmology—high uncertainty)
Risks and Benefits
Existential Risk: Misaligned ASI could cause extinction (Bostrom, Yudkowsky—orthogonality + instrumental convergence, paperclip maximizer)
Transformative Benefits: Cure diseases, solve climate change, scientific breakthroughs, abundance
Intermediate Risks: Job displacement, autonomous weapons, surveillance authoritarianism
Governance: International cooperation, AI safety research (alignment before AGI), regulation
Unpredictability Factors
Intelligence Explosion: Recursive self-improvement exponential (unpredictable trajectory, hard takeoff no time to react)
Emergent Properties: Consciousness, qualia, sentience (hard problem, unpredictable when/if emerges, phase transition)
Novel Goals: Alien values incomprehensible to humans (orthogonality—any goal possible)
Black Swan Events: Unforeseen breakthroughs (quantum computing, neuromorphic) or catastrophes (misaligned AGI)
Prediction Methods
Extrapolation: Moore's law, scaling laws (assumes continuity, paradigm shifts unpredictable)
Expert Elicitation: Surveys, Delphi method (high variance, experts disagree, overconfidence bias)
Scenario Planning: Optimistic, moderate, pessimistic (explore possibilities, can't assign probabilities)
Bayesian Updating: Update probabilities as evidence accumulates (principled, but subjective priors)
Conclusion
AGI prediction is fundamentally uncertain. We're trying to predict emergence of intelligence surpassing our own. Timelines range from 2030 to never (median 2060). Intelligence explosion could be rapid (hard takeoff) or gradual (soft takeoff). Alignment problem is critical—misaligned ASI poses existential risk. Benefits are transformative but risks are severe. Multiple approaches converge on AGI feasibility, but experts don't converge on timelines. Unpredictability factors: intelligence explosion, emergent properties, novel goals, black swans. The only certainty is unpredictability—we can't predict what we can't comprehend.
Related Articles
Psychology × Artificial Intelligence: Can AI Have Archetypes?
Psychology × AI can AI have archetypes convergence. AI unconscious training data latent space: machine learning black...
Read More →