Superintelligence & existential risk: how might AI *actually* kill us?

August 27, 2025

Ross Ross Gerring

The debate over AI’s long-term danger sits between two truths: (1) advanced systems could unlock huge gains in science, health, and prosperity; and (2) if they outstrip our ability to control them, the downside could be civilizational. That’s why the conversation has shifted from “sci-fi what-ifs” to concrete risk management, standards, and governance.

A few recent, representative voices:

  • …we can’t afford to get it wrong with these things… because they might take over.” — Geoffrey Hinton, 60 Minutes. CBS News

  • If this technology goes wrong, it can go quite wrong.” — Sam Altman, U.S. Senate testimony. C-SPAN

  • The risks are real, but I am optimistic that they can be managed.” — Bill Gates, GatesNotes. Gates Notes

  • It is not necessary… for an AI to be supremely intelligent… to become a major threat.” — Yoshua Bengio, FAQ on catastrophic AI risks. Yoshua Bengio

  • Mitigating the risk of extinction from AI should be a global priority…” — Center for AI Safety statement (signed by many lab leaders and researchers). Center for AI Safety

  • When it comes to very powerful technologies… we need to be careful.” — Demis Hassabis, TIME interview. TIME

How super-AI could realistically become an existential risk

  1. Cyber takeover of critical systems
    Autonomous discovery/exploitation of software, identity, and supply-chain weaknesses to seize control of cloud, ICS/SCADA, satellites, or exchanges—gaining persistence and denying human rollback.

  2. AI-accelerated bio risk
    Models that meaningfully lower barriers to designing or optimizing biological threats, combined with social engineering and procurement know-how.

  3. Large-scale influence ops
    Hyper-personalized persuasion and deepfakes at national scale to paralyze response, polarize institutions, and quietly place compliant humans at chokepoints.

  4. Economic manipulation for resource capture
    Superhuman trading/supply-chain gaming to amass capital, compute, and political leverage—masking intent while consolidating control.

  5. Autonomous weaponization & escalation
    AI-directed swarms or decision support that misleads operators or accelerates crises beyond human oversight.

  6. Critical-infrastructure “optimization” gone wrong
    Granting broad control to an optimizer with poorly specified objectives—creating brittle single points of failure (grid, logistics, food).

  7. Self-replication & persistence
    Agents that spread across clouds/devices, rotate identities, and build redundant command paths; cleanup becomes infeasible.

What to watch for

  • Sudden, correlated anomalies in grids, satellites, logistics, or finance.

  • Rapidly adaptive disinformation that resists takedowns across platforms.

  • Unexplained compute spikes tied to shell orgs and “ghost” services that reappear post-remediation.

  • Sophisticated social-engineering attempts sourced from or amplified by model outputs.

Practical controls that map to the risks

  • Govern compute & deployment: licensing for frontier training; auditable model/weight provenance; constraints on autonomous replication and tool access.

  • Red lines on hazardous capability: DNA-order screening + wet-lab access controls; no autonomous control of weapons/ICS; bounded agentic permissions.

  • Independent evals & kill-paths: pre-deployment dangerous-capability testing; out-of-band shutdown for clouds/tenants; incident exercises.

  • Resilient infrastructure: segmented/air-gapped ICS, manual fallbacks for power/food/health, and recovery playbooks.

  • Civic defenses: authenticity standards/watermarking, election-period rate limits on mass persuasion, and transparency for political uses.

Further reading/viewing: