- The AI Pulse
- Posts
- š§ Is Predicting Existential Risks From Advanced AI Models Pointless?
š§ Is Predicting Existential Risks From Advanced AI Models Pointless?
PLUS: What Fears of a Black Hole Consuming Earth Teaches Us About Predicting Existential Risks
Welcome back AI prodigies!
In todayās Sunday Special:
š§ØAn AI-Powered Catastrophic Event, Described.
šHow Do Policymakers Predict the Odds?
šChallenges of Inductive Estimation Methods
šShortcomings of Deductive Estimation Methods
š„½Limitations of Subjective Estimation Methods
šKey Takeaway
Read Time: 7 minutes
šKey Terms
Cyberattacks: An intentional effort to steal, expose, or disable data or gain unauthorized access to computers and networks.
Base Rate: The underlying probability of an event occurring without considering any other factors.
Reference Class: A method for predicting the future by looking at similar past situations and their outcomes.
Generative AI (GenAI): Uses AI models pre-trained on text, image, audio, video, or code data to generate new content.
Artificial General Intelligence (AGI): A theoretical concept where AI systems achieve human-level learning, perception, and cognitive ability.
š©ŗ PULSE CHECK
In your opinion, is it useful to estimate the odds of catastrophic events?Vote Below to View Live Results |
š§ØAN AI-POWERED CATASTROPHIC EVENT, DESCRIBED.
The possibility of existential risks from advanced AI models has garnered significant attention in recent years. Tech billionaire Elon Musk explained: āWeāre building advanced AI models much smarter than we can comprehend. Itās like having a child thatās a super genius you know will be smarter than you.ā
One frequently discussed threat is AI-powered Cyberattacks, which automate and enhance traditional Cyberattacks by making them more targeted, sophisticated, and challenging to detect.
What Is the Use Case?
For example, OpenAIās ChatGPT has given bad actors the ability to craft, test, and send highly personalized Spear Phishing emails that exploit a personās āroutines, interests, or characteristicsā to access their sensitive information.
Thatās just the tip of the iceberg. In the future, AI-powered Cyberattacks will autonomously perform:
Obfuscation: Employ techniques such as rerouting commands through multiple compromised computers to avoid being traced.
Malware Creation: Deploy AI-powered algorithms to design malware tailored to exploit the weaknesses of specific computers and networks.
Execution: Shut down critical elements of computers and networks to impact essential infrastructure like power grids or transportation systems (e.g., air traffic control).
What Are the Actual Chances?
Earlier last year, the U.S. Department of State collaborated with Gladstone AI to develop an action plan outlining catastrophic national security risks posed by advanced AI models. It detailed several AI-powered threats, including AI-powered Cyberattacks on Americaās power grids. But what are the actual chances of this happening?
šHOW DO POLICYMAKERS PREDICT THE ODDS?
Policymakers often rely on estimation methods provided by statisticians. However, the reliability of these estimates is questionable, raising concerns about their utility in informing policy decisions. Assigning precise probabilities to complex, unprecedented events like AI-powered Cyberattacks on Americaās power grids can create a false sense of certainty. Estimation Methods often rest on speculative assumptions about advanced AI models and how bad actors may deploy them irresponsibly. Such predictions can obscure the issueās complexity, causing overconfidence and misallocating resources to combat the wrong threats.
Nevertheless, Pseudo-Quantification, or generating error-prone percentages, is a dominant approach to estimating the probability of catastrophic events. For example, predicting a 5% chance of global conflict within the next 5 years.
Policymakers and statisticians use inductive, deductive, and subjective estimation methods to determine the possibility of catastrophic events occurring in the future. Today, weāll focus on how each estimation method prevents statisticians from generating accurate and actionable assessments of existential risks from advanced AI models.
šCHALLENGES OF INDUCTIVE ESTIMATION METHODS
Inductive estimation methods rely on historical data to predict future catastrophic events. For example, meteorologists predict the likelihood of a hurricane hitting a specific area in weather forecasts by analyzing past hurricane patterns and historical weather data. However, existential risks from advanced AI models lack a relevant Reference Class.
Unlike hurricane prediction, which builds on decades of empirical data, thereās no precedent for catastrophic events perpetrated by advanced AI models because the Base Rate for human extinction is extremely low. In the early 2000s, some feared that the Large Hadron Collider (LHC), a machine that accelerates and collides tiny particles for scientific research, could create miniature Black Holes: areas of gravity from which nothing can escape. So, some feared that a small Black Hole created by LHC could destroy the Earth. These fears were based on speculative predictions without historical precedent, and no such black holes appeared. Despite capturing the public imagination, the LHC incident demonstrated the importance of a Reference Class and a Base Rate when estimating the probability of a catastrophic event.
In other inductive estimation efforts, statisticians compare advanced AI models to nuclear weapon development, as both present existential risks. However, these comparisons break down when considering the decentralized nature of AI-powered ecosystems. During the Cold War, nuclear technology was controlled by two state actors: the United States (U.S.) and the Soviet Union (U.S.S.R.). In contrast, cutting-edge AI research occurs globally, often in open-source environments. Thus, many AI firms and governments control critical AI resources, making predictability more difficult.
šSHORTCOMINGS OF DEDUCTIVE ESTIMATION METHODS
Deductive estimation methods involve deriving probabilities from logical reasoning based on existing knowledge. When it comes to existential risks from advanced AI models, this approach requires detailing how AI developments will unfold, their interactions with society, and their unintended consequences. Creating such comprehensive guesses is nearly impossible due to the unpredictability of technical advances and societal responses.
Consider Teslaās Full Self-Driving (FSD) program. Despite extensive frameworks of traffic rules, urban environments, and driver behavior, the program faces frequent challenges from Edge Cases: unexpected situations not covered by training data, such as unusual road markings, debris on highways, or emergency vehicles parked irregularly. These novel driving conditions have led to FSD malfunctions and crashes, highlighting the difficulty in anticipating every possible scenario in real-world driving environments. Similarly, deductive estimation methods for predicting existential risk from advanced AI models are vulnerable to such unpredictable variables. Thus, theyāre inherently incomplete forecasts of what might happen.
š„½LIMITATIONS OF SUBJECTIVE ESTIMATION METHODS
Subjective estimation methods lean on personal judgment to assign probabilities to outcomes. However, these forecasts usually reflect the biases, assumptions, and perspectives of the person making them.
For example, AI researchers sometimes forecast high probabilities for scenarios where an unaligned AI system optimizes for harmful goals, such as maximizing resource consumption at the expense of human life. These forecasts depend heavily on unscientific assumptions, such as the inevitability of a single AGI system gaining uncontested dominance. Critics argue these predictions often overestimate the probability of such outcomes as they donāt fully consider human intervention or regulatory constraints.
Without empirical data to validate such claims, subjective estimation methods become more reflective of personal beliefs than objective truths. Policymakers relying on these forecasts may inadvertently prioritize fringe scenarios over more pressing concerns like GenAI misuse through disinformation campaigns on social media platforms.
šKEY TAKEAWAY
Assessing existential risks from advanced AI models is filled with challenges that undermine the reliability of estimation methods.
Inductive estimation methods fail due to a lack of relevant historical data, such as trying to predict the risk of a black hole consuming Earth. Deductive estimation methods struggle to account for Edge Cases, like in Teslaās FSD program. Subjective estimation methods are often shaped by personal biases rather than objective realities.
These limitations highlight the need for humility when assessing the probability of existential risks from advanced AI models. Policymakers must prioritize adaptability rather than fixating on uncertain estimation methods when making policy decisions.
šFINAL NOTE
FEEDBACK
How would you rate todayās email?It helps us improve the content for you! |
ā¤ļøTAIP Review of The Week
āYour AI Ethics section on death, memory, and immortality was absolutely mind-blowing!ā
REFER & EARN
šYour Friends Learn, You Earn!
You currently have 0 referrals, only 1 away from receiving āļøUltimate Prompt Engineering Guide.
Refer 3 friends to learn how to š·āāļøBuild Custom Versions of OpenAIās ChatGPT.
Copy and paste this link to friends: https://theaipulse.beehiiv.com/subscribe?ref=PLACEHOLDER
Reply