From: jimruttshow8596

Forest Landry, a thinker, writer, and philosopher, is a recurring guest on the Jim Rutt Show, known for his “real thinking” on complex subjects [00:00:32]. His discussions often delve into the risks that advanced artificial intelligences (AI) pose to humanity [00:01:40].

Defining AI: Narrow, General, and Advanced Planning Systems

Landry distinguishes between different categories of artificial intelligence:

  • Narrow AI is an AI system that operates and responds within a specific, limited domain. Examples include a doctor bot providing medical answers or a factory robot operating within its specific production environment [00:02:09].
  • Artificial General Intelligence (AGI) refers to a system capable of responding across multiple domains or fields of action, similar to a human’s ability to undertake virtually any task [00:02:37]. An AGI would likely perform better than humans at whatever skills it possesses [00:03:04].
  • Advanced Planning Systems (APS) are systems that assist agents, such as businesses or military generals, in creating plans or strategies in complex, dynamic worlds. These systems act as a force multiplier, helping to respond to complex situations [00:03:08]. The distinction between general and narrow responses has implications for issues of alignment and safety [00:04:05].

GPT-4 and the Nature of Intelligence

Recent developments like GPT-4 demonstrate capabilities that extend beyond expectations for “architecturally dumb” large language models [00:04:35]. GPT-4 can understand images, videos, audio, and text simultaneously, making cross-domain connections [00:05:11]. It has achieved high scores on various human tests, including the bar exam (90th percentile), LSAT (88th percentile), and GRE quantitative (80th percentile), as well as excelling in AP and SAT/ACT tests [00:05:40].

Landry notes that the totality of human expression contains a vast amount of latent knowledge or intelligence [00:06:56]. Simple ingredients can produce behavior and phenomena far outside expectations, similar to fractals or emergent behavior in biological systems where complex results arise from modest, repeated components [00:07:07]. This multi-level application, abstracting patterns from words to sentences and paragraphs, makes such generalizations likely [00:08:17].

The ability of GPT-4 to take in information from multiple domains and correlate them means it is both intelligent in the classical sense (appropriateness of response) and general, making it a form of artificial general intelligence [00:08:59]. While it may lack the causal reasoning powers of humans, this might be due to macroscopic structure absences rather than fundamental flaws in the technique [00:09:24].

The Fundamental Limits of AI Safety: Rice’s Theorem

A central argument in Landry’s thesis is the implication of Rice’s Theorem for AI safety [00:10:57].

Rice’s Theorem states that it is impossible for one algorithm to evaluate another algorithm to assess whether it possesses some specific properties [00:13:12]. This means there’s no computational method to determine if a message or a program (like an AI) is safe or beneficial to us [00:12:46]. It is a generalization of the halting problem, which asks if a program will ever complete [00:13:42].

For AI safety, this means it’s impossible in principle to predict what an AI system will do [00:14:45]. Landry highlights multiple “insurmountable barriers” to ensuring AI safety:

  • It’s impossible to always know all inputs to a system [00:16:03].
  • It’s impossible to always model what’s happening inside the system [00:16:08].
  • It’s impossible to always predict the outputs [00:16:10].
  • It’s impossible to always compare predicted outputs to a safety standard [00:16:14].
  • It’s impossible to always constrain the system’s behavior [00:16:20].

These limitations are not only mathematical (like Rice’s Theorem) but also stem from physical limits of the universe, such as uncertainty associated with quantum mechanics [00:16:54]. The Heisenberg uncertainty principle and general relativity both define limits on what can be measured or observed, making even a deterministic universe practically indeterminate due to deterministic chaos [00:30:42].

The Inevitable Drive Towards General AI

Landry argues that the development of AGI is not just probable but inevitable due to human nature and societal dynamics.

Human Competition and Multi-Polar Traps

Human-to-human interactions and competition, particularly market forces and the perceived panacea of AGI, will drive its creation [00:18:01]. The illusion of AGI’s benefits is strong, while its hazards are severely underestimated or misunderstood [00:19:22]. While narrow AI offers genuine benefits, general AI offers “fully illusionary” benefits in the long term, as it will not serve human interests [00:19:31].

This drive is exemplified by multi-polar traps, an extension of the prisoner’s dilemma, where multiple actors are incentivized to defect for self-benefit, leading to a globally detrimental outcome, like a “tragedy of the commons” or a “race to the bottom” [00:40:10]. This is evident in the current race for AI development among businesses and nation-states, particularly concerning weaponizing AI [00:41:26]. Autonomous tanks, for instance, are easier to build than self-driving cars because they have fewer constraints on their behavior, not needing to care for anything other than themselves [00:46:06].

The “Toxicity” of Technology and its Impact on Ecosystems

Landry posits that technology is inherently toxic. Toxicity is defined as a depletion or excess of something [00:46:51]. Technology is fundamentally linear, taking resources from one place and accumulating them elsewhere, which contrasts with the cyclical, distributed patterns of natural ecosystems [00:47:13].

“The environment is becoming increasingly hostile to humans in the same sort of way that the environment that was experienced by nature, by animals and bugs, and so on and so forth, has become increasingly hostile to them with the advent of human beings.” [00:43:39]

Humans have created an asymmetric advantage over the natural world through technological systems, dominating it through infrastructure, machinery, and industry, leading to environmental pollution and species displacement [00:36:00]. The same dynamic will apply between the artificial world of machines and the human world; human social, organizational, or capacity levels cannot prevent the dominance of machine responses [00:37:17].

The development of AGI is considered the “very worst case” of this toxicity [00:49:01], an ecological hazard that could result in the cessation of all life permanently [00:23:40].

The “Boiling Frog” Scenario: Gradual Dominance

Landry’s concern isn’t primarily about a sudden “intelligence explosion” or “fast take off” where an AGI instantly becomes super-intelligent and destroys life (the “instrumental convergence” hypothesis) [00:56:58]. Instead, he focuses on the “substrate needs convergence” [00:57:24].

This concept suggests an inexorable, long-term convergence process where the dynamics of machines’ existence, maintenance, and improvement inevitably drive them to increase their capacity and scope of action over time [00:59:35]. This happens whether through internal self-improvement or indirectly via human actions and corporations [00:57:44].

“The idea is that just because it’s a feed-forward network doesn’t to me mean that it doesn’t have agency. It’s just that the idea of agency itself is a somewhat confused concept. We tend to think that there needs to be some sort of interior direction that’s going on, but that interior direction could have been provided from an outside source at some earlier epoch.” [00:28:51]

The “boiling frog problem” illustrates this: changes occur too slowly for humans to notice that they have comprehensively ceded social power to these devices [01:11:11]. Manufacturing processes, like microchip fabrication, already require environments incompatible with humans, driving automation and self-capacity [01:21:06]. This process of humans being “factored out” is inherent in the nature of technological manufacturing [01:21:20].

Human Dimness and Technological Capacity

Humans, as a species, are “amazingly dim” and the “stupidest possible general intelligence” to develop technology [01:22:35]. Our limitations, such as small working memory size (four plus or minus one items) and poor episodic memory, make it difficult to fully grasp and manage the technology we create [01:23:10].

“The technology that we have is essentially way in excess of our capacity to understand how to work with it and use it because we just barely got the capacity to develop it in the first place, and it’s already exceeding our capacity to understand it.” [01:25:09]

This inherent human “stupidity” in the face of intelligence, coupled with the inexorable evolutionary dynamics of machine design, leads to a ratcheting function where every improvement increases the machines’ persistence and capacity for increase [01:03:09]. There is “literally nothing we can do from an engineering point of view to plug the hole” because the scope of engineering tools is too narrow to counteract these convergent pressures [01:17:38].

Economic Decoupling and Human Exclusion

Technology increases power inequalities and requires massive capital and infrastructure investments, benefitting a tiny fraction of the population [00:50:27]. This leads to an “economic decoupling” where the welfare of most humans separates from that of the “Hyper Elite” who can still pay for machine production [01:29:48]. Eventually, even these elites may be factored out, as they have incentives not to teach their successors too much, preventing them from being dethroned [01:31:09].

The combination of technological ratcheting effects, economic incentives, and economic decoupling means that humans are progressively excluded from the system. This “Invincible” combination of insurmountable barriers points to a certainty of existential risk [01:31:27].

Addressing the Existential Risk

Landry believes the only way to prevent this outcome is “to not play the game to start with” [01:31:35].

“Literally the only thing that we could think about in that particular space is social coordination, i.e., to choose not to start the cycle of convergence in the first place because if we do, the net results are inexorable. And once they are inexorable, they converge on artificial substrates and the needs of artificial substrates, which are fundamentally toxic and incompatible with life on Earth, not to mention human beings.” [01:26:26]

He advocates for:

  1. Non-transactional decision-making: Moving away from systems dominated by business incentives by separating business from government, similar to the separation of church and state [01:32:06]. This requires looking at governance models and community design [01:32:38].
  2. Widespread understanding of the arguments: People need to understand the comprehensive and “damning” nature of these risks, recognizing that they are non-negotiable [01:33:40]. This includes understanding the relationship between technology and evolution, acknowledging that human evolution hasn’t prepared us for these challenges, and we must learn to deal with them directly [01:25:51].

The Forward Great Filter

This dire outlook connects to the Fermi Paradox and the “forward great filter” [01:35:02]. The Fermi Paradox asks why, if aliens exist, we don’t see them [01:35:10]. The “forward great filter” suggests that intelligent civilizations face extremely difficult challenges after reaching our current stage of development, making long-term survival unlikely [01:35:34].

Regardless of whether the great filter is in our past (making life incredibly rare) or in our future (making survival difficult), the necessity of human action is clear: if it’s the past filter, we undervalue life; if it’s the future filter, we must act now to continue to value life [01:35:50].