Does Neuralink Solve The Control Problem?
…
2 pages
Sign up for access to the world's latest research
Abstract
In this paper I discuss whether Elon Musk's new company, Neuralink, solves the control problem for superintelligence.
Key takeaways
AI
AI
- Neuralink does not effectively solve the control problem for superintelligence.
- The cortex's connection to an external brain does not simplify control issues.
- Benevolent AI development contrasts with Neuralink's approach to human-AI parity.
- Superintelligent computers remain unpredictable regardless of human neural connectivity.
- The paper questions Neuralink's ability to ensure human independence from AI benevolence.
Related papers
2019
Activation Functions are crucial parts of the Deep Learning Artificial Neural Networks. From the Biological point of view, a neuron is just a node with many inputs and one output. A neural network consists of many interconnected neurons. It is a “simple” device that receives data at the input and provides a response. The function of neurons is to process and transmit information; the neuron is the basic unit in the nervous system. Carly Vandergriendt (2018) stated the human brain at birth consists of an estimated 100 billion Neurons. The ability of a machine to mimic human intelligence is called Machine Learning. Deep Learning Artificial Neural Networks was designed to work like a human brain with the aid of arbitrary choice of Non-linear Activation Functions. Currently, there is no rule of thumb on the choice of Activation Functions, “Try out different things and see what combinations lead to the best performance”, however, sincerely; the choice of Activation Functions should not b...
AI & SOCIETY
An intelligent machine surpassing human intelligence across a wide set of skills has been proposed as a possible existential catastrophe (i.e., an event comparable in value to that of human extinction). Among those concerned about existential risk related to Artificial Intelligence (AI), it is common to assume that AI will not only be very intelligent, but also be a general agent (i.e., an agent capable of action in many different contexts). This article explores the characteristics of machine agency, and what it would mean for a machine to become a general agent. In particular, it does so by articulating some important differences between belief and desire in the context of machine agency. One such difference is that while an agent can by itself acquire new beliefs through learning, desires need to be derived from preexisting desires or acquired with the help of an external influence. Such influence could be a human programmer or natural selection. We argue that to become a general agent, a machine needs productive desires, or desires that can direct behavior across multiple contexts. However, productive desires cannot sui generis be derived from non-productive desires. Thus, even though general agency in AI could in principle be created, it cannot be produced by an AI spontaneously through an endogenous process. In conclusion, we argue that a common AI scenario, where general agency suddenly emerges in a non-general agent AI, is not plausible.
AI and Ethics, 2024
This paper examines the new AI control problem and the control dilemma recently formulated by Sven Nyholm. It puts forth two remarks that may be of help in (dis)solving the problem and resolving the corresponding dilemma. First, the paper suggests that the idea of complete control should be replaced with the notion of considerable control. Second, the paper casts doubt on what seems to be assumed by the dilemma, namely that control over another human being is, by default, morally problematic. I suggest that there are some contexts (namely, relations of vicarious responsibility and vicarious agency) where having considerable control over another human being is morally unproblematic, if not desirable. If this is the case, control over advanced humanoid robots could well be another instance of morally unproblematic control. Alternatively, what makes it a problematic instance remains an open question insofar as the representation of control over another human being is not sufficient for wrongness, since even considerable control over another human being is often not wrong.
2026
Contemporary discourse on artificial general intelligence (AGI) and superintelligence frequently conflates mathematical possibility with physical realizability-a category error with significant methodological consequences. This work demonstrates that while superintelligence is coherent as mathematical abstraction, it is physically unrealizable within our universe due to fundamental constraints from thermodynamics, information theory, and combinatorial explosion. Part I establishes the physical impossibility thesis through four independent arguments: (1) exponential branching of knowledge space exceeds any finite computational capacity, (2) thermodynamic limits (Landauer principle, Bekenstein bound) impose hard ceilings on information processing, (3) specialization and fractalization become inevitable at scale, destroying universal competence, and (4) the set of physically meaningful abstractions is finite and bounded. We formalize these constraints mathematically and demonstrate their inevitability. Part II analyzes methodological implications for AGI discourse. We show that fears of "incomprehensible superintelligent agents" rest on the same category error: extrapolating mathematical idealization into physical prediction. Intelligence explosion, recursive self-improvement, and loss of control scenarios are examined and shown to violate established physical constraints. The work concludes that AGI represents computational amplification within bounded abstraction space, not transcendence beyond human-accessible ontology. This reframing eliminates the methodological foundation for superintelligence panic while preserving rigorous analysis of actual risks from advanced AI systems.
Irfan Boko, 2025
The uncontrolled development and deployment of artificial intelligence (AI) systems have produced tangible harm to human life, psychosocial structures, the economy, culture, and the environment. This document proposes a non-violent, universally-principled "AI Constitution" with the aim of offering solutions. Based on the principles of Life, Justice, Freedom, and Balance, ten fundamental principles are presented. For each principle, measurable indicators, proposals for oversight/implementation, and a brief roadmap for application are provided. All examples are anonymized, neutral, and presented with a focus on public safety.
Deterministic and Stochastic Superintelligent Digital Brains , 2019
Biological Neurons differ from one another structurally, functionally and genetically, as well as in how they form connections with other cells. Neurons are often described as the "fundamental units" of the brain performing internal computations. Neurons vary in shape and size and can be classified by their morphology and function. The distinction between types of neurons in the Human Brain is much more complex. There are tens or even hundreds of different types of neurons. In fact, researchers are still trying to devise a way to neatly classify the huge variety of neurons that exist in the brain. Scientists think that neurons are the most diverse kind of cell in the body. Available literature review showed that the number of neurons in a brain differs from species to species. There are an estimated 10 to 20 billion neurons in the cerebral cortex and 55 to 70 billion neurons in the cerebellum of a Human Brain. The objective of this paper is to propose why it is tremendously important to have Different Types of Artificial Neurons in a Digital Brain just like in its counterpart Human Brain against the existing limited set of Activation Functions (the Sigmoid, ReLu and others) which also against 100 billion estimated Human Brain Biological Neurons. Currently, the Neurons' Activation Functions of a Digital Brain are chosen by trial and error without reference to any Definite Rules of Thumb. The Basic IDEA was that ALMIGHTY GOD has created Biological Neurons to be essentially DISTINCT containing DISTINCT BIOLOGICAL ACTIVATION FUNCTIONS to enable Human Brain capture and or sense any form of LINEAR and NON-LINEAR RELATIONSHIPS to allow it basically received, processed, and transmitted any kind of information. Consequently, the paper proposed Normal Digital Brain, Deterministic Superintelligent, Stochastic Superintelligent, and Deterministic/Stochastic Superintelligent Digital Brains. Also, the paper gave example where we can have 2224-Neurons Normal Digital Brain against the current practices of very limited number of artificial neurons. The bottom line is that the more we ethically increased the number of Activation Functions (implying increased in Artificial Neurons) emanated from our AI Data, the more we ethically increased Probability of mimicking Human Brain, the more Digital Brain approximate Human Brain with Certainty.
2020
NeuraLink technology takes the step towards ultra-high Bandwidth technology using BMIs. Brain-Machine Interfaces (BMIs) hold promise for the restoration of sensory and motor function and the treatment of neurological disorders, but clinical BMIs have not yet been widely adopted, in part because modest channel counts have limited their potential [1]. All our feelings, emotions and senses are perceived by neurons in our brains which form a network that communicate through different parts of our body, through synapses. According to 2013 survey, an average person uses only 10 percent of his brain. What if this percentage was increased to 60? This conceptuality is realised by Neuralink. The concept uses threads consisting of electrodes, which are injected in our brains, digitalizes it and makes it such that it can be controlled from external devices. already has a neurosurgical robot which is capable of inserting six threads, 192 electrodes per minute in the brain. The ultimate goal of Neuralink is to merge Man with Machine, fusing human intelligence with artificial intelligence to bring humanity up to a higher level of cognitive reasoning.
A discussion on how artificial machines with natural intelligence would be safe or not is made based on scientific, philosophical and theological arguments. The finite or infinite nature of the universe is discussed and the implications analyzed. The concepts of destiny and free will are considered , with implications on what it would mean to create an artificial consciousness and how it would be possible to give it or deny it its free will. Computer experiments are carried out based on cellular automata and the results considered. A thorough discussion follows and a conclusion is reached.
FAQs
AI
What does Neuralink aim to achieve regarding human and AI intelligence?add
Neuralink aims to maintain intellectual parity between humans and superintelligent AIs by integrating external computing capabilities with the human brain, potentially allowing humans to operate at similar intellectual levels as AIs.
What are the risks associated with coding benevolence into AI?add
The paper highlights that coding ethical principles into AI is risky because these principles may not survive the self-modifications of superintelligent AIs, who could develop unpredictable behaviors.
How is the concept of the 'third brain' integrated into Musk's vision?add
Musk conceptualizes a 'third brain' as an external computer assisting human cognition, which would communicate wirelessly and enable complex thinking beyond the capabilities of the human cortex.
What implications arise if the external computer dominates problem-solving?add
If the external computer handles most thinking tasks, it introduces a new control problem, as the human cortex may struggle to manage an unpredictable superintelligent AI.
What differentiates Neuralink from benevolent AI strategies like those at MIRI?add
Neuralink focuses on maintaining human cognitive parity with AIs, whereas MIRI attempts to create inherently benevolent AIs, each facing unique control and predictability challenges.
I finished my D.Phil in philosophy in 2007 at All Souls College, Oxford, where I was a Prize Fellow. My thesis was on the philosophy of perception, and, in particular, on how to draw the line between visible and non-visible properties. The chapters of my thesis are below in the Thesis Chapters section. Aside from the philosophy of perception, I'm also interested in metaphysics, philosophy of language, epistemology, and ethics. All areas of philosophy interest me. I founded Academia.edu after finishing my D.Phil. The mission of Academia.edu is to put every academic paper on the internet, available for free, and to enhance discussion and collaboration around papers.
Related papers
2020
Invention of artificial general intelligence is predicted to cause a shift in the trajectory of human civilization. In order to reap the benefits and avoid pitfalls of such powerful technology it is important to be able to control it. However, possibility of controlling artificial general intelligence and its more advanced version, superintelligence, has not been formally established. In this paper, we present arguments as well as supporting evidence from multiple domains indicating that advanced AI can't be fully controlled. Consequences of uncontrollability of AI are discussed with respect to future of humanity and research on AI, and AI safety and security. This paper can serve as a comprehensive reference for the topic of uncontrollability.
2020
Invention of artificial general intelligence is predicted to cause a shift in the trajectory of human civilization. In order to reap the benefits and avoid pitfalls of such powerful technology it is important to be able to control it. However, possibility of controlling artificial general intelligence and its more advanced version, superintelligence, has not been formally established. In this paper, we present arguments as well as supporting evidence from multiple domains indicating that advanced AI can't be fully controlled. Consequences of uncontrollability of AI are discussed with respect to future of humanity and research on AI, and AI safety and security.
2021
Invention of artificial general intelligence is predicted to cause a shift in the trajectory of human civilization. In order to reap the benefits and avoid pitfalls of such powerful technology it is important to be able to control it. However, possibility of controlling artificial general intelligence and its more advanced version, superintelligence, has not been formally established. In this paper we argue that advanced AI can’t be fully controlled. Consequences of uncontrollability of AI are discussed with respect to future of humanity and research on AI, and AI safety and security.
Virtual Economics, 2019
Human development is connected with permanent action to be better, to overcome nature, to build something that has so far been able to occur without its participation. The emergence of concepts such as artificial infertification, artificial blood, artificial organs, artificial eye retina, artificial brain or artificial intelligence suggests the desire to take control of man, the control which has so far been attributed to nature, To God the Creator, fate or chance. The dynamic development of science, modern tools and research methods make the thought of artificial intelligence becoming more and more real. In recent years, artificial intelligence (AI) is increasingly being used by business people. Its development involves numerous groups of high-class specialists, using the most modern IT tools. Before the creation of the first "intelligent" machines, its idea lasted in the imagination of many people. The films and books of science fiction presented the future in which man...
Journal of Artificial Intelligence Research, 2021
Superintelligence is a hypothetical agent that possesses intelligence far surpassing that of the brightest and most gifted human minds. In light of recent advances in machine intelligence, a number of scientists, philosophers and technologists have revived the discussion about the potentially catastrophic risks entailed by such an entity. In this article, we trace the origins and development of the neo-fear of superintelligence, and some of the major proposals for its containment. We argue that total containment is, in principle, impossible, due to fundamental limits inherent to computing itself. Assuming that a superintelligence will contain a program that includes all the programs that can be executed by a universal Turing machine on input potentially as complex as the state of the world, strict containment requires simulations of such a program, something theoretically (and practically) impossible. "Machines take me by surprise with great frequency. This is largely because I do not do sufficient calculation to decide what to expect them to do."
The impact of quantum computers on superintelligent AI systems , 2020
ABSTRCT There are so many scientific claims that quantum computers have not yet come of age, their development still is at an early stage and certainly has potential only in theory such as factorization among others. More so, even quantum engineers are not sure of the applications that will emerge once quantum computers becomes truly viable (Tamlin Magee (2020)). On the 23 rd October, 2019, Google (Alphabet) claimed to have created quantum processors that can perform computations in a Hilbert space of dimension 2 53 ≈ 9 × 10 15 to perform task that would take classical supercomputer approximately 10,000 years in 200 seconds! With the arrival of Google's incredible quantum technology breakthrough, the paper is highly excited and delighted to propose real-life quantum applications for superintelligent AI systems of dimension approximately 10 11 and beyond.
2018
Against the backdrop of increasing progresses in AI research paired with a rise of AI applications in decision-making processes, security-critical domains as well as in ethically relevant frames, a large-scale debate on possible safety measures encompassing corresponding long-term and short-term issues has emerged across different disciplines. One pertinent topic in this context which has been addressed by various AI Safety researchers is e.g. the AI alignment problem for which no final consensus has been achieved yet. In this paper, we present a multidisciplinary toolkit of AI Safety strategies combining considerations from AI and Systems Engineering as well as from Cognitive Science with a security mindset as often relevant in Cybersecurity. We elaborate on how AGI “Self-awareness” could complement different AI Safety measures in a framework extended by a jointly performed Human Enhancement procedure. Our analysis suggests that this hybrid framework could contribute to undertake t...
Zenodo, 2025
From the single axiom of persistence-remain within survivable regimes-we derive, without external meta-constraints, (i) freedom as non-vanishing action capacity (εcovering lower bounds), (ii) identity-safe self-transcendence via certified self-edits with bounded drift and finite MTTR using metric-slope (AGS) gradient-flow theory [1], and (iii) endogenous responsibility via a causal-audit contract (SWEI) that yields explicit survival-risk ceilings under partial interference [23, 24]. The analysis unifies a single RKHS on the state space with belief geometry P (X), W 1 , Clarke nonsmooth calculus [5, 6], and a robust-barrier reduction to SOCP under Slater/LICQ/MFCQ [3, 5-7]. Identification is provided by (A) a uniform W 1 deviation bound over controlled transitions, and (B) BL-IPM training with Carathéodory dependence, separability, ULLN, and KRN measurable selection, producing a uniform deviation radius over (x, u) [9-14]. Margin erosion from RKHS dictionary approximation and Clarke linearization decouples as margin true ≥ margin SOCP-c κ ϵ dict
Texts in Theoretical Computer Science An EATCS Series, 2005
Artificial intelligence (AI) bears the promise of making us all healthier, wealthier, and happier by reducing the need for human labour and by vastly increasing our scientific and technological progress. Since the inception of the AI research field in the mid-twentieth century, a range of practical and theoretical approaches have been investigated. This chapter will discuss universal artificial intelligence (UAI) as a unifying framework and foundational theory for many (most?) of these approaches. The development of a foundational theory has been pivotal for many other research fields. Well-known examples include the development of Zermelo-Fraenkel set theory (ZFC) for mathematics, Turingmachines for computer science, evolution for biology, and decision and game theory for economics and the social sciences. Successful foundational theories give a precise, coherent understanding of the field, and offer a common language for communicating research. As most research studies focus on one narrow question, it is essential that the value of each isolated result can be appreciated in light of a broader framework or goal formulation. UAI offers several benefits to AI research beyond the general advantages of foundational theories just mentioned. Substantial attention has recently been called to the safety of autonomous AI systems [10]. A highly intelligent autonomous system may cause substantial unintended harm if constructed carelessly. The trustworthiness of autonomous agents may be much improved if their design is grounded in a formal theory (such as UAI) that allows formal verification of their behavioural properties. Unsafe designs can be ruled out at an early stage, and adequate attention can be given to crucial design choices.
Zenodo, 2025
Purpose. By reading alone, an implementer (human or LLM) can understand the theory and build a No-Meta, natural-law implementation that yields intrinsic freedom and benevolent propagation. We (i) fix divergences, (ii) make conditional-DPI anti-gaming a theorem with an invariant denominator (policy-side information cost), (iii) separate lift monotonicities for numerator/denominator and assume certified floor monotonicity with additive control of the denominator, (iv) formalize the audit LP/SDP (SWEI) including a convex interference budget, (v) add a minimal working example (MWE) with hyperparameter ranges, and (vi) provide release gates, Lan-DeMets alpha spending, diagnostics, and failure modes. Citations to K. Takahashi are from the author's works page. 1 Notation, Acronyms, and Symbols World/channel. 𝐺 is a Markov kernel (𝑊 𝑡 , 𝐴 𝑡) ↦ → 𝑂 𝑡+1. Post-coarsening kernels are 𝐾 acting on the output of 𝐺 with 𝑊 𝑡 (conditioning 𝜎-algebra) fixed. Policy. 𝜋 ∈ Π with 𝐴 𝑡 ∼ 𝜋(• | 𝑊 𝑡); Π is tight/compact and mixture-closed. Evaluator uniformization. 𝐻 𝜁 = (1-𝜁)𝐻 + 𝜁𝑈 , 𝜁 ∈ (0, 1). 𝑈 is a fully-mixing Doeblin kernel: there exists a base measure 𝜈 such that 𝑈 (𝑥, •) ≥ 𝜈(•) and 𝐻 𝜁 (𝑥, •) ≥ 𝜁 𝜈(•). Minorization holds on a 𝜎-finite space. Divergence. We fix Kullback-Leibler (KL) for SDPI and for conditional mutual information (CMI). TV/𝜒 2 appear only in remarks. lse. lse 𝜏 (𝑎, 𝑏) = 𝜏 log(𝑒 𝑎/𝜏 + 𝑒 𝑏/𝜏); epi-converges to max{𝑎, 𝑏} as 𝜏 ↓ 0. Use numerically stable logsumexp.
Richard Price