Jump to content
Main menu
Main menu
move to sidebar
hide
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Search
Search
Create account
Log in
Personal tools
Create account
Log in
Pages for logged out editors
learn more
Contributions
Talk
Editing
Mind uploading
Page
Discussion
English
Read
Edit
Edit source
View history
Tools
Tools
move to sidebar
hide
Actions
Read
Edit
Edit source
View history
Move
General
What links here
Related changes
Special pages
Page information
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
{{short description|Hypothetical process of digitally emulating a brain}} {{redirect|Mind transfer}} {{cyborg}} '''Mind uploading''' is a speculative process of '''whole brain emulation''' in which a [[brain scanning|brain scan]] is used to completely [[emulator|emulate]] the mental state of the individual in a [[Digital data|digital]] [[computer]]. The computer would then run a [[computer simulation|simulation]] of the brain's information processing, such that it would respond in essentially the same way as the original brain and experience having a [[sentient]] [[conscious]] [[mind]].<ref name="sim.me.uk">{{cite journal | last=Bamford | first=Sim | title=A framework for approaches to transfer of a mind's substrate | journal=International Journal of Machine Consciousness | volume=04 | issue=1 | year=2012 | issn=1793-8430 | doi=10.1142/s1793843012400021 | pages=23–34 |url=http://www.sim.me.uk/neural/JournalArticles/Bamford2012IJMC.pdf}}</ref><ref>{{Cite journal |last1=Goertzel |first1=Ben |last2=Ikle' |first2=Matthew |year=2012 |title=Introduction |journal=International Journal of Machine Consciousness |volume=04 |pages=1–3 |doi=10.1142/S1793843012020015}}</ref><ref name="kajsotala.fi">{{cite journal |last1=Sotala |first1=Kaj |last2=Valpola |first2=Harri |url=http://kajsotala.fi/Papers/CoalescingMinds.pdf |title=Coalescing minds: brain uploading-related group mind scenarios |journal=International Journal of Machine Consciousness |date=June 2012 |volume=04 |issue=1 |pages=293–312 |doi=10.1142/S1793843012400173 |s2cid=6230653}}</ref> Substantial mainstream research in related areas is being conducted in [[neuroscience]] and [[computer science]], including animal brain mapping and simulation,<ref name="ReferenceA">{{Cite journal |last1=Fan |first1=Xue |last2=Markram |first2=Henry |date=2019-05-07 |title=A Brief History of Simulation Neuroscience |journal=Frontiers in Neuroinformatics |volume=13 |page=32 |doi=10.3389/fninf.2019.00032 |doi-access=free |issn=1662-5196 |pmc=6513977 |pmid=31133838}}</ref> development of faster supercomputers, [[virtual reality]], [[brain–computer interface]]s, [[connectomics]], and information extraction from dynamically functioning brains.<ref>{{cite journal|vauthors=Kay KN, Naselaris T, Prenger RJ, Gallant JL|date=March 2008|title=Identifying natural images from human brain activity|journal=Nature|volume=452|issue=7185|pages=352–5|bibcode=2008Natur.452..352K|doi=10.1038/nature06713|pmc=3556484|pmid=18322462}}</ref> According to supporters, many of the tools and ideas needed to achieve mind uploading already exist or are under active development; however, they will admit that others are, as yet, very speculative, but say they are still in the realm of engineering possibility. Mind uploading may potentially be accomplished by either of two methods: copy-and-upload or copy-and-delete by gradual replacement of [[neurons]] (which can be considered as a gradual destructive uploading), until the original [[Organic matter|organic]] brain no longer exists and a [[computer program]] emulating the brain takes control of the body. In the case of the former method, mind uploading would be achieved by scanning and [[Brain mapping|mapping]] the salient features of a biological brain, and then by storing and copying that information state into a computer system or another computational device. The [[biological]] brain may not survive the copying process or may be deliberately destroyed during it in some variants of uploading. The simulated mind could be within a virtual reality or [[simulated reality|simulated world]], supported by an anatomic 3D body simulation model. Alternatively, the simulated mind could reside in a computer inside—or either connected to or remotely controlled by—a (not necessarily [[humanoid]]) [[robot]], biological, or [[cyborg|cybernetic]] body.<ref name=Roadmap> {{Cite book |first1=Anders |last1=Sandberg | author1-link=Anders Sandberg |first2=Nick |last2=Boström | author2-link=Nick Bostrom |title=Whole Brain Emulation: A Roadmap |quote=The basic idea is to take a particular brain, scan its structure in detail, and construct a software model of it that is so faithful to the original that, when run on appropriate hardware, it will behave in essentially the same way as the original brain. |url=http://www.fhi.ox.ac.uk/Reports/2008-3.pdf |access-date=5 April 2009 |series= Technical Report #2008-3 |year=2008 | publisher = Future of Humanity Institute, Oxford University }}</ref> Among some [[futurist]]s and within part of [[Transhumanism|transhumanist]] movement, mind uploading is treated as an important proposed [[life extension]] or [[immortality]] technology (known as "[[digital immortality]]"). Some believe mind uploading is humanity's current best option for preserving the identity of the species, as opposed to [[cryonics]]. Another aim of mind uploading is to provide a permanent backup to our "mind-file", to enable interstellar space travel, and a means for human culture to survive a global disaster by making a functional copy of a human society in a computing device. Whole-brain emulation is discussed by some futurists as a "logical endpoint"<ref name=Roadmap/> of the topical [[computational neuroscience]] and [[neuroinformatics]] fields, both about [[brain simulation]] for medical research purposes. It is discussed in [[artificial intelligence]] research publications as an approach to [[computational theory of mind|strong AI]] ([[artificial general intelligence]]) and to at least weak [[superintelligence]]. Another approach is [[seed AI]], which would not be based on existing brains. Computer-based intelligence such as an upload could think much faster than a biological human even if it were no more intelligent. A large-scale society of uploads might, according to futurists, give rise to a [[technological singularity]], meaning a sudden time constant decrease in the exponential development of technology.<ref>{{cite journal |last=Goertzel|first=Ben| author-link=Ben Goertzel| date= December 2007|title=Human-level artificial general intelligence and the possibility of a technological singularity: a reaction to Ray Kurzweil's The Singularity Is Near, and McDermott's critique of Kurzweil |journal= Artificial Intelligence| volume=171|issue=18, Special Review Issue|pages=1161–1173| doi= 10.1016/j.artint.2007.10.011| url= https://scholar.google.com/scholar?cluster=4599202374830957934|doi-access=free}}</ref> Mind uploading is a central conceptual feature of [[Mind uploading in fiction|numerous science fiction novels, films, and games]].<ref>{{Cite web |title=SFE: Upload |url=https://sf-encyclopedia.com/entry/upload |access-date=2024-03-24 |website=sf-encyclopedia.com}}</ref> ==Overview== Many neuroscientists believe that the human [[mind]] is largely an [[Emergence|emergent]] property of the information processing of its [[neuronal network]].<ref>{{Cite journal|last=Hopfield|first=J. J.|date=1982-04-01|title=Neural networks and physical systems with emergent collective computational abilities.|journal=Proceedings of the National Academy of Sciences| volume=79|issue=8|pages=2554–2558| doi= 10.1073/pnas.79.8.2554| pmid= 6953413 |pmc= 346238 |bibcode= 1982PNAS...79.2554H |issn=0027-8424 |doi-access=free}}</ref> Neuroscientists have stated that important functions performed by the mind, such as learning, memory, and consciousness, are due to purely physical and electrochemical processes in the brain and are governed by applicable laws. For example, [[Christof Koch]] and [[Giulio Tononi]] wrote in ''[[IEEE Spectrum]]'': {{Blockquote|Consciousness is part of the natural world. It depends, we believe, only on mathematics and logic and on the imperfectly known laws of physics, chemistry, and biology; it does not arise from some magical or otherworldly quality.<ref name="ieee_conscious">{{cite journal |doi= 10.1109/MSPEC.2008.4531463 |title=Can machines be conscious? |year=2008 |last1=Koch |first1=Christof |author-link1= Giulio Tononi |last2=Tononi |first2= Giulio |journal= IEEE Spectrum |volume=45 |pages=55 |issue=6|s2cid=7226896 |url=https://authors.library.caltech.edu/11693/1/KOCieeespec08.pdf }}</ref>}} Eminent [[computer scientists]] and neuroscientists have predicted that advanced computers will be capable of thought and even attain consciousness, including Koch and Tononi,<ref name="ieee_conscious"/> [[Douglas Hofstadter]],<ref name="spectrum.ieee.org">{{cite web |title=Tech Luminaries Address Singularity |url=http://www.spectrum.ieee.org/jun08/6277 |url-status=dead |archive-url=https://web.archive.org/web/20090501102356/http://www.spectrum.ieee.org/jun08/6277 |archive-date=2009-05-01 |access-date=2009-04-02 |work=[[IEEE]]}}</ref> [[Jeff Hawkins]],<ref name="spectrum.ieee.org"/> [[Marvin Minsky]],<ref>{{cite journal| first= Marvin | last= Minsky| author-link= Marvin Minsky| title= Conscious Machines| journal= Machinery of Consciousness| publisher= Proceedings, National Research Council of Canada, 75th Anniversary Symposium on Science in Society |date= June 1991}}</ref> Randal A. Koene, and [[Rodolfo Llinás]].<ref name="Llinas">{{cite book | last = Llinas | first = R |pages = 261–262 | title = I of the vortex: from neurons to self | year = 2001 | publisher = MIT Press | location = Cambridge | isbn = 978-0-262-62163-2 | title-link = I of the vortex: from neurons to self }}</ref> Many theorists have presented models of the brain and have established a range of estimates of the amount of computing power needed for partial and complete simulations.<ref name="ReferenceA"/><ref name=Roadmap/> Using these models, some have estimated that uploading may become possible within decades if trends such as [[Moore's law]] continue.<ref>{{cite web| first= Ray |last= Kurzweil|author-link=Ray Kurzweil|url=http://www.kurzweilai.net/live-forever-uploading-the-human-braincloser-than-you-think|title=Live Forever–Uploading The Human Brain...Closer Than You Think|work=Psychology Today |date= February 2000}}</ref> As of December 2022, this kind of technology is almost entirely theoretical. ==Theoretical benefits and applications== ==="Immortality" or backup=== {{Main|Digital immortality}} In theory, if the information and processes of the mind can be disassociated from the biological body, they are no longer tied to the individual limits and lifespan of that body. Furthermore, information within a brain could be partly or wholly copied or transferred to one or more other substrates (including digital storage or another brain), thereby—from a purely mechanistic perspective—reducing or eliminating "mortality risk" of such information. This general proposal was discussed in 1971 by [[biomedical gerontology|biogerontologist]] [[George M. Martin]] of the [[University of Washington]].<ref name="Martin 1971 339">{{cite journal |last= Martin |first= G. M. |title=Brief proposal on immortality: an interim solution |journal=Perspectives in Biology and Medicine |volume=14 |issue=2 |pages=339–340 |year=1971 |pmid= 5546258 |doi=10.1353/pbm.1971.0015|s2cid=71120068 }}</ref> This questions the concept of identity. From the perspective of the biological brain, the simulated brain may just be a copy, even if it is conscious and has an indistinguishable character. As such, the original biological being, before the uploading, might consider the digital twin to be a new and independent being rather than the future self.<ref>{{Cite web |last=Rosenberg |first=Louis |date=2022-08-13 |title=Identity crisis: Artificial intelligence and the flawed logic of 'mind uploading' |url=https://venturebeat.com/virtual/identity-crisis-artificial-intelligence-and-the-flawed-logic-of-mind-uploading/ |access-date=2024-05-10 |website=VentureBeat |language=en-US}}</ref> ===Space exploration=== An "uploaded astronaut" could be used instead of a "live" astronaut in [[human spaceflight]], avoiding the perils of [[zero gravity]], the [[vacuum]] of space, and [[cosmic radiation]] to the human body. It would allow for the use of smaller spacecraft, such as the proposed [[StarChip (spacecraft)|StarChip]], and it would enable virtually unlimited [[interstellar travel]] distances.<ref>{{cite web| last1= Prisco| first1= Giulio |title= Uploaded e-crews for interstellar missions |date= 12 December 2012 |url= http://www.kurzweilai.net/uploaded-e-crews-for-interstellar-missions |website= kurzweilai.net| access-date= 31 July 2015}}</ref> === Mind editing === While some researchers believe editing human brains to be physically possible in theory, for example by performing neurosurgery with [[nanobots]], it would require particularly advanced technology. Editing an uploaded mind would be much easier, as long as the exact edits to be made are known.<ref>{{Cite book |last=Bostrom |first=Nick |title=Deep Utopia: Life and Meaning in a Solved World |date=March 27, 2024 |isbn=978-1646871643 |chapter=Handout 10: downloading and brain editing|publisher=Ideapress }}</ref> This would facilitate [[cognitive enhancement]] and the precise control of the well-being, motivations or personality of the emulated beings.<ref>{{Cite book |last=Bostrom |first=Nick |title=Deep Utopia: Life and Meaning in a Solved World |date=March 27, 2024 |isbn=978-1646871643 |chapter=Handout 2: Some capabilities at technological maturity|publisher=Ideapress }}</ref> === Speed === Although the number of neuronal connections in the human brain is very significant (around 100 trillions<ref>{{Cite web |last=Caruso |first=Catherine |date=January 19, 2023 |title=A New Field of Neuroscience Aims to Map Connections in the Brain |url=https://hms.harvard.edu/news/new-field-neuroscience-aims-map-connections-brain |access-date=2024-05-10 |website=Harvard medical school}}</ref>), the frequency of activation of biological neurons is limited to around 200 Hz, whereas electronic hardware can easily operate at multiple GHz. With sufficient hardware parallelism, a simulated brain could thus in theory be made to run faster than a biological brain. Uploaded beings may therefore not only be more efficient, but also supposedly have a faster rate of [[subjective experience]] than biological brains (e.g. experiencing an hour of lifetime in a single second of real time).<ref>{{Cite book |last=Bostrom |first=Nick |title=Superintelligence: paths, dangers, strategies |date=2017 |publisher=Oxford University Press |isbn=978-0-19-967811-2 |edition= |location= |chapter=Speed superintelligence}}</ref> == Relevant technologies and techniques == The focus of mind uploading, in the case of copy-and-transfer, is on data acquisition, rather than data maintenance of the brain. A set of approaches known as loosely coupled off-loading (LCOL) may be used in the attempt to characterize and copy the mental contents of a brain.<ref name=SIM>{{cite web|url=http://www.carboncopies.org/substrate-independent-minds| title=Substrate-Independent Minds |publisher= Carboncopies.org Foundation |work=carboncopies.org|access-date=2014-01-03 |archive-url= https://web.archive.org/web/20140103135545/http://www.carboncopies.org/substrate-independent-minds |archive-date= 2014-01-03|url-status=dead}}</ref> The LCOL approach may take advantage of self-reports, life-logs and video recordings that can be analyzed by artificial intelligence. A bottom-up approach may focus on the specific resolution and morphology of neurons, the spike times of neurons, the times at which neurons produce action potential responses. ===Computational complexity=== [[File:Whole brain emulation.svg|thumb|upright=1.8|Estimates of how much processing power is needed to emulate a human brain at various levels, along with the fastest and slowest [[supercomputer]]s from [[TOP500]] and a $1000 PC. Note the logarithmic scale. The (exponential) trend line for the fastest supercomputer reflects a doubling every 14 months. Kurzweil believes that mind uploading will be possible at neural simulation, while the Sandberg & Bostrom report is less certain about where consciousness arises.<ref>Roadmap. "Given the complexities and conceptual issues of consciousness we will not examine criteria 6abc, but mainly examine achieving criteria 1–5." p. 11.</ref>]] Advocates of mind uploading point to Moore's law to support the notion that the necessary computing power is expected to become available within a few decades. However, the actual computational requirements for running an uploaded human mind are very difficult to quantify, potentially rendering such an argument specious. Regardless of the techniques used to capture or recreate the function of a human mind, the processing demands are likely to be immense, due to the large number of neurons in the human brain along with the considerable complexity of each neuron. Required computational capacity strongly depends on the chosen level of simulation model scale:<ref name="Roadmap" /> {| class="wikitable" ! Level ||CPU demand<br />(FLOPS)||Memory demand<br />(Tb)||$1 million super‐computer<br />(Earliest year of making) |- | Analog network population model||10<sup>15</sup>||10<sup>2</sup>||2008 |- | [[Spiking neural network]] ||10<sup>18</sup>||10<sup>4</sup>||2019 |- | [[Electrophysiology]]||10<sup>22</sup>||10<sup>4</sup>||2033 |- | [[Metabolome]] ||10<sup>25</sup>||10<sup>6</sup>||2044 |- | [[Proteome]]||10<sup>26</sup>||10<sup>7</sup>||2048 |- | States of protein complexes ||10<sup>27</sup>||10<sup>8</sup>||2052 |- | Distribution of complexes ||10<sup>30</sup>||10<sup>9</sup>||2063 |- | Stochastic behavior of single molecules||10<sup>43</sup>||10<sup>14</sup>||2111 |- |+ align="bottom" |Estimates from ''[[Anders Sandberg|Sandberg]], [[Nick Bostrom|Bostrom]], 2008'' |} ===Scanning and mapping scale of an individual=== When modelling and simulating the brain of a specific individual, a brain map or connectivity database showing the connections between the neurons must be extracted from an anatomic model of the brain. For whole brain simulation, this network map should show the connectivity of the whole [[nervous system]], including the [[spinal cord]], [[sensory receptor]]s, and [[muscle cells]]. Destructive scanning of a small sample of tissue from a [[mouse brain]] including synaptic details is possible as of 2010.<ref>{{cite web|title=New imaging method developed at Stanford reveals stunning details of brain connections|date=21 December 2009 |url=http://med.stanford.edu/news/all-news/2010/11/new-imaging-method-developed-at-stanford-reveals-stunning-details-of-brain-connections.html|publisher=Stanford Medicine}}</ref> However, if [[short-term memory]] and [[working memory]] include prolonged or repeated firing of neurons, as well as intra-neural dynamic processes, the electrical and chemical signal state of the synapses and neurons may be hard to extract. The uploaded mind may then perceive a [[memory loss]] of the events and mental processes immediately before the time of brain scanning.<ref name=Roadmap/> A full brain map has been estimated to occupy less than 2 x 10<sup>16</sup> bytes (20,000 TB) and would store the addresses of the connected neurons, the synapse type and the synapse "weight" for each of the brains' 10<sup>15</sup> synapses.<ref name=Roadmap/>{{Failed verification|date=March 2012}} However, the biological complexities of true brain function (e.g. the epigenetic states of neurons, protein components with multiple functional states, etc.) may preclude an accurate prediction of the volume of binary data required to faithfully represent a functioning human mind. ===Serial sectioning=== [[File:User-FastFission-brain.gif|thumb|right|Serial sectioning of a brain]] A possible method for mind uploading is serial sectioning, in which the brain tissue and perhaps other parts of the nervous system are frozen and then scanned and analyzed layer by layer, which for frozen samples at nano-scale requires a cryo-[[ultramicrotome]], thus capturing the structure of the neurons and their interconnections.<ref name="Merkle 1989">Merkle, R., 1989, [http://www.merkle.com/merkleDir/brainAnalysis.html Large scale analysis of neural structures] {{Webarchive|url=https://web.archive.org/web/20120204045807/http://www.merkle.com/merkleDir/brainAnalysis.html |date=2012-02-04 }}, CSL-89-10 November 1989, [P89-00173]</ref><ref>{{Cite journal |last=Chalmers |first=David J. |date=2014-06-13 |title=Uploading: A Philosophical Analysis |url=http://dx.doi.org/10.1002/9781118736302.ch6 |journal=Intelligence Unbound |pages=102–118 |doi=10.1002/9781118736302.ch6|isbn=9781118736418 }}</ref> The exposed surface of frozen nerve tissue would be scanned and recorded, and then the surface layer of tissue removed. While this would be a very slow and labor-intensive process, research is underway to automate the collection and microscopy of serial sections.<ref name="ATLUM Project">[http://www.mcb.harvard.edu/lichtman/ATLUM/ATLUM_web.htm ATLUM Project], {{webarchive|url=https://web.archive.org/web/20080201055512/http://www.mcb.harvard.edu/lichtman/ATLUM/ATLUM_web.htm|date=2008-02-01}}.</ref> The scans would then be analyzed, and a model of the neural net recreated in the system into which the mind was being uploaded. There are uncertainties with this approach using current microscopy techniques. If it is possible to replicate neuron function from its visible structure alone, then the resolution afforded by a [[scanning electron microscope]] would suffice for such a technique.<ref name="ATLUM Project"/> However, as the function of brain tissue is partially determined by molecular events (particularly at [[synapse]]s, but also at [[Ion channel|other]] places on the neuron's cell [[Membrane potential|membrane]]), this may not suffice for capturing and simulating neuron functions. It may be possible to extend the techniques of serial sectioning and to capture the internal molecular makeup of neurons, through the use of sophisticated [[immunohistochemistry]] staining methods that could then be read via [[confocal laser scanning microscopy]]. However, as the physiological genesis of 'mind' is not currently known, this method may not be able to access all of the necessary biochemical information to recreate a human brain with sufficient fidelity. ===Brain imaging=== [[File:Connectome extraction procedure.jpg|thumb|upright=3|Process from MRI acquisition to whole brain structural network<ref>{{cite journal |doi=10.1371/journal.pbio.0060159 |title=Mapping the Structural Core of Human Cerebral Cortex |year=2008 |last1=Hagmann |first1=Patric |last2=Cammoun |editor1-first=Karl J. |first2=Leila |last3=Gigandet |first3=Xavier |last4=Meuli |editor1-last=Friston |first4=Reto |last5=Honey |first5=Christopher J. |last6=Wedeen |first6=Van J. |last7=Sporns |first7=Olaf |last8=Friston |first8=Karl J. |journal=PLOS Biology |volume=6 |pages=e159 |pmid=18597554 |issue=7 |pmc=2443193 |doi-access=free }}</ref>]] [[File:Magnetoencephalography.svg|thumb|right|[[Magnetoencephalography]]]] It may be possible to create functional 3D maps of the brain activity, using advanced [[neuroimaging]] technology, such as [[functional MRI]] (fMRI, for mapping change in blood flow), [[magnetoencephalography]] (MEG, for mapping of electrical currents), or combinations of multiple methods, to build a detailed three-dimensional model of the brain using non-invasive and non-destructive methods. Today, fMRI is often combined with MEG for creating functional maps of human cortex during more complex cognitive tasks, as the methods complement each other. Even though current imaging technology lacks the spatial resolution needed to gather the information needed for such a scan, important recent and future developments are predicted to substantially improve both spatial and temporal resolutions of existing technologies.<ref>{{cite journal |last1=Glover |first1=Paul |last2=Bowtell |first2=Richard |year=2009 |title=Medical imaging: MRI rides the wave |journal=Nature |volume=457 |issue=7232 |pages=971–972 |bibcode=2009Natur.457..971G |doi=10.1038/457971a |pmid=19225512 |s2cid=205044426 |doi-access=free}}</ref> ===Brain simulation=== {{main|Brain simulation}} Ongoing work in the field of brain simulation includes partial and whole simulations of some animals.<ref name="ReferenceA"/> For example, the ''C. elegans'' roundworm,<ref>{{Cite journal |last1=Niebur |first1=E. |last2=Erdös |first2=P. |date=November 1993 |title=Theory of the locomotion of nematodes: control of the somatic motor neurons by interneurons |url=https://pubmed.ncbi.nlm.nih.gov/8260760 |journal=Mathematical Biosciences |volume=118 |issue=1 |pages=51–82 |doi=10.1016/0025-5564(93)90033-7 |issn=0025-5564 |pmid=8260760}}</ref> ''Drosophila'' fruit fly,<ref>{{Cite book |date=2010 |title=The 2010 International Joint Conference on Neural Networks (IJCNN) |chapter-url=https://ieeexplore.ieee.org/document/5596513 |journal=IEEE|doi=10.1109/IJCNN.2010.5596513 |last1=Arena |first1=Paolo |last2=Patane |first2=Luca |last3=Termini |first3=Pietro Savio |chapter=An insect brain computational model inspired by Drosophila melanogaster: Simulation results |pages=1–8 |isbn=978-1-4244-6916-1 }}</ref> and mouse<ref>{{Cite news |date=2007-04-27 |title=Mouse brain simulated on computer |url=http://news.bbc.co.uk/2/hi/technology/6600965.stm |access-date=2024-05-10 |work=BBC |language=en-GB}}</ref> have all been simulated to various degrees. The [[Blue Brain Project]], initiated by the Brain and Mind Institute of the ''[[École Polytechnique Fédérale de Lausanne]]'' in Switzerland, is an attempt to create a synthetic brain by reverse-engineering mammalian brain circuitry, in order to accelerate experimental research on the brain.<ref>{{cite web |date=19 May 2015 |title=Bluebrain - EPFL |url=http://bluebrainproject.epfl.ch/FAQs.htm |archive-url=https://web.archive.org/web/20060114070312/http://bluebrainproject.epfl.ch/FAQs.htm |archive-date=2006-01-14 |work=EPFL |quote=Once this replica is built, we will be able to do experiments that normally take us years or are prohibitively expensive or too difficult to perform. This will greatly accelerate the pace of research.}}</ref> In 2009, after a successful simulation of part of a rat brain, the director [[Henry Markram]] claimed that "A detailed, functional artificial human brain can be built within the next 10 years".<ref>{{cite news |last=Fildes |first=Jonathan |date=22 July 2009 |title=Artificial brain '10 years away' |url=http://news.bbc.co.uk/2/hi/8164060.stm |access-date= |website=bbc.co.uk |publisher=[[BBC News]]}}</ref> In 2013, Markram became the director of the new decade-long [[Human Brain Project]]. But less than two years into it, the project was recognized to be mismanaged and its claims overblown, and Markram was asked to step down.<ref>{{Cite web |last=Epstein |first=Robert |date=18 May 2016 |title=The empty brain |url=https://aeon.co/essays/your-brain-does-not-process-information-and-it-is-not-a-computer |access-date=2021-04-04 |website=Aeon |language=en}}</ref><ref>{{Cite web |last=Theil |first=Stefan |date=2015-10-01 |title=Why the human brain project went wrong and how to fix it |url=https://www.scientificamerican.com/article/why-the-human-brain-project-went-wrong-and-how-to-fix-it/ |access-date=2021-04-04 |website=[[Scientific American]]}}</ref> ==Issues== ===Philosophical issues=== The main philosophical problem faced by "mind uploading" or mind copying is the [[hard problem of consciousness]]: the difficulty of explaining how a physical entity such as a human can have [[qualia]], [[Consciousness#Types of consciousness|phenomenal consciousness]], or [[Subjective character of experience|subjective experience]].<ref name="original-paper2">{{cite journal |last=Chalmers |first=David |date=1995 |title=Facing up to the problem of consciousness |journal=[[Journal of Consciousness Studies]] |volume=2 |issue=3 |pages=200–219}}</ref> Many philosophical responses to the hard problem entail that mind uploading is fundamentally or practically impossible, while others are compatible with at least some formulations of mind uploading. Many proponents of mind uploading defend the possibility of mind uploading by recourse to [[physicalism]], which includes the philosophical belief that consciousness is an [[emergence|emergent]] feature that arises from large neural network high-level patterns of organization, which could be realized in other processing devices. Mind uploading relies on the idea that the human mind (the "self" and the long-term memory) reduces to the current neural network paths and the weights of synapses in the brain. In contrast, many [[Dualism (philosophy of mind)|dualistic]] and [[Idealism|idealistic]] accounts seek to avoid the hard problem of consciousness by explaining it in terms of immaterial (and presumably inaccessible) substances like [[soul]], which would pose a fundamental or at least practical challenge to the feasibility of artificial consciousness in general.<ref>{{Cite journal |last1=Kastrup |first1=Bernardo |year=2018 |title=The Universe in Consciousness |url=https://philpapers.org/rec/KASTUI |journal=Journal of Consciousness Studies |volume=25 |issue=5–6 |pages=125–155}}</ref> Assuming physicalism is true, the mind can be defined as the information state of the brain, so it is immaterial only in the same sense as the information content of a data file, or the state of software residing in a computer's memory. In this case, data specifying the information state of the neural network could be captured and copied as a "computer file" from the brain and re-implemented into a different physical form.<ref>{{cite web|url=http://hplusmagazine.com/2013/06/17/clearing-up-misconceptions-about-mind-uploading/|title=Clearing Up Misconceptions About Mind Uploading|work=h+ Media|date=June 17, 2013 |first= Franco |last= Cortese}}</ref> This is not to deny that minds are richly adapted to their substrates.<ref>{{cite journal|url=http://faculty.cs.tamu.edu/choe/ftp/publications/choe-ijmc12-preprint.pdf|title=Time, Consciousness, and Mind Uploading|journal=International Journal of Machine Consciousness|year=2012|doi=10.1142/S179384301240015X|volume=04|issue=1|pages=257 |author1=Yoonsuck Choe |author2= Jaerock Kwon |author3= Ji Ryang Chung }}</ref> An analogy to mind uploading is to copy the information state of a computer program from the memory of the computer on which it is executing to another computer and then continue its execution on the second computer. The second computer may perhaps have different hardware architecture, but it [[computer emulator|emulates]] the hardware of the first computer. These philosophical issues have a long history. In 1775, [[Thomas Reid]] wrote: “I would be glad to know... whether when my brain has lost its original structure, and when some hundred years after the same materials are fabricated so curiously as to become an intelligent being, whether, I say that being will be me; or, if, two or three such beings should be formed out of my brain; whether they will all be me, and consequently one and the same intelligent being.”<ref>{{cite web |title=The Duplicates Paradox (The Duplicates Problem) |url=http://www.benbest.com/philo/doubles.html |work=benbest.com}}</ref> Although the name of ''the hard problem of consciousness'' was coined in 1994, debate surrounding the problem itself is ancient. [[Augustine of Hippo]] argued against physicalist "Academians" in the 5th century, writing that consciousness cannot be an illusion because only a conscious being can be deceived or experience an illusion.<ref>{{cite book |last=Augustine of Hippo |title=City of God |chapter=Book 11, Chapter 26}}</ref> [[René Descartes]], the founder of [[Mind–body dualism|mind-body dualism]], made a similar objection in the 17th century, coining the popular phrase ''"Je pense, donc je suis"'' ("I think, therefore I am").<ref>{{cite book |last=Descartes |first=René |title=Discourse on the Method |year=1637 |chapter=4}}</ref> Although physicalism is known to have been proposed in ancient times, [[Thomas Henry Huxley|Thomas Huxley]] was among the first to describe mental experience as merely an [[epiphenomenon]] of interactions within the brain, having no causal power of its own and being entirely downstream from the brain's activity.<ref>{{Citation |last=Robinson |first=William |title=Epiphenomenalism |date=2023 |encyclopedia=The Stanford Encyclopedia of Philosophy |editor-last=Zalta |editor-first=Edward N. |url=https://plato.stanford.edu/archives/sum2023/entries/epiphenomenalism/ |access-date=2024-05-16 |edition=Summer 2023 |publisher=Metaphysics Research Lab, Stanford University |editor2-last=Nodelman |editor2-first=Uri}}</ref> A considerable portion of [[transhumanists]] and [[singularitarians]] place great hope in the belief that they may become immortal, by creating one or many non-biological functional copies of their brains, thereby leaving their "biological shell". However, the philosopher and transhumanist [[Susan Schneider (philosopher)|Susan Schneider]] claims that at best, uploading would create a copy of the original person's mind.<ref name="Schneider">{{cite news|last=Schneider|first=Susan|title=The Philosophy of 'Her'|url=http://opinionator.blogs.nytimes.com/2014/03/02/the-philosophy-of-her/?_php=true&_type=blogs&_r=0|access-date=May 7, 2014|newspaper=The New York Times|date=March 2, 2014}}</ref> Schneider agrees that consciousness has a computational basis, but this does not mean we can upload and survive. According to her views, "uploading" would probably result in the death of the original person's brain, while only outside observers can maintain the illusion of the original person still being alive. For it is implausible to think that one's consciousness would leave one's brain and travel to a remote location; ordinary physical objects do not behave this way. Ordinary objects (rocks, tables, etc.) are not simultaneously here, and elsewhere. At best, a copy of the original mind is created.<ref name="Schneider" /> [[Neural correlates of consciousness]], a sub-branch of neuroscience, states that consciousness may be thought of as a state-dependent property of some undefined [[Complex systems|complex]], adaptive, and highly interconnected biological system.<ref>{{Cite book|title= Fundamental neuroscience|date=2008|publisher=Elsevier / Academic Press| last= Squire| first= Larry R.|isbn= 9780123740199|edition= 3rd|location=Amsterdam|oclc=190867431}}</ref> Others have argued against such conclusions. For example, Buddhist transhumanist James Hughes has pointed out that this consideration only goes so far: if one believes the self is an illusion, worries about survival are not reasons to avoid uploading,<ref name="The Transhumanist Reader">{{cite book|last=Hughes|first=James|title=Transhumanism and Personal Identity|date=2013|publisher=Wiley|url=http://eu.wiley.com/WileyCDA/WileyTitle/productCd-1118334299.html}}</ref> and Keith Wiley has presented an argument wherein all resulting minds of an uploading procedure are granted equal primacy in their claim to the original identity, such that survival of the self is determined retroactively from a strictly subjective position.<ref>{{cite news|last=Wiley|first=Keith|title=Response to Susan Schneider's "Philosophy of 'Her"|url=http://hplusmagazine.com/2014/03/26/response-to-susan-schneiders-the-philosophy-of-her/|access-date=7 May 2014| work= H+Magazine|date=March 20, 2014}}</ref><ref name="WileyK_Taxonomy">{{cite book |last1= Wiley |first1= Keith |title= A Taxonomy and Metaphysics of Mind-Uploading|date=September 2014|publisher=Humanity+ Press and Alautun Press| isbn= 978-0692279847|edition= 1st|url=http://alautunpress.com|access-date=16 October 2014}}</ref> Some have also asserted that consciousness is a part of an extra-biological system that is yet to be discovered; therefore it cannot be fully understood under the present constraints of neurobiology. Without the transference of consciousness, true mind-upload or perpetual immortality cannot be practically achieved.<ref>{{Cite web|url=https://medium.com/@anthrobot/on-achieving-immortality-3ed1d567f7a2|title=On Achieving Immortality|last=Ruparel|first=Bhavik|date=2018-07-30|website= medium.com |access-date=2018-07-31}}</ref> Another potential consequence of mind uploading is that the decision to "upload" may then create a mindless symbol manipulator instead of a conscious mind (see [[philosophical zombie]]).<ref>{{cite journal |url= https://www.academia.edu/1246312|title=My Brain, my Mind, and I: Some Philosophical Problems of Mind-Uploading |volume= 4 |issue= 1 |pages= 187–200| first= Michael |last= Hauskeller|journal=Academia.edu|year=2012}}</ref><ref>{{cite web| url= http://io9.com/you-ll-probably-never-upload-your-mind-into-a-computer-474941498|title=You Might Never Upload Your Brain Into a Computer| first= George |last= Dvorsky|work=io9 |date=April 17, 2013}}</ref> If a computer could process sensory inputs to generate the same outputs that a human mind does (speech, muscle movements, etc.) without necessarily having any experience of consciousness, then it may be impossible to determine whether the uploaded mind is truly conscious, and not merely an automaton that externally behaves the way a human would. Thought experiments like the [[Chinese room]] raise fundamental questions about mind uploading: If an upload displays behaviors that are highly indicative of consciousness, or even verbally insists that it is conscious, does that prove it is conscious?<ref>{{cite web |url= http://degreesofclarity.com/writing/oto_mind_uploading.pdf|title=Seeking normative guidelines for novel future forms of consciousness| first= Brandon |last= Oto|publisher=University of California, Santa Cruz|year=2011|access-date=2014-01-03 |archive-date=2014-01-03|archive-url= https://web.archive.org/web/20140103132727/http://degreesofclarity.com/writing/oto_mind_uploading.pdf|url-status=dead}}</ref> There might also be an absolute upper limit in processing speed, above which consciousness cannot be sustained. The subjectivity of consciousness precludes a definitive answer to this question.<ref>{{cite web |url= http://goertzel.org/Goertzel_IJMC_Special_Issue.pdf|title=When Should Two Minds Be Considered Versions of One Another?| first= Ben |last= Goertzel |year=2012 }}</ref> Numerous scientists, including [[Ray Kurzweil]], believe that whether a separate entity is conscious is impossible to know with confidence, since consciousness is inherently subjective (see [[solipsism]]). Regardless, some scientists believe consciousness is the consequence of computational processes which are substrate-neutral. Still other scientists believe consciousness may emerge from some form of quantum computation that is dependent on the organic substrate (see [[quantum mind]]).<ref>{{cite web| url= http://hplusmagazine.com/2013/04/21/goertzel-contra-dvorsky-on-mind-uploading/|title=Goertzel Contra Dvorsky on Mind Uploading|date=April 21, 2013| first= Sally |last= Morem|work=h+ Media}}</ref><ref>{{cite journal |url= http://www.terasemcentral.org/docs/Terasem%20Mind%20Uploading%20Experiment%20IJMC.pdf |title= The Terasem Mind Uploading Experiment| first= Martine |last= Rothblatt|pages=141–158|year=2012 |journal= [[International Journal of Machine Consciousness]] |volume=4|issue=1|doi=10.1142/S1793843012400070 |url-status=dead |archive-url= https://web.archive.org/web/20130827213457/http://www.terasemcentral.org/docs/Terasem%20Mind%20Uploading%20Experiment%20IJMC.pdf |archive-date=2013-08-27}}</ref><ref>{{cite journal|url=http://home.millsaps.edu/hopkipd/IJMC-Preprint-HopkinsUploading.pdf|title=Why Uploading Will Not Work, or, the Ghosts Haunting Transhumanism| first= Patrick D.| last= Hopkins| year= 2012 |journal=International Journal of Machine Consciousness|volume=4|issue=1|pages=229–243| doi= 10.1142/S1793843012400136| url-status= dead|archive-url= https://web.archive.org/web/20120906145410/http://home.millsaps.edu/hopkipd/IJMC-Preprint-HopkinsUploading.pdf |archive-date= 2012-09-06}}</ref> In light of uncertainty about whether mind uploads are conscious, Sandberg proposes a cautious approach:<ref name= SandbergEthics2014 /> {{Blockquote|Principle of assuming the most (PAM): Assume that any emulated system could have the same mental properties as the original system and treat it correspondingly.}} ===Ethical and legal implications=== The process of developing emulation technology raises ethical issues related to [[animal welfare]] and [[artificial consciousness]].<ref name=SandbergEthics2014>{{cite journal| first= Anders |last= Sandberg|title=Ethics of brain emulations |journal=Journal of Experimental & Theoretical Artificial Intelligence| date=14 April 2014 |volume= 26 |issue= 3 |pages= 439–457| doi=10.1080/0952813X.2014.895113|s2cid=14545074}}</ref> The neuroscience required to develop brain emulation would require animal experimentation, first on invertebrates and then on small mammals before moving on to humans. Sometimes the animals would just need to be euthanized in order to extract, slice, and scan their brains, but sometimes behavioral and ''in vivo'' measures would be required, which might cause pain to living animals.<ref name=SandbergEthics2014 /> In addition, the resulting animal emulations themselves might suffer, depending on one's views about consciousness.<ref name=SandbergEthics2014 /> Bancroft argues for the plausibility of consciousness in brain simulations on the basis of the "[[Qualia#David Chalmers|fading qualia]]" thought experiment of [[David Chalmers]]. He then concludes:<ref name= Bancroft2013>{{cite journal |first= Tyler D. |last= Bancroft|title=Ethical Aspects of Computational Neuroscience |journal= Neuroethics |date=Aug 2013|volume=6|issue=2|pages=415–418|doi=10.1007/s12152-012-9163-7| s2cid= 145511899 |issn= 1874-5504}}</ref> “If, as I argue above, a sufficiently detailed computational simulation of the brain is potentially operationally equivalent to an organic brain, it follows that we must consider extending protections against suffering to simulations.” Chalmers himself has argued that such virtual realities would be genuine realities.<ref>{{cite book |last=Chalmers |first=David |author-link=David Chalmers |date=2022 |title=Reality+: Virtual Worlds and the Problems of Philosophy |url=https://wwnorton.com/books/reality |location=New York |publisher=W. W. Norton & Company |isbn= 9780393635805}}</ref> However, if mind uploading occurs and the uploads are not conscious, there may be a significant opportunity cost. In the book ''[[Superintelligence: Paths, Dangers, Strategies|Superintelligence]]'', [[Nick Bostrom]] expresses concern that we could build a "Disneyland without children."<ref name="bostrom2014">{{cite book |last=Bostrom |first=Nick |title=Superintelligence: Paths, Dangers, Strategies |date=2014 |publisher=Oxford University Press |isbn=978-0199678112 |location=Oxford, England}}</ref> It might help reduce emulation suffering to develop virtual equivalents of anaesthesia, as well as to omit processing related to pain and/or consciousness. However, some experiments might require a fully functioning and suffering animal emulation. Animals might also suffer by accident due to flaws and lack of insight into what parts of their brains are suffering.<ref name=SandbergEthics2014 /> Questions also arise regarding the moral status of partial brain emulations, as well as creating neuromorphic emulations that draw inspiration from biological brains but are built somewhat differently.<ref name=Bancroft2013 /> Brain emulations could be erased by computer viruses or malware, without the need to destroy the underlying hardware. This may make assassination easier than for physical humans. The attacker might take the computing power for its own use.<ref name=EckersleySandberg2013>{{cite journal|first1= Peter |last1= Eckersley|first2= Anders |last2= Sandberg|title=Is Brain Emulation Dangerous?|journal=Journal of Artificial General Intelligence|date=Dec 2013|volume=4|issue=3|pages=170–194|doi=10.2478/jagi-2013-0011|issn=1946-0163|bibcode=2013JAGI....4..170E|doi-access=free}}</ref> Many questions arise regarding the legal personhood of emulations.<ref name=Muzyka2013 /> Would they be given the rights of biological humans? If a person makes an emulated copy of themselves and then dies, does the emulation inherit their property and official positions? Could the emulation ask to "pull the plug" when its biological version was terminally ill or in a coma? Would it help to treat emulations as adolescents for a few years so that the biological creator would maintain temporary control? Would criminal emulations receive the death penalty, or would they be given forced data modification as a form of "rehabilitation"? Could an upload have marriage and child-care rights?<ref name=Muzyka2013>{{cite journal|first= Kamil |last= Muzyka|title=The Outline of Personhood Law Regarding Artificial Intelligences and Emulated Human Entities|journal=Journal of Artificial General Intelligence|date=Dec 2013|volume=4|issue=3|pages=164–169| doi= 10.2478/jagi-2013-0010|issn=1946-0163|bibcode=2013JAGI....4..164M|doi-access=free}}</ref> If simulated minds would come true and if they were assigned rights of their own, it may be difficult to ensure the protection of "digital human rights". For example, social science researchers might be tempted to secretly expose simulated minds, or whole isolated societies of simulated minds, to controlled experiments in which many copies of the same minds are exposed (serially or simultaneously) to different test conditions.{{citation needed|date=June 2014}} Research led by cognitive scientist Michael Laakasuo has shown that attitudes towards mind uploading are predicted by an individual's belief in an afterlife; the existence of mind uploading technology may threaten religious and spiritual notions of immortality and divinity.<ref name="laakasuo2022">{{cite journal |first1= Michael |last1= Laakasuo |first2= Jukka |last2= Sundvall |first3= Marianna |last3= Drosinou |display-authors= 3| author4=Ivar Hannikainen |author5=Anton Kunnari |author6=Kathryn B. Francis |author7=Jussi Palomäki |date=2023 |title=Would you exchange your soul for immortality? – Existential Meaning and Afterlife Beliefs Predict Mind Upload Approval |journal=Frontiers in Psychology |volume=14 |doi=10.3389/fpsyg.2023.1254846 |doi-access=free |pmid=38162973 |pmc=10757642 }}</ref> ===Political and economic implications=== {{Update section|date=June 2024|reason=may not be relevant anymore, considering recent progress in large multimodal models}} Emulations might be preceded by a technological arms race driven by [[First-mover advantage|first-strike advantages]]. Their emergence and existence may lead to increased risk of war, including inequality, power struggles, strong loyalty and willingness to die among emulations, and new forms of racism, xenophobia, and religious prejudice.<ref>{{Cite journal |last=Hurtado Hurtado |first=Joshua |date=2022-07-18 |title=Envisioning postmortal futures: six archetypes on future societal approaches to seeking immortality |journal=Mortality |volume=29 |pages=18–36 |doi=10.1080/13576275.2022.2100250 |s2cid=250650618 |issn=1357-6275|doi-access=free }}</ref><ref name=EckersleySandberg2013 /> If emulations run much faster than humans, there might not be enough time for human leaders to make wise decisions or negotiate. It is possible that humans would react violently against the growing power of emulations, especially if that depresses human wages. Emulations may not trust each other, and even well-intentioned defensive measures [[security dilemma|might be interpreted as offense]].<ref name=EckersleySandberg2013 /> The book ''[[The Age of Em]]'' by [[Robin Hanson]] poses many hypotheses on the nature of a society of mind uploads, including that the most common minds would be copies of adults with personalities conducive to long hours of productive specialized work.<ref name="hanson">{{cite book |last=Hanson |first=Robin |author-link=Robin Hanson |url=https://ageofem.com/ |title=The Age of Em |date=2016 |publisher=Oxford University Press |isbn=9780198754626 |location=Oxford, England |page=528}}</ref> ===Emulation timelines and AI risk=== {{Update section|date=June 2024|reason=most sources are more than 10 years old, and may not reflect the current state of the debate}} Kenneth D. Miller, a professor of neuroscience at Columbia University and a co-director of the Center for Theoretical Neuroscience, raised doubts about the practicality of mind uploading. His major argument is that reconstructing neurons and their connections is in itself a formidable task, but it is far from being sufficient. Operation of the brain depends on the dynamics of electrical and biochemical signal exchange between neurons; therefore, capturing them in a single "frozen" state may prove insufficient. In addition, the nature of these signals may require modeling at the molecular level and beyond. Therefore, while not rejecting the idea in principle, Miller believes that the complexity of the "absolute" duplication of an individual mind is insurmountable for the nearest hundreds of years.<ref name="abondonallhopetoupload">{{Cite news |url=https://www.nytimes.com/2015/10/11/opinion/sunday/will-you-ever-be-able-to-upload-your-brain.html |title=Will You Ever Be Able to Upload Your Brain? |work=New York Times |first=Kenneth D. |last=Miller |date=October 10, 2015}}</ref> There are very few feasible technologies that humans have refrained from developing<!-- what technologies? -->. The neuroscience and computer-hardware technologies that may make brain emulation possible are widely desired for other reasons, and logically their development will continue into the future. We may also have brain emulations for a brief but significant period on the way to non-emulation based human-level AI.<ref name="hanson" /> Assuming that emulation technology will arrive, a question becomes whether we should accelerate or slow its advance.<ref name=EckersleySandberg2013 /> Arguments for speeding up brain-emulation research: * If neuroscience is the bottleneck on brain emulation rather than computing power, emulation advances may be more erratic and unpredictable based on when new scientific discoveries happen.<ref name=EckersleySandberg2013 /><ref name=ShulmanSandberg2010>{{cite journal|last1= Shulman |first1= Carl |first2= Anders |last2= Sandberg|title=Implications of a Software-Limited Singularity|journal=ECAP10: VIII European Conference on Computing and Philosophy |year= 2010 |url= http://intelligence.org/files/SoftwareLimited.pdf|access-date=17 May 2014|editor1-first=Klaus|editor1-last=Mainzer}}</ref><ref name=Hanson2009 /> Limited computing power would mean the first emulations would run slower and so would be easier to adapt to, and there would be more time for the technology to transition through society.<ref name=Hanson2009>{{cite web|last1=Hanson|first1=Robin|title=Bad Emulation Advance|url=http://www.overcomingbias.com/2009/11/bad-emulation-advance.html|website=Overcoming Bias|access-date=28 June 2014|date=26 Nov 2009}}</ref> * Improvements in manufacturing, 3D printing, and nanotechnology may accelerate hardware production,<ref name =EckersleySandberg2013 /> which could increase the "computing overhang"<ref name="MuehlhauserSalamon2012">{{cite book |last1=Muehlhauser |first1=Luke |title=Singularity Hypotheses: A Scientific and Philosophical Assessment |last2=Salamon |first2=Anna |publisher=Springer |year=2012 |editor=Eden |editor-first=Amnon |chapter=Intelligence Explosion: Evidence and Import |editor2=Søraker |editor-first2=Johnny |editor3=Moor |editor-first3=James H. |editor4=Steinhart |editor-first4=Eric |chapter-url=http://intelligence.org/files/IE-EI.pdf}}</ref> from excess hardware relative to neuroscience. * If one AI-development group had a lead in emulation technology, it would have more subjective time to win an arms race to build the first superhuman AI. Because it would be less rushed, it would have more freedom to consider AI risks.<ref name=SalamonMuehlhauser2012 /><ref name=Bostrom2014 /> Arguments for slowing brain-emulation research: * Greater investment in brain emulation and associated cognitive science might enhance the ability of artificial intelligence (AI) researchers to create "neuromorphic" (brain-inspired) algorithms, such as neural networks, reinforcement learning, and hierarchical perception. This could accelerate [[Existential risk from artificial general intelligence|risks from uncontrolled AI]].<ref name=EckersleySandberg2013 /><ref name=Bostrom2014>{{cite book| last1= Bostrom |first1= Nick |title=Superintelligence: Paths, Dangers, Strategies|chapter=Ch. 14: The strategic picture| date= 2014 |publisher= Oxford University Press|isbn=978-0199678112}}</ref> Participants at a 2011 AI workshop estimated an 85% probability that neuromorphic AI would arrive before brain emulation. This was based on the idea that brain emulation would require understanding of the workings and functions of the different brain components, along with the technological know-how to emulate neurons. To counter this idea, reverse engineering the Microsoft Windows code base is already hard, so reverse engineering the brain would likely be much harder. By a very narrow margin, the participants on balance leaned toward the view that accelerating brain emulation would increase expected AI risk.<ref name=SalamonMuehlhauser2012>{{cite web| first1= Anna |last1= Salamon | first2= Luke |last2= Muehlhauser|title=Singularity Summit 2011 Workshop Report |url= https://intelligence.org/files/SS11Workshop.pdf|website=Machine Intelligence Research Institute| date=2012 |access-date= 28 June 2014}}</ref> * Waiting might give society more time to think about the consequences of brain emulation and develop institutions to improve cooperation.<ref name=EckersleySandberg2013 /><ref name=Bostrom2014 /> Emulation research would also accelerate neuroscience as a whole, which might accelerate medical advances, cognitive enhancement, lie detectors, and capability for [[psychological manipulation]].<ref name=Bostrom2014 /> Emulations might be easier to control than ''de novo'' AI because: # Human abilities, behavioral tendencies, and vulnerabilities are more thoroughly understood, thus control measures might be more intuitive and easier to plan.<ref name=SalamonMuehlhauser2012 /><ref name=Bostrom2014 /> # Emulations could more easily inherit human motivations.<ref name=Bostrom2014 /> # Emulations are harder to manipulate than ''de novo'' AI, because brains are messy and complicated; this could reduce risks of their rapid takeoff.<ref name=EckersleySandberg2013 /><ref name=Bostrom2014 /> Also, emulations may be bulkier and require more hardware than AI, which would also slow the speed of a transition.<ref name=Bostrom2014 /> Unlike AI, an emulation would not be able to rapidly expand beyond the size of a human brain.<ref name=Bostrom2014 /> Emulations running at digital speeds would have less intelligence differential vis-à-vis AI and so might more easily control AI.<ref name=Bostrom2014 /> As counterpoint to these considerations, Bostrom notes some downsides: # Even if we better understand human behavior, the ''evolution'' of emulation behavior under self-improvement might be much less predictable than the evolution of safe ''de novo'' AI under self-improvement.<ref name=Bostrom2014 /> # Emulations may not inherit all human motivations. Perhaps they would inherit our darker motivations or would behave abnormally in the unfamiliar environment of cyberspace.<ref name=Bostrom2014 /> # Even if there is a slow takeoff toward emulations, there would still be a second transition to ''de novo'' AI later on. Two intelligence explosions may mean more total risk.<ref name=Bostrom2014 /> Because of the postulated difficulties that a whole brain emulation-generated [[superintelligence]] would pose for the control problem, computer scientist [[Stuart J. Russell]] in his book ''[[Human Compatible]]'' rejects creating one, simply calling it "so obviously a bad idea".<ref>{{Cite book|last=Russell|first=Stuart|author-link=Stuart J. Russell|title=[[Human Compatible|Human Compatible: Artificial Intelligence and the Problem of Control]]|publisher=[[Viking Press]]|year=2019|isbn=978-0-525-55861-3|oclc=1113410915}}</ref> ==Advocates== In 1979, [[Hans Moravec]] (1979) described and endorsed mind uploading using a brain surgeon.<ref>{{cite web| url= https://www.frc.ri.cmu.edu/~hpm/project.archive/general.articles/1978/analog.1978.html |title= Today's Computers, Intelligent Machines and Our Future| publisher= [[Field Robotics Center]], [[Robotics Institute]], [[Carnegie Mellon University]] |archive-url= https://web.archive.org/web/20221219111328/https://www.frc.ri.cmu.edu/~hpm/project.archive/general.articles/1978/analog.1978.html |archive-date=2022-12-19 |first= Hans |last= Moravec| year= 1979| access-date= }} [[wikidata:Q115765733|wikidata]]</ref> Moravec used a similar description in 1988, calling it "transmigration".<ref name= "moravec1988">{{cite book| title= Mind Children |first= Hans |last= Moravec| year= 1988| publisher= }}</ref> [[Ray Kurzweil]], director of engineering at [[Google]], has long predicted that people will be able to "upload" their entire brains to computers and become "digitally immortal" by 2045. Kurzweil made this claim for many years, e.g. during his speech in 2013 at the [[2045 Initiative|Global Futures 2045]] International Congress in New York, which claims to subscribe to a similar set of beliefs.<ref>{{cite web|url=https://www.huffpost.com/entry/mind-uploading-2045-futurists_n_3458961 |title='Mind Uploading' & Digital Immortality May Be Reality By 2045, Futurists Say |work=Huffington Post |date=June 18, 2013| access-date= March 8, 2024}}</ref> Mind uploading has also been advocated by a number of researchers in [[neuroscience]] and [[artificial intelligence]], such as Marvin Minsky.{{Citation needed|date=March 2016}} In 1993, Joe Strout created a small web site called the Mind Uploading Home Page, and began advocating the idea in [[cryonics]] circles and elsewhere on the net. That site has not been actively updated in recent years, but it has spawned other sites including MindUploading.org, run by [[Randal A. Koene]], who also moderates a mailing list on the topic. These advocates see mind uploading as a medical procedure which could eventually save countless lives. Many transhumanists look forward to the development and deployment of mind uploading technology, with transhumanists such as [[Nick Bostrom]] predicting that it will become possible within the 21st century due to technological trends such as Moore's law.<ref name=Roadmap/> [[Michio Kaku]], in collaboration with [[Science (TV network)|Science]], hosted a documentary, ''Sci Fi Science: Physics of the Impossible'', based on his book ''[[Physics of the Impossible]]''. Episode four, titled "How to Teleport", mentions that mind uploading via techniques such as [[quantum entanglement]] and whole brain emulation using an advanced [[Magnetic resonance imaging|MRI machine]] may enable people to be transported vast distances at near light-speed. The book ''Beyond Humanity: CyberEvolution and Future Minds'' by [[Gregory S. Paul]] & Earl D. Cox, is about the eventual (and, to the authors, almost inevitable) evolution of computers into [[sentience|sentient]] beings, but also deals with human mind transfer. [[Richard Doyle (actor)|Richard Doyle]]'s ''Wetwares: Experiments in PostVital Living'' deals extensively with uploading from the perspective of distributed embodiment, arguing for example that humans are currently part of the "artificial life phenotype". Doyle's vision reverses the polarity on uploading, with artificial life forms such as uploads actively seeking out biological embodiment as part of their reproductive strategy. == In fiction == {{main|Mind uploading in fiction}} Mind uploading—transferring an individual's personality to a computer—appears in several works of [[science fiction]].<ref name="SFEUpload">{{Cite encyclopedia |year=2022 |title=Upload |encyclopedia=[[The Encyclopedia of Science Fiction]] |url=https://sf-encyclopedia.com/entry/upload |access-date=2024-03-29 |author1-link=David Langford |editor1-last=Clute |editor1-first=John |edition=4th |author2-last=Stableford |author2-first=Brian |author1-last=Langford |author1-first=David |author2-link=Brian Stableford |editor1-link=John Clute |editor2-last=Langford |editor2-first=David |editor2-link=David Langford |editor3-last=Sleight |editor3-first=Graham |editor3-link=Graham Sleight}}</ref> It is distinct from the concept of transferring a consciousness from one human body to another.<ref name="WebbMindUploading">{{Cite book |last=Webb |first=Stephen |author-link=Stephen Webb (scientist) |title=All the Wonder that Would Be: Exploring Past Notions of the Future |date=2017 |publisher=Springer |isbn=978-3-319-51759-9 |series=Science and Fiction |pages=276–278 |language=en |chapter=Mind Uploading |doi=10.1007/978-3-319-51759-9_10 |chapter-url=https://books.google.com/books?id=TVPJDgAAQBAJ&pg=PA276}}</ref><ref name="Fischer">{{multiref2|{{Cite book |last1=Fischer |first1=John Martin |title=Immortal Engines: Life Extension and Immortality in Science Fiction and Fantasy |last2=Curl |first2=Ruth |publisher=[[University of Georgia Press]] |year=1996 |isbn=0-8203-1733-0 |editor-last=Slusser |editor-first=George |editor-link=George Edgar Slusser |location=[[Athens, Georgia]] |pages=3–12 |chapter=Philosophical Models of Immortality in Science Fiction |author-link=John Martin Fischer |oclc=34319944 |editor-last2=Westfahl |editor-first2=Gary |editor-link2=Gary Westfahl |editor-last3=Rabkin |editor-first3=Eric S.}}|{{Cite book |last1=Fischer |first1=John Martin |title=Our Stories: Essays on Life, Death, and Free Will |last2=Curl |first2=Ruth |publisher=Oxford University Press |year=2009 |isbn=978-0-19-537495-7 |editor-last=Fischer |editor-first=John Martin |editor-link=John Martin Fischer |pages=93–101 |language=en |chapter=Appendix to Chapter 6: Philosophical Models of Immortality in Science Fiction |author-link=John Martin Fischer |chapter-url=https://books.google.com/books?id=xeAiDCSqEBMC&pg=PA93}} }}</ref> It is sometimes applied to a single person and other times to an entire society.<ref name="GreenwoodComputers">{{Cite book |last=Langford |first=David |author-link=David Langford |title=[[The Greenwood Encyclopedia of Science Fiction and Fantasy: Themes, Works, and Wonders]] |date=2005 |publisher=Greenwood Publishing Group |isbn=978-0-313-32951-7 |editor-last=Westfahl |editor-first=Gary |editor-link=Gary Westfahl |pages=154 |language=en |chapter=Computers |chapter-url=https://archive.org/details/greenwoodencyclo0000unse_k2b9/page/154/mode/2up}}</ref> Recurring themes in these stories include whether the computerized mind is truly [[Consciousness|conscious]], and if so, whether [[Personal identity|identity]] is preserved.<ref name="BlackfordReshapingTheHuman">{{Cite book |last=Blackford |first=Russell |author-link=Russell Blackford |title=Science Fiction and the Moral Imagination: Visions, Minds, Ethics |date=2017 |publisher=Springer |isbn=978-3-319-61685-8 |series=Science and Fiction |pages=173–174 |language=en |chapter=Reshaping the Human |chapter-url=https://books.google.com/books?id=jlU0DwAAQBAJ&pg=PA173}}</ref> It is a common feature of the [[cyberpunk]] subgenre,<ref name="HistoricalDictionaryOfScienceFictionInLiteraturArtificialIntelligence">{{Cite book |last=Booker |first=M. Keith |title=Historical Dictionary of Science Fiction in Literature |date=2014 |publisher=Rowman & Littlefield |isbn=978-0-8108-7884-6 |pages=28 |language=en |chapter=Artificial Intelligence (AI) |quote=Cyberpunk writers and their successors have also frequently imagined the uploading of human minds into computers, thus creating a special sort of artificial intelligence that can free individuals of the limitations of biological bodies, a notion that would be notably extended in the work of Greg Egan. |chapter-url=https://books.google.com/books?id=WRi7BAAAQBAJ&pg=PA28}}</ref> sometimes taking the form of [[digital immortality]].<ref name="Fischer" /><ref name="GreenwoodEncyclopediaImmortality">{{Cite book |last=Westfahl |first=Gary |author-link=Gary Westfahl |title=[[The Greenwood Encyclopedia of Science Fiction and Fantasy: Themes, Works, and Wonders]] |date=2005 |publisher=Greenwood Publishing Group |isbn=978-0-313-32951-7 |editor-last=Westfahl |editor-first=Gary |editor-link=Gary Westfahl |pages=418–420 |language=en |chapter=Immortality and Longevity |chapter-url=https://archive.org/details/greenwoodencyclo0000unse_k2b9/page/418/mode/2up}}</ref> ==See also== {{div col|colwidth=21em}} * [[BRAIN Initiative]] * [[Brain transplant]] * [[Brain-reading]] * [[Cyborg]] * [[Cylon (reimagining)]] * [[Democratic transhumanism]] * [[Human Brain Project]] * [[Isolated brain]] * [[Neuralink]] * [[Open individualism]] * [[Posthumanization]] * [[Robotoid]] * [[Ship of Theseus]]—thought experiment asking if objects having all parts replaced fundamentally remain the same object * [[Simulation hypothesis]] * [[Synthetic telepathy#Synthetic telepathy.2Fsilent communication|Technologically enabled telepathy]] * [[Teletransportation paradox]] * [[Thought recording and reproduction device]] * [[Turing test]] * ''[[The Future of Work and Death]]'' * [[Vertiginous question]] * [[Chinese room]] * [[2045 Initiative]] * [[Dmitry Itskov]] * [[Miguel Nicolelis]] * [[Neural network (machine learning)]] {{div col end}} ==References== {{Reflist}} {{Life extension}} {{BCI}} {{emerging technologies|topics=yes|neuro=yes|infocom=yes}} {{Science fiction}} {{DEFAULTSORT:Mind Uploading}} [[Category:Fictional technology]] [[Category:Hypothetical technology]] [[Category:Immortality]] [[Category:Neurotechnology]] [[Category:Transhumanism]] [[Category:Posthumanism]]
Summary:
Please note that all contributions to Ikwipedia are considered to be released under the Creative Commons Attribution-ShareAlike (see
Ikwipedia:Copyrights
for details). If you do not want your writing to be edited mercilessly and redistributed at will, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource.
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Toggle limited content width