I think SOMA made it pretty clear we're never uploading jack shit, at best we're making a copy for whom it'll feel as if they've been uploaded, but the original remains behind as well.
I mean, if I die instantaneously and painlessly, and conciousness is seemingly continuous for the surviving copy, why would I care?
My conciousness might not continue but I lose consciousness every day. Someone exists who is me and lives their (my) life. I totally understand peoples aversion to death but I also don't see any difference to falling asleep and waking up. You lose consciousness, then a person who's lived your life and is you regains consciousness. Idk
That's actually not true. When you cut/paste a file on your computer (for most computers), it's much faster than copying the file. Deleting the file is also not instant, so copy and delete should be the slowest of the three operations.
When you cut and paste a file, you're just renaming the file or updating the file database. It's different how that works depending on your file system, but it typically never involves rewriting much of the data of the file.
That's what I've always thought more or less, to have a chance you would need a method where mental processing starts to be shared in both, then transfers more and more to the inorganic platform till it's 100% and the organic isn't working anymore.
Yeah, like replacing individual braincells with more durable mechanisms.
Idk, maybe they would be cellular as well.
..that makes me wonder, maybe it is possible to transfer consciousness even with traditional biological mechanism?
You would have to functionally duplicate the exact structure of the brain or its consciousness while having the duplication mechanism destroy the thing it was reading at almost exactly the same time. And even then, that's not really solving the issue.
I was just annoyed at the protagonist for expecting anything else. The exact same thing already happened 2 times to the protagonist (initial copy at beginning of the game, then move to the other suit). Plus it's reinforced in the found notes for good measure. So by the ending, the player knows exactly what's going to happen and so should the protagonist, but somehow he's surprised.
Ahh, but here's the question. Who are you? The you who did the upload, or the you that got uploaded, retaining the memories of everything you did before the upload? Go on, flip that coin.
If you are the version doing the upload, you're staying behind. The other "you" pops into existence feeling as if THEY are the original, so from their perspective, it's as if they won the coin flip.
Why would you want a simulation version? You will get saved at "well rested." It will be an infinite loop of put to work for several hours and then deleted. You won't even experience that much, your consciousness is gone.
If anyone's interested in a hard sci-fi show about uploading consciousness they should watch the animated series Pantheon. Not only does the technology feel realistic, but the way it's created and used by big tech companies is uncomfortably real.
The show got kinda screwed over on advertising and fell to obscurity because of streaming service fuck ups and region locking, and I can't help but wonder if it's at least partially because of its harsh criticisms of the tech industry.
Well yeah, if you passed a reference then once the original is destroyed it would be null. The real trick is to make a copy and destroy the original reference at the same time, that way it never knows it wasn't the original.
You see, with Effective Altruism, we'll destroy the world around us to serve a small cadre of ubermensch tech bros, who will then somehow in the next few centuries go into space and put supercomputers on other planets that run simulations of people. You might actually be in one of those simulations right now, so be grateful.
We are very smart and not just reckless, over-indulged douchebags who jerk off to the smell of our own farts.
Sure, some probably do. And you can be sceptical and discuss why that's a dangerous and undemocratic direction. Effective Altruism is a question, not an answer. In thr community, asking for and being open to critical feedback is encouraged as the main tenet of good culture.
But if you look at the amounts, most EAs donate most to helping the poorest people alive today. Because it is so obviously good, and proven to work with high certainty.
I think that really depends on the implementation details. For example, consider a thought experiment where artificial neurons can be created that behave just the same as biological ones. Then each of your neurons is replaced by an artificial version while you are still conscious. You wouldn't notice losing a single neuron at a time, in fact this regularly happens already. Yet, over time, all your biological neurons could be replaced by artificial ones at which point your consciousness will have migrated to a new substrate.
Alternatively, what if one of your hemispheres was replaced by an artificial one. What if an artificial hemisphere was added into the mix in addition to the two you have. What if a dozen artificial hemispheres were added, or a thousand, would the two original biological ones still be the most relevant parts of you?
A value is trying to be set on a copy of a slice from a DataFrame. Try using .loc[row_indexer,col_indexer] = value instead See the caveats in the documentation: http://pandas.pydata.org/pandas-docs/stable/indexing.html#indexing-view-versus-copy
The Closest-Continuer schema is a theory of identity according to which identity through time is a function of appropriate weighted dimensions. A at time 1 and B at time 2 are the same just in case B is the closest continuer of A, according to a metric determined by continuity of the appropriate weighted dimensions.
What do you think happens in the infinite loop that "runs you" moment to moment? Passing the same instance of consciousness to itself, over and over?
Consciousness isn't an instance. It isn't static, it's a constantly self-modifying waveform that remembers bits about its former self from moment to moment.
You can upload it without destroying the original if you can find a way for it to meaningfully interact with processing architecture and media that are digital in nature; and if you can do that without shutting you off. Here's the kinky part: We can already do this. You can make a device that takes a brain signal and stimulates a remote device; and you can stimulate a brain with a digital signal. Set it up for feedback in a manner similar to the ongoing continuous feedback of our neural structures and you have now extended yourself into a digital device in a meaningful way.
Then you just keep adding to that architecture gradually, and gradually peeling away redundant bits of the original brain hardware, until most or all of you is being kept alive in the digital device instead of the meat body. To you, it's continuous and it's still you on the other end. Tada, consciousness uploaded.
It would be easier to record than upload. Since upload requires at least a decode steps. Given the fleeting nature of existence how does one confirm the decoding? This also requires we create a simulated brain, which seems more difficult and resource intensive than forming a new biological brain remotely connected to your nervous system inputs.
Recording all inputs in real time and play them back across a blank nervous system will create an active copy. The inputs can be saved so they can be played back later in case of clone failure. As long as the inputs are record until the moment of death, the copy will be you minus the death so you wouldn't be aware you're a copy. Attach it to fresh body and off you go.
Failure mode would take your literal lifetime to reform your consciousness but what's a couple decades to an immortal.
We already have the program to create new brains. It's in our DNA. A true senior developer knows better than to try and replicate black box code that's been executing fine. We don't even understand consciousness enough to pretend we're going to add new features so why waste the effort creating a parallel system of a black box.
Scheduled reboots of a black box system is common practice. Why pretend we're capable of skipping steps.
It wouldn’t really matter until you get to the brain. Very little of your body’s “processing” happens outside of your brain. Basically all of your consciousness is in there. There are some quick nerve paths that loop through your spine for things like moving your hand away when you touch a hot object, but that’s not really consciousness.