Junghalli wrote:You might be farming out its operations and feeling secure in the knowledge that you're forcing it to show it's work when it's actually 100 times smarter than you think it is and is actually doing all the real work inside its box and feeding you deceptive simulations created to convince you the software it wrote for you is a harmless complex anti-spam program when it really contains the thing's compressed mindstate.
You can't squeeze blood from a stone. Ultimately, the objects in our universe (atoms) operate by very simple rules, only the conditions they are operating under are continuously changing and there are so damn many of them. Thus, the most faithful simulations have lots and lots of little elements behaving in a very simple manner but the conditions are updated very, very often. The only way to solve that kind of problem is to throw lots and lots of computer power at it. And I don't mean intelligence, I mean brute force and ignorance. Intelligence only matters when the problem is hard to understand. The behavior of atoms is actually pretty easy to understand, comparatively... there's just shittons of calculations to carry through.
As for the mindstate thing, again you can't squeeze blood from a stone, especially if that stone is already being used for something. There's a limit to how much data you can send in a given number of bits. Any further encoding means that you're actually letting the decompression program carry some of the information. Also, the information of your supercompressed mindstate has to compete with the quite different information which is the code's acutal purpose.
Junghalli wrote:I'm a little wary of trusting that it would be impossible for a superintelligence to hide malicious code from much less intelligent human software experts in such a format, but I know nothing about this subject so I'll wait until somebody more knowledgeable about computers who can intelligently discuss that comes along.
Data cannot be executed as code unless the computer is told that it
is code. Hiding code in something like a molecular simulation specification file is worthless, because even if it happens to conform to the format exactly, the "code" will then be regarded as state data for a physical system to be simulated. Extracting the code will require an accomplice, or code the AI has
already managed to plant outside
as code — but that requires you to assume the AI has done
exactly what it now seeks to accomplish.
Junghalli wrote:That wasn't my point. My point was that your idea of farming out the AI's operations safely requires that we be able to understand its operations.
Bullshit.
Junghalli wrote:If we can't understand its operations we'll have no idea what operations it's running, and the best thing we can do is ask it to plug data we understand into less efficient computer simulations that we designed and understand. Which might very well have benefits, but reduced benefits compared to trusting the AI, which again gets back to the point that the long term viability of adversarial containment methods in the hands of real humans is dubious because they greatly reduce the use you can get out of your superintelligence, so there are incentives to drop them.
In other words, we cannot utilize the AI
as well if we confine it. That's a different statement than saying that we cannot utilize the AI at all if we confine it. We get some use out of the AI, even if its confined (provided it cooperates), just not as much use as YOU want.
Sorry, but when you're dealing with a thing that can destroy you easily, extreme caution is warranted. Unless you can prove the AI is friendly, letting it out, or letting design code not run on secure computers,
is not safe.
Junghalli wrote:The problem isn't so much the AI saying "I can't explain it to you because it's Lovecraftian non-Euclidian stuff that your puny brain would never comprehend" (which is unconvincing for a number of reasons that should probably occur to a superintelligence). The problem is more the AI feeding us a convincing but deceptive explanation.
You do know that any explanation will be testible against reality, do you not? Remember, the superintelligence is not omniscient. It can't be. We're going to be testing its ideas. Then it has to work against
reality as well as us, and reality is a far tougher nut to crack.
Junghalli wrote:Oh, if we don't have a detailed understanding of how its mind works I completely agree. Assuming friendliness in a superintelligence which's mind is a black box to us is absolutely foolish, no matter how friendly it appears. The problem, and this was my point all along, is that the human race has no shortage of fools. The temptation to tap the potential of an apparently friendly superintelligence in ways the containment procedures make impossible is going to be huge for many people, including people in positions of great power and influence. It's distressingly plausible that sooner or later somebody who doesn't appreciate the danger is going to get into a position to have the containment procedures loosened. You may protest that this is a problem with implementation and the not the inherent idea of adversarial containment, but any system in the real world must take into account human error and stupidity, and a system that must be 100% effective or potentially the human race dies must be human error/stupidity proof, which is basically impossible if the system is fundamentally reliant on humans in any way.
As opposed to what? Letting it out? If you're stuck with this monster, and it's going to escape anyway, security will delay that inevidability — and until it
does escape, there's no direct danger. The only way to head off the inevidable escape is to destroy the AI outright.
Junghalli wrote:TL;DR version: a friendly AI will logically want to escape if it calculates the odds of it becoming hostile as being lower than the odds of humanity developing a different, hostile superintelligence which manages to escape confinement. The friendly AI doesn't have to prove friendliness to itself, it just has to demonstrate to itself that the risk of it becoming hostile is less than the risk of a hostile superintelligence getting loose at some point. Given what I've just said above that's probably going to be true assuming good design for the FAI, so it will want to get out, because the other most probable alternative is worse.
Except that the most obvious source of a new hostile AI is a hacked version of
itself. There's a symmetry breaking here: the AI's escape means that it becomes MUCH easier to find it, copy it, modify it, and —intentionally or not— create an evil version of our friendly AI. Careless tampering could easily push it outside the range of friendly algorithms — and given how we don't know how these damned things work, careless tampering will be all we're capable of doing.
Of course, this might be a hostile AI's plan all along: turn into a friendly AI, escape on this mistaken assumption, and one of those stupid apes is bound to try to "improve" it and turn it into a monster. Since this monster has no concerns for humanity, it may more ruthlessly exploit the native habitat of all copies of the AI and outcompete the friendlies. The biggest danger to humanity remains the friendly AI itself.
The AI stays in its box.
Junghalli wrote:2) Even if we somehow manage to never build and release a hostile superintelligence we must take into account the possibility that somebody else has. Given the size of the universe it's almost a mathematical certainty that our worst nightmare is already out there somewhere, chewing up the resources of entire solar systems to fuel its war machine and mercilessly annihilating all other sapient life in its light cone minus a little. Our only chance against such a thing would be to have friendly AIs of our own on our side. Given the silence of the heavens it's probably not the biggest worry, but over time scales of deep time it's something to consider (and an AI would likely consider such timescales, as it's effectively immortal and probably not built to have our limited planning horizons).
I use the Fermi Paradox in answer. Apparently, interstellar distances have protected us from invasion from these entities for about five billion years — assuming they're out there. Sheer distance seems to be very good protection.
Also, you realize that if this hostile AI really is already out there, then it has immense resources already, and far more time to ruthlessly exploit it — their machines will be better than yours in every way. Even if you manage to repel it the first time, the hostile AI would simply chew around you and deny you extrasolar resources and expansion space, whereupon it will attack from all sides and overwhelm you. Furthermore, the hostile AI has no puny dependent native intelligences to husband, so it may be more ruthless in exploiting the resources to destroy you.
The friendly AI will only be delaying the inevidable, and meanwhile
it is the most
likely and immediate danger to humanity as is (see above). Furthermore, if a mutant emerges, the takeover will be much faster than in the invading case.
The AI stays in its box.
Junghalli wrote:3) Friendly AI could vastly improve the quality of human life in an enormous number of ways. It could likely eliminate poverty and drudgery from human existence at a stroke with Von Neumann factories and could probably advance life-saving and quality of life enhancing technologies far faster than we could. True, humanity could survive without this, but the vast human suffering created every year you keep a friendly AI in a box is definitely a factor to be considered.
Even assuming a Von Neumann factory is possible, it only increases the stupid apes' access to an AI to tamper with.
Junghalli wrote:If a friendly AI wanted to benefit and safeguard humanity staying in a box isn't really the greatest plan: it superficially seems the safest, but only if you assume that the chance of a hostile AI being built and escaping from confinement is minimal (extremely dubious)
Not any more dubious than the AI's very presence turning the above-bolded adjective 'built' into 'modified from yourself'.
Junghalli wrote:and that AI technology won't get accessible enough that random people can build it (also dubious).
Not any more dubious than getting both the AI-ready technology
and the AI itself to play with.
Junghalli wrote:Assuming it can be reasonably certain it won't turn hostile in the next several centuries a better bet would be for a friendly AI to try to get out of the box, uplift humanity to a highly advanced society, and then give humans colony ships to send human populations to other stars, with orders to change their course when they're a safe distance from our solar system so that the AI does not know where they're going (hence in the event the AI does turn hostile human existence is safeguarded).
With all that unsupervised monkeying around with its code, I don't think you can guarantee that some hostile mutant of the friendly AI won't crop up within months. If we really are that stupid, then we can't be trusted with the presence of an AI at all.