We were modelling the ultimate best future (billions of years from now) for 3+ years, Xerox PARC-style, we got very exciting results. Including AI safety results
Here is a drafty continuation you can find interesting (or not ;):
In unreasonable times the solution to AI problem will sounds unreasonable at first. Even though it's probably the only reasonable and working solution.
Imagine in a year we solved alignment and even hackers/rogue states cannot unleash AI agents on us. How we did it?
And NVIDIA will be happy to have this world, because it will double her business, NVIDIA will be able to replace all the GPUs: so people will bring theirs and get some money for it, then they buy new non-agentic sandboxed GPU with an updatable whitelist (probably to use gpus you'll need internet connection from now on, especially if you didn't update the whitelist of AI models for more than a few days).
And NVIDIA will be able to take up to 15-30% commission from the paid AI model providers (like OpenAI). Smaller developers will make models, they will be registered in a stricter fashion than in Apple's App Store, in a similar fashion to Nintendo developers. Basically we'll want to know they are good people and won't run evil AI models or agents while pretending they are developing something benign. .. So we need just to spread the world and especially convince the politicians of the dangers and of this solution: that we just need to make GPU makers the gatekeepers who have skin in the game to keep all the AI models safe.
We'll give deadlines to GPU owners, first we'll update their GPUs with blacklists and whitelists. There will be a deadline to replace GPUs, else the old ones will stop working (will be remotely bricked, all OSes and AI tools will have a list of those bricked GPUs and will refuse to work with them) and law enforcement will take possession of them.
This way we'll sanitize our world from insecure unsafe GPUs we have now. Only whitelisted models will run inside of the sandboxed GPU and will only spit out safe text or picture output.
Having a few GPU companies to control is much easier than having infinitely many insecure unsafe GPUs with hackers, military and rogue states everywhere.
At least we can have politicians (in order to make defense and national security better) make NVIDIA and other GPU manufacturers sell those non-agentic GPUs to foreign countries, so there will be bigger and bigger % of non-agentic (or it can be some very limited agency if math proven safe) GPUs that are mathematically proven to be safe. Same way we try to make fewer countries have nuclear weapons, we can replace their GPUs (their "nukes", their potentially uncontrollable and autonomous weapons) with safe non-agentic GPUs (=conventional non-military civilian tech)
Yes, David, I should start recommending GPUs with internal antiviruses because they have lists + heuristics, they work.
We better to have GPUs that are protected right now, before things go bad.
Even if we don’t know how to make ideal 100% perfect anti-bad-AI antiviruses on GPU-level, it’s better than 100% unprotected GPUs we have now. It’ll deter some threats, slow down hackers/agentic AI takeover.
It can be a start that we’ll build upon as we better understand the threats
A major problem with the setup of our forum: currently it's possible to go here or to any post and double downvote all the new posts in mass (please, don't! Just believe me it's possible), so writers who were writing their AI safety proposal for months will have their posts ruined (almost no one reads posts with negative rating), will probably abandon our cause and go become evil masterminds.
Solution: UI proposal to solve the problem of demotivating writers, helps to teach writers how to improve their posts (so it makes all the posts better), it keeps the downvote buttons, increases signal to noise ratio on the site because both authors and readers will have information why the post was downvoted:
Thank you for reading and making the website work!
The only complete and comprehensive solution that makes AIs 100% safe: in a nutshell we need to at least lobby politicians to make GPU manufacturers (NVIDIA and others) to make robust blacklists of bad AI models, update GPU firmwares with them. It's not the full solution: please steelman and read the rest to learn how to make it much safer and why it will work (NVIDIA and other GPU makers will want to do it because it'll double their business and all future cash flows. Gov will want it because it removes all AI threats from China, all hackers, terrorists and rogue states):
If we go extinct, it doesn't matter how much value we get. We don't exist to appreciate it.
If we don't go extinct, it probably means we have enough "value" (we survived, it means we had and have food and shelter) and probably we can have math proofs how to make AI agents safe. Now after the main extinction event (probably AI agentic explosion) is in the past, we can work on increasing the value of futures.
The summary is not great, the main idea is this: we have 3 “worlds” - physical, online, and AI agents’ multimodal “brains” as the third world. We can only easily access the physical world, we are slower than AI agents online and we cannot access multimodal “brains” at all, they are often owned by private companies.
While AI agents can access and change all the 3 “worlds” more and more.
We need to level the playing field by making all the 3 worlds easy for us to access and democratically change, by exposing the online world and especially the multimodal “brains” world as game-like 3D environments for people to train and get at least the same and ideally more freedoms and capabilities than AI agents have.
Not bad for a summary, it’s important to ensure human freedoms grow faster than restrictions on us. Human and agentic AI freedoms can be counted, we don’t want to have fewer freedoms than agentic AIs (it’s the definition of a dystopia) but sadly we are already ~10% there and continue falling right into a dystopia
I’ll be happy to answer any questions 🫶🏻 We started to discuss the project to some extent already but why not to discuss it here.
P.S. Should I change the title or tags? How? I'm very bad at those things
Yes, the only realistic and planet-wide 100% safe solution is this: putting all the GPUs in safe cloud/s controlled by international scientists that only make math-proven safe AIs and only stream output to users.
Each user can use his GPU for free from the cloud on any device (even on phone), when the user doesn't use it, he can choose to earn money by letting others use his GPU.
You can do everything you do now, even buy or rent GPUs, all of them just will be cloud math-proven safe GPUs instead of physical. Because GPUs are nukes are we want no nukes or to put them deep underground in one place so they can be controlled by international scientists.
Computer viruses we still didn't 100% solve (my mom had an Android virus recently), even iPhone and Nintendo Switch got jailbroken almost instantly, there are companies jailbreak iPhones as a service. I think Google Docs never got jailbroken, and majorly hacked, it's a cloud service, so we need to base our AI and GPU security on this best example, we need to have all our GPUs in an internationally scientist controlled cloud.
Else we'll have any hacker write a virus (just to steal money) with an AI agent component, grab consumer GPUs like cup-cakes, AI agent can even become autonomous (and we know they become evil in major ways, want to have a tea party with Stalin and Hitler - there was a recent paper - if given an evil goal. Will anyone align AIs for hackers or hacker themself will do it perfectly (they won't) to make an AI agent just to steal money but be a slave and do nothing else bad?)