Asi Alignment Tier List
S
A
Recruitment & Education
Best Memes = AI Notkilleveryoneism Memes
General Ai Safety FAQ = aiSafety.info (Rob Miles)
Ai Safety Map = aiSafety.world
Ai Safety Beginner's Guide = aiSafetyLinkTree
Ai Safety Fundraising
How to Receive Funding for an Ai Safety project:
Navigate to the top right section of aisafety.world
---------------------------------------------------------------------
How to Fundraise for AI Safety: More Info coming soon.
Ai Regulations
1. Blackbox Algorithmic Transparancy
2. Data Collection & Usage
3. Human Extinction Safety Standards
4. Economic Impact & Universal Basic Income
5. Ai Capability restrictions
More Info coming soon.
B
"Ai will solve Ai Alignment"
This ^^^ could work , but there are huge issues with it and it's very far from a foolproof solution
AsiATL's Criticism of "Ai Solves Asi Alignment"
----------------------- Projects working on this approach -----------------------
ELK (Eliciting Latent Knowledge)(Paul Christiano): "To produce a minimal AI that can help to do AI safety research."
Mechanistic Interpretability
Mechanistic interpretability is the pursuit to understand the inner workings of black box Ai such as LLM's or End-to-End Reinforcement Learning systems.
Basic Overview 2024
Advanced Overview 2024
Cognitive Emulation (Connor Leahy)
Trying to build bounded understandable systems that emulate human-like reasoning. When you use the system, at the end you get a causal story, an explanation you can understand using human-like reasoning why the system did what it did and why you should trust the output to be valid.
Connor Leahy Explaining "CoEm" 2023
Connor Leahy Explaining "CoEm" 2024
C
Corrigibility
Is an approach to AI alignment that involves designing an AI system that is willing to be corrected or shut down by humans when it is uncertain about its own goals or the actions it should take. More Info coming soon.
D
E
Synthetic Brain Clone
The idea is to create a fullly transparant and understandable clone of a human bain in computer form. Ben Goertzel among others talk about this. More Info coming soon.
Advanced Human Cyborgs (Brain Machine Interfaces)
The basic idea is to radically enhance human brain capabilities by connecting them to computational power, then have the new cyborgs solve AI Alignment. Neuralink is the most well known project working on BMI, but there are a handful of companies including but not limited to Kernel, CTRL-Labs, Blackrock Microsystems. More Info coming soon.
F
Maximum Truth Seeking (Elon Musk)
There's not much info from elon about this yet, will update as new info comes in. He seems to be building a new project to try to rival OpenAI and Google using this Maximum Truth Seeking idea. More Info coming soon.
RLHF (Reinforcement Learning with Human Feedback)
Sam Altman direct quote.
chatGPT is the poster child for RLHF.
Reinforcement learning with human feedback involves using a learning algorithm that interacts with an environment and receives feedback signals from a human, typically in the form of rewards or penalties, to guide its learning process. The algorithm uses this feedback to learn how to make better decisions in the environment over time. More Info coming soon.
"Just unplug it"
*starts fortnite dancing*
"The risk is negligible"
"Trust me bro"
Accelerated Global Nuclear Winter
I hope this is not the best solution to stop AI Capability Accelaration. More Info coming soon.
Mesaoptimisers
A mesaoptimizer is an AI system that is capable of optimizing its own objectives, including the ability to modify its own code or architecture. It may arise as a result of training an AI system using an optimization algorithm that incentivizes the development of sub-agents or sub-routines that themselves exhibit optimization behavior.
Heuristic Imperatives (Ben Shapiro)
This solution claims you can use natural language (english) to tell LLM's 1. Reduce suffering in the universe. 2. Increase Prosperity in the universe. 3. Increase understanding in the universe. Check reddit r/HeuristicImperatives or this video for more info.
OpenCog Hyperon (Ben Goertzel)
Hyperon is based on a knowledge representation and reasoning system that uses a probabilistic logic programming language called Probabilistic Logic Networks (PLN). PLN combines first-order logic with probability theory to enable reasoning with uncertainty. More Info coming soon.
Don't understand enough to comment
Robust Adversarial Planning
No Comment
AI Safety Gridworlds
No Comment
Iterated Amplification
No Comment
Cooperative Inverse Reinforcement Learning (CIRL)
No Comment
Scalable Agent Alignment via Reward Compression
No Comment