> alignment problems don’t disappear just because the agents can negotiate with each other.
I feel that current agents weren't trained to do that; Once this becomes economically feasible to build trader agents for yourself, labs will RL those AIs to be more effective traders
What if the experiment also involved "evolution", where different randomized prompts are given, and agents can self-modify themselves, and ineffective agents are "fired" and replaced with more effective?
Or maybe I didn't fully understand the setup
Dec 9
at
12:42 AM
Relevant people
Log in or sign up
Join the most interesting and insightful discussions.