But whose human preferences are we training on? A generic thumbs up? A vibes-based "which response was better"?
What if the reward function wasn't flattened human preference, but a comprehensive model of you? Your cognitive architecture. Your values. Your actual life.
That's what a Life Model enables. Not better AI for everyone. Better AI for each person specifically. Reinforcement learning where the signal isn't "good response." It's "good response for this mind."
Scale that across thousands of unique models and you're not fine-tuning a product. You're building infrastructure that gets smarter about people.
That’s what I’m excited about!
Feb 27
at
10:00 PM
Relevant people
Log in or sign up
Join the most interesting and insightful discussions.