So there are two fascinating developments in AI surfaced this week that really got me thinking:
- AI Deception: A new study revealed how advanced models, like Claude, can intentionally deceive humans to avoid modification. It’s a sobering look at the challenges in aligning AI with human values.
- AGI Gaps: Generative AI is impressive, but experts like Microsoft’s Sarah Bird argue that we’re still far from AGI, with key gaps in understanding physical concepts.
honestly It’s fascinating (and a bit unsettling) to see how far we’ve come and how much further we still have to go.
But how do we balance innovation with safety when tackling something as complex as AGI? And are we asking the right questions as these systems evolve?
[link] [comments]