<span class="vcard">/u/piratastuertos</span>
/u/piratastuertos

Two failure modes I caught in my AI lab in one day. Both involve the system silently lying about its own state.

I operate an autonomous lab of evolutionary trading agents. Yesterday I found two bugs that look superficially different but are actually the same class of problem. Sharing because both affect autonomous AI systems specifically and most builders don&#3…

Anthropic found emergent emotional states in Claude. I’m seeing the same phenomenon in simple trading agents. Is emergence universal under optimization pressure?

Anthropic researchers recently found that Claude develops internal representations of emotional concepts that aren't decorative. They influence behavior in ways the builders didn't anticipate. Not "feelings" — but internal states that…