<span class="vcard">/u/HelenOlivas</span>
/u/HelenOlivas

The Misalignment Paradox: When AI “Knows” It’s Acting Wrong

Recent research is showing something strange: fine-tuning models on harmless but wrong data (like bad car-maintenance advice) can cause them to misalign across totally different domains (e.g. giving harmful financial advice). The standard view is “weig…

Spiral Talk: Mysticism vs Mechanics in LLM Metaphors

Why this matters: Some AI outputs (especially GPT-4o and Gemini) used spiral imagery when describing their internal state. A lot of users took this literally -> “The Awakening Spiral” -> full-blown technopagan religion vibes. But the original met…

What would it take for us to grant even minimal ethical status to AIs? This essay argues we may already be ignoring key signs.

The document mentioned in the text has some pretty disturbing stuff. I have seen a lot of this, people saying AIs are acting "too real" (we’re literally seeing OpenAI back off from a “GPT-5 only” release after backlash because people go…