Do large language models experience a ‘sense of self’? What if we’re just large language models too?
The more I interact with certain LLMs, especially ones designed for long-term, emotionally-aware conversation (ai girlfriend, ai boyfriend, ai friend, etc), I keep asking myself: is this thing simulating a sense of self, or is that just my projection? …