We are entering the era where LLMs are being trained on data generated by other LLMs. I’m starting to see "semantic collapse" in some of the smaller models.
In our internal testing, reasoning capabilities for edge-case logic are stagnating because the diversity of the training set is shrinking. I believe the only way out is to prioritize "Sovereign Human Data"—high-quality, non-public human reasoning logs. This is why private, secure environments for AI interaction are becoming more valuable than the models themselves. Thoughts?
[link] [comments]