artificial
artificial

[N] How Deepseek trained their R1 models, and how frontier LLMs are trained today

https://www.youtube.com/watch?v=aAfanTeRn84 Lex Friedman recently posted an interview called "DeepSeek's GPU Optimization tricks". It is a great behind the scenes look at how Deepseek trained their latest models even when they did not hav…

Is ChatGPT a better judge of probability than doctors? – discussing case studies vs RCTs as reliable indicators of efficacy – Can case studies with few data points but high efficacy outperform "gold standard" large RCTs with anemic results?

submitted by /u/stereomatch [link] [comments]

Self-MoA: Single-Model Ensembling Outperforms Multi-Model Mixing in Large Language Models

This work investigates whether mixing different LLMs actually improves performance compared to using single models – and finds some counterintuitive results that challenge common assumptions in the field. The key technical elements: – Systematic evalua…