<span class="vcard">/u/theMonarch776</span>
/u/theMonarch776

Why Groq? Why?

groq Speed Insights Tokens 507 – Input tokens 37 – Output tokens 544 – Tokens Inference Time 0.04 – Input seconds 0.09 – Output seconds 0.13 – Seconds Tokens / second 12434 – Input 409 – Output 4145 – Total Round trip time: 0.84s Model: groq/compound-m…

Evolution Tree of Encoder / Decoder / Encoder-Decoder Models.

The following Image was shared on X by Yan LeCun.. The architecture nomenclature for LLMs is somewhat confusing and unfortunate. What's called "encoder only" actually has an encoder and a decoder (just not an auto-regressive decoder…

Is this the End of Epochs?

1960s: "COBOL will let non-programmers make the software!" 1980s: "4GLs will let non-programmers make the software!" 2000s: "UML will let non-programmers make the software!" 2020s: "Al will let non-programmers make th…

Will AI give better answer when you threaten it ?

Old news, but wild enough to resurface. Google co-founder Sergey Brin once said on the All-In podcast that Al models (including Google's Gemini) actually perform better when you threaten them. "Not just our models, but all models tend to…