Shaw TalebiHow to Train LLMs to “Think” (o1 & DeepSeek-R1)Advanced reasoning models explainedFeb 123Feb 123
Alberto RomeroDebunking 10 Popular Myths About DeepSeekI’ve heard too many wrong takes that need correctionFeb 414Feb 414
InThe GeneratorbyThomas SmithWhy DeepSeek Doesn’t MatterThe new AI model is a hack, not an “extinction level event.” Here’s why.Jan 2957Jan 2957
InGenerative AIbyCezary GesikowskiDeep Research in ChatGPT–AI Agent for In-Depth Knowledge WorkThe days of million active browsers tabs are numberedFeb 31Feb 31
Jan KammerathDeepSeek: Is It A Stolen ChatGPT?While I was drowning in emails, fiddling around with Xcode and the Neural Cores in my MacBook, DeepSeek popped up on X and Reddit. It…Jan 2752Jan 2752
InLevel Up CodingbyDr. Ashish BamaniaDeepSeek-R1 Beats OpenAI’s o1, Revealing All Its Training Secrets Out In The OpenA deep dive into how DeepSeek-R1 was trained from scratch and how this open-source research will accelerate AI progress like never before.Jan 2732Jan 2732
Alberto RomeroDeepSeek Is Chinese But Its AI Models Are From Another PlanetOpenAI and the US are in deep troubleJan 22193Jan 22193
InTowards AIbyYash ThubeDeepSeek-R1: The Open-Source AI That Thinks Like OpenAI’s BestWhat’s the hype about?👀Jan 213Jan 213
Alberto RomeroThis Rumor About GPT-5 Changes EverythingLet’s start the year on an exciting noteJan 1630Jan 1630
InLevel Up CodingbyFareed KhanBuilding a 2 Billion Parameter LLM from Scratch Using PythonIt starts making senseJan 1520Jan 1520
InAI MindbyMr Tony MomohThe $6 Million AI That’s Making OpenAI Nervous: How a Tiny Chinese Startup Is Disrupting Silicon…In the world of artificial intelligence, David just threw a stone at Goliath. And this time, David spent 99% less money.Jan 228Jan 228
Rendy DalimuntheUnderstanding Test-Time Compute: A New Mechanism Allowing AI to “Think Harder”Exploring How AI Adapts to Complex Tasks with Dynamic Reasoning PowerNov 22, 2024Nov 22, 2024
InAI AdvancesbyTim Urista | Senior Cloud EngineerDramatically Reduce Inference Costs with DeepSeek-V3: A New Era in Open-Source LLMsIntroductionDec 27, 20241Dec 27, 20241
InAIGuysbyVishal RajputOpenAI Achieved AGI With Its New o3 model?Does adding extensive search during inference means better generalization.Dec 23, 20242Dec 23, 20242
InAIGuysbyVishal RajputAre Tiny Transformers The Future Of Scaling?Can we reduce the O(N)² of Attnetion Mechanism?Nov 12, 20241Nov 12, 20241
DevanshLLMs are NOT reaching their limits.A response to Gary Marcus and many other “AI skeptics”Nov 10, 20249Nov 10, 20249
InGenerative AIbyYuki ShizuyaUnlocking Mixture-of-Experts (MoE) LLM : Your MoE model can be embedding model for freeMixture-of-experts (MoE) LLM can be used as an embedding model for free.Nov 3, 20245Nov 3, 20245
Don LimReasoning tokens and techniques used in System 2 LLMs such as OpenAI o1What is the System 2 model?Sep 16, 20242Sep 16, 20242
InSyncedReviewbySyncedStanford’s Landmark Study: AI-Generated Ideas Rated More Novel Than Expert ConceptsRecent advancements in large language models (LLMs) have generated enthusiasm about their potential to accelerate scientific innovation…Sep 18, 20241Sep 18, 20241