Subscribe
Sign in
Home
Notes
Chat
Support
LLMs From Scratch Book
Reasoning From Scratch Book
Archive
About
Latest
Top
Discussions
The State Of LLMs 2025: Progress, Problems, and Predictions
A 2025 review of large language models, from DeepSeek R1 and RLVR to inference-time scaling, benchmarks, architectures, and predictions for 2026.
22 hrs ago
•
Sebastian Raschka, PhD
188
14
15
LLM Research Papers: The 2025 List (July to December)
In June, I shared a bonus article with my curated and bookmarked research paper lists to the paid subscribers who make this Substack possible.
22 hrs ago
•
Sebastian Raschka, PhD
9
From DeepSeek V3 to V3.2: Architecture, Sparse Attention, and RL Updates
Understanding How DeepSeek's Flagship Open-Weight Models Evolved
Dec 3
•
Sebastian Raschka, PhD
225
11
26
November 2025
Beyond Standard LLMs
Linear Attention Hybrids, Text Diffusion, Code World Models, and Small Recursive Transformers
Nov 4
•
Sebastian Raschka, PhD
340
24
36
October 2025
Understanding the 4 Main Approaches to LLM Evaluation (From Scratch)
Multiple-Choice Benchmarks, Verifiers, Leaderboards, and LLM Judges with Code Examples
Oct 5
•
Sebastian Raschka, PhD
343
25
32
September 2025
Understanding and Implementing Qwen3 From Scratch
A Detailed Look at One of the Leading Open-Source LLMs
Sep 6
•
Sebastian Raschka, PhD
112
6
10
August 2025
From GPT-2 to gpt-oss: Analyzing the Architectural Advances
And How They Stack Up Against Qwen3
Aug 9
•
Sebastian Raschka, PhD
611
46
54
July 2025
The Big LLM Architecture Comparison
From DeepSeek-V3 to Kimi K2: A Look At Modern LLM Architecture Design
Jul 19
•
Sebastian Raschka, PhD
1,633
78
149
LLM Research Papers: The 2025 List (January to June)
A topic-organized collection of 200+ LLM research papers from 2025
Jul 1
•
Sebastian Raschka, PhD
88
5
9
June 2025
Understanding and Coding the KV Cache in LLMs from Scratch
KV caches are one of the most critical techniques for efficient inference in LLMs in production.
Jun 17
•
Sebastian Raschka, PhD
425
37
33
May 2025
Coding LLMs from the Ground Up: A Complete Course
Why build LLMs from scratch? It's probably the best and most efficient way to learn how LLMs really work. Plus, many readers have told me they had a lot…
May 10
•
Sebastian Raschka, PhD
247
4
18
April 2025
The State of Reinforcement Learning for LLM Reasoning
Understanding GRPO and New Insights from Reasoning Model Papers
Apr 19
•
Sebastian Raschka, PhD
466
31
39
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts