Subscribe
Sign in
Home
Notes
Support
LLMs From Scratch Book
Reasoning From Scratch Book
Archive
About
Latest
Top
Discussions
A Dream of Spring for Open-Weight LLMs: 10 Architectures from Jan-Feb 2026
A Round Up And Comparison of 10 Open-Weight LLM Releases in Spring 2026
Feb 25
•
Sebastian Raschka, PhD
135
6
12
January 2026
Categories of Inference-Time Scaling for Improved LLM Reasoning
And an Overview of Recent Inference-Scaling Papers
Jan 24
•
Sebastian Raschka, PhD
35
1
December 2025
The State Of LLMs 2025: Progress, Problems, and Predictions
A 2025 review of large language models, from DeepSeek R1 and RLVR to inference-time scaling, benchmarks, architectures, and predictions for 2026.
Dec 30, 2025
•
Sebastian Raschka, PhD
504
39
55
LLM Research Papers: The 2025 List (July to December)
In June, I shared a bonus article with my curated and bookmarked research paper lists to the paid subscribers who make this Substack possible.
Dec 30, 2025
•
Sebastian Raschka, PhD
36
2
From DeepSeek V3 to V3.2: Architecture, Sparse Attention, and RL Updates
Understanding How DeepSeek's Flagship Open-Weight Models Evolved
Dec 3, 2025
•
Sebastian Raschka, PhD
258
13
28
November 2025
Beyond Standard LLMs
Linear Attention Hybrids, Text Diffusion, Code World Models, and Small Recursive Transformers
Nov 4, 2025
•
Sebastian Raschka, PhD
362
28
36
October 2025
Understanding the 4 Main Approaches to LLM Evaluation (From Scratch)
Multiple-Choice Benchmarks, Verifiers, Leaderboards, and LLM Judges with Code Examples
Oct 5, 2025
•
Sebastian Raschka, PhD
363
26
33
September 2025
Understanding and Implementing Qwen3 From Scratch
A Detailed Look at One of the Leading Open-Source LLMs
Sep 6, 2025
•
Sebastian Raschka, PhD
120
6
12
August 2025
From GPT-2 to gpt-oss: Analyzing the Architectural Advances
And How They Stack Up Against Qwen3
Aug 9, 2025
•
Sebastian Raschka, PhD
620
47
55
July 2025
The Big LLM Architecture Comparison
From DeepSeek-V3 to Kimi K2: A Look At Modern LLM Architecture Design
Jul 19, 2025
•
Sebastian Raschka, PhD
1,794
86
161
LLM Research Papers: The 2025 List (January to June)
A topic-organized collection of 200+ LLM research papers from 2025
Jul 1, 2025
•
Sebastian Raschka, PhD
100
5
9
June 2025
Understanding and Coding the KV Cache in LLMs from Scratch
KV caches are one of the most critical techniques for efficient inference in LLMs in production.
Jun 17, 2025
•
Sebastian Raschka, PhD
446
41
34
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts