Subscribe
Sign in
Home
Notes
LLM Gallery
Support
LLMs From Scratch Book
Reasoning From Scratch Book
Archive
About
The Big LLM Architecture Comparison
From DeepSeek-V3 to Kimi K2: A Look At Modern LLM Architecture Design
Jul 19, 2025
•
Sebastian Raschka, PhD
1,922
94
171
Most Popular
View all
Understanding Reasoning LLMs
Feb 5, 2025
•
Sebastian Raschka, PhD
1,303
46
122
Understanding and Coding Self-Attention, Multi-Head Attention, Causal-Attention, and Cross-Attention in LLMs
Jan 14, 2024
498
41
20
Understanding Large Language Models
Apr 16, 2023
•
Sebastian Raschka, PhD
960
52
51
Latest
Top
Discussions
My Workflow for Understanding LLM Architectures
A learning-oriented workflow for understanding new open-weight model releases
Apr 18
•
Sebastian Raschka, PhD
47
2
4
Components of A Coding Agent
How coding agents use tools, memory, and repo context to make LLMs work better in practice
Apr 4
•
Sebastian Raschka, PhD
767
59
76
A Visual Guide to Attention Variants in Modern LLMs
From MHA and GQA to MLA, sparse attention, and hybrid architectures
Mar 22
•
Sebastian Raschka, PhD
367
15
32
A Dream of Spring for Open-Weight LLMs: 10 Architectures from Jan-Feb 2026
A Round Up And Comparison of 10 Open-Weight LLM Releases in Spring 2026
Feb 25
•
Sebastian Raschka, PhD
205
12
20
Categories of Inference-Time Scaling for Improved LLM Reasoning
And an Overview of Recent Inference-Scaling Papers
Jan 24
•
Sebastian Raschka, PhD
41
2
The State Of LLMs 2025: Progress, Problems, and Predictions
A 2025 review of large language models, from DeepSeek R1 and RLVR to inference-time scaling, benchmarks, architectures, and predictions for 2026.
Dec 30, 2025
•
Sebastian Raschka, PhD
518
39
55
LLM Research Papers: The 2025 List (July to December)
In June, I shared a bonus article with my curated and bookmarked research paper lists to the paid subscribers who make this Substack possible.
Dec 30, 2025
•
Sebastian Raschka, PhD
36
3
See all
Ahead of AI
Ahead of AI focuses on machine learning and AI research and is read by more than 150,000 researchers and practitioners who want to stay ahead in a rapidly evolving field.
Subscribe
Ahead of AI
Subscribe
About
Archive
Recommendations
Sitemap
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts