Subscribe
Sign in
Home
Support
LLMs From Scratch Book
Archive
About
Latest
Top
Discussions
Building LLMs from the Ground Up: A 3-hour Coding Workshop
If your weekend plans include catching up on AI developments and understanding Large Language Models (LLMs), I've prepared a 1-hour presentation on the…
Aug 31
•
Sebastian Raschka, PhD
337
Share this post
Building LLMs from the Ground Up: A 3-hour Coding Workshop
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
13
New LLM Pre-training and Post-training Paradigms
A Look at How Moderns LLMs Are Trained
Aug 17
•
Sebastian Raschka, PhD
253
Share this post
New LLM Pre-training and Post-training Paradigms
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
20
July 2024
Instruction Pretraining LLMs
The Latest Research in Instruction Finetuning
Jul 20
•
Sebastian Raschka, PhD
170
Share this post
Instruction Pretraining LLMs
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
13
June 2024
Developing an LLM: Building, Training, Finetuning
A Deep Dive into the Lifecycle of LLM Development
Jun 8
•
Sebastian Raschka, PhD
344
Share this post
Developing an LLM: Building, Training, Finetuning
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
18
LLM Research Insights: Instruction Masking and New LoRA Finetuning Experiments
Discussing the Latest Model Releases and AI Research in May 2024
Jun 2
•
Sebastian Raschka, PhD
71
Share this post
LLM Research Insights: Instruction Masking and New LoRA Finetuning Experiments
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
9
May 2024
How Good Are the Latest Open LLMs? And Is DPO Better Than PPO?
Discussing the Latest Model Releases and AI Research in April 2024
May 12
•
Sebastian Raschka, PhD
117
Share this post
How Good Are the Latest Open LLMs? And Is DPO Better Than PPO?
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
4
April 2024
Using and Finetuning Pretrained Transformers
What are the different ways to use and finetune pretrained large language models (LLMs)? The most common ways to use and finetune pretrained LLMs…
Apr 20
•
Sebastian Raschka, PhD
112
Share this post
Using and Finetuning Pretrained Transformers
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
13
March 2024
Tips for LLM Pretraining and Evaluating Reward Models
Discussing AI Research Papers in March 2024
Mar 31
•
Sebastian Raschka, PhD
133
Share this post
Tips for LLM Pretraining and Evaluating Reward Models
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
16
Research Papers in February 2024: A LoRA Successor, Small Finetuned LLMs Vs Generalist LLMs, and Transparent LLM Research
Once again, this has been an exciting month in AI research. This month, I'm covering two new openly available LLMs, insights into small finetuned LLMs…
Mar 3
•
Sebastian Raschka, PhD
79
Share this post
Research Papers in February 2024: A LoRA Successor, Small Finetuned LLMs Vs Generalist LLMs, and Transparent LLM Research
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
13
February 2024
Improving LoRA: Implementing Weight-Decomposed Low-Rank Adaptation (DoRA) from Scratch
Low-rank adaptation (LoRA) is a machine learning technique that modifies a pretrained model (for example, an LLM or vision transformer) to better suit a…
Feb 18
•
Sebastian Raschka, PhD
168
Share this post
Improving LoRA: Implementing Weight-Decomposed Low-Rank Adaptation (DoRA) from Scratch
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
37
Model Merging, Mixtures of Experts, and Towards Smaller LLMs
Model Merging, Mixtures of Experts, and Towards Smaller LLMs
Feb 3
•
Sebastian Raschka, PhD
176
Share this post
Model Merging, Mixtures of Experts, and Towards Smaller LLMs
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
15
January 2024
Understanding and Coding Self-Attention, Multi-Head Attention, Cross-Attention, and Causal-Attention in LLMs
This article will teach you about self-attention mechanisms used in transformer architectures and large language models (LLMs) such as GPT-4 and Llama.
Jan 14
284
Share this post
Understanding and Coding Self-Attention, Multi-Head Attention, Cross-Attention, and Causal-Attention in LLMs
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
33
Share
Copy link
Facebook
Email
Note
Other
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts