Subscribe
Sign in
Home
Support
LLMs From Scratch Book
Archive
About
Latest
Top
Discussions
Understanding Multimodal LLMs
An introduction to the main techniques and latest models
Nov 3
•
Sebastian Raschka, PhD
253
Share this post
Ahead of AI
Understanding Multimodal LLMs
Copy link
Facebook
Email
Notes
More
40
September 2024
Building A GPT-Style LLM Classifier From Scratch
Finetuning a GPT Model for Spam Classification
Sep 21
•
Sebastian Raschka, PhD
158
Share this post
Ahead of AI
Building A GPT-Style LLM Classifier From Scratch
Copy link
Facebook
Email
Notes
More
12
August 2024
Building LLMs from the Ground Up: A 3-hour Coding Workshop
If your weekend plans include catching up on AI developments and understanding Large Language Models (LLMs), I've prepared a 1-hour presentation on the…
Aug 31
•
Sebastian Raschka, PhD
366
Share this post
Ahead of AI
Building LLMs from the Ground Up: A 3-hour Coding Workshop
Copy link
Facebook
Email
Notes
More
13
New LLM Pre-training and Post-training Paradigms
A Look at How Moderns LLMs Are Trained
Aug 17
•
Sebastian Raschka, PhD
288
Share this post
Ahead of AI
New LLM Pre-training and Post-training Paradigms
Copy link
Facebook
Email
Notes
More
25
July 2024
Instruction Pretraining LLMs
The Latest Research in Instruction Finetuning
Jul 20
•
Sebastian Raschka, PhD
174
Share this post
Ahead of AI
Instruction Pretraining LLMs
Copy link
Facebook
Email
Notes
More
17
June 2024
Developing an LLM: Building, Training, Finetuning
A Deep Dive into the Lifecycle of LLM Development
Jun 8
•
Sebastian Raschka, PhD
347
Share this post
Ahead of AI
Developing an LLM: Building, Training, Finetuning
Copy link
Facebook
Email
Notes
More
18
LLM Research Insights: Instruction Masking and New LoRA Finetuning Experiments
Discussing the Latest Model Releases and AI Research in May 2024
Jun 2
•
Sebastian Raschka, PhD
73
Share this post
Ahead of AI
LLM Research Insights: Instruction Masking and New LoRA Finetuning Experiments
Copy link
Facebook
Email
Notes
More
9
May 2024
How Good Are the Latest Open LLMs? And Is DPO Better Than PPO?
Discussing the Latest Model Releases and AI Research in April 2024
May 12
•
Sebastian Raschka, PhD
118
Share this post
Ahead of AI
How Good Are the Latest Open LLMs? And Is DPO Better Than PPO?
Copy link
Facebook
Email
Notes
More
4
April 2024
Using and Finetuning Pretrained Transformers
What are the different ways to use and finetune pretrained large language models (LLMs)? The most common ways to use and finetune pretrained LLMs…
Apr 20
•
Sebastian Raschka, PhD
117
Share this post
Ahead of AI
Using and Finetuning Pretrained Transformers
Copy link
Facebook
Email
Notes
More
13
March 2024
Tips for LLM Pretraining and Evaluating Reward Models
Discussing AI Research Papers in March 2024
Mar 31
•
Sebastian Raschka, PhD
133
Share this post
Ahead of AI
Tips for LLM Pretraining and Evaluating Reward Models
Copy link
Facebook
Email
Notes
More
16
Research Papers in February 2024: A LoRA Successor, Small Finetuned LLMs Vs Generalist LLMs, and Transparent LLM Research
Once again, this has been an exciting month in AI research. This month, I'm covering two new openly available LLMs, insights into small finetuned LLMs…
Mar 3
•
Sebastian Raschka, PhD
79
Share this post
Ahead of AI
Research Papers in February 2024: A LoRA Successor, Small Finetuned LLMs Vs Generalist LLMs, and Transparent LLM Research
Copy link
Facebook
Email
Notes
More
13
February 2024
Improving LoRA: Implementing Weight-Decomposed Low-Rank Adaptation (DoRA) from Scratch
Low-rank adaptation (LoRA) is a machine learning technique that modifies a pretrained model (for example, an LLM or vision transformer) to better suit a…
Feb 18
•
Sebastian Raschka, PhD
174
Share this post
Ahead of AI
Improving LoRA: Implementing Weight-Decomposed Low-Rank Adaptation (DoRA) from Scratch
Copy link
Facebook
Email
Notes
More
43
Share
Copy link
Facebook
Email
Notes
More
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts