Subscribe
Sign in
Home
Support
LLMs From Scratch Book
Archive
About
Latest
Top
Discussions
Instruction Pretraining LLMs
The Latest Research in Instruction Finetuning
Jul 20
•
Sebastian Raschka, PhD
139
Share this post
Instruction Pretraining LLMs
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
8
June 2024
Developing an LLM: Building, Training, Finetuning
A Deep Dive into the Lifecycle of LLM Development
Jun 8
•
Sebastian Raschka, PhD
333
Share this post
Developing an LLM: Building, Training, Finetuning
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
18
LLM Research Insights: Instruction Masking and New LoRA Finetuning Experiments
Discussing the Latest Model Releases and AI Research in May 2024
Jun 2
•
Sebastian Raschka, PhD
71
Share this post
LLM Research Insights: Instruction Masking and New LoRA Finetuning Experiments
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
9
May 2024
How Good Are the Latest Open LLMs? And Is DPO Better Than PPO?
Discussing the Latest Model Releases and AI Research in April 2024
May 12
•
Sebastian Raschka, PhD
114
Share this post
How Good Are the Latest Open LLMs? And Is DPO Better Than PPO?
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
4
April 2024
Using and Finetuning Pretrained Transformers
What are the different ways to use and finetune pretrained large language models (LLMs)? The most common ways to use and finetune pretrained LLMs…
Apr 20
•
Sebastian Raschka, PhD
110
Share this post
Using and Finetuning Pretrained Transformers
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
13
March 2024
Tips for LLM Pretraining and Evaluating Reward Models
Discussing AI Research Papers in March 2024
Mar 31
•
Sebastian Raschka, PhD
130
Share this post
Tips for LLM Pretraining and Evaluating Reward Models
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
16
Research Papers in February 2024: A LoRA Successor, Small Finetuned LLMs Vs Generalist LLMs, and Transparent LLM Research
Once again, this has been an exciting month in AI research. This month, I'm covering two new openly available LLMs, insights into small finetuned LLMs…
Mar 3
•
Sebastian Raschka, PhD
78
Share this post
Research Papers in February 2024: A LoRA Successor, Small Finetuned LLMs Vs Generalist LLMs, and Transparent LLM Research
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
13
February 2024
Improving LoRA: Implementing Weight-Decomposed Low-Rank Adaptation (DoRA) from Scratch
Low-rank adaptation (LoRA) is a machine learning technique that modifies a pretrained model (for example, an LLM or vision transformer) to better suit a…
Feb 18
•
Sebastian Raschka, PhD
166
Share this post
Improving LoRA: Implementing Weight-Decomposed Low-Rank Adaptation (DoRA) from Scratch
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
37
Model Merging, Mixtures of Experts, and Towards Smaller LLMs
Model Merging, Mixtures of Experts, and Towards Smaller LLMs
Feb 3
•
Sebastian Raschka, PhD
171
Share this post
Model Merging, Mixtures of Experts, and Towards Smaller LLMs
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
15
January 2024
Understanding and Coding Self-Attention, Multi-Head Attention, Cross-Attention, and Causal-Attention in LLMs
This article will teach you about self-attention mechanisms used in transformer architectures and large language models (LLMs) such as GPT-4 and Llama…
Jan 14
273
Share this post
Understanding and Coding Self-Attention, Multi-Head Attention, Cross-Attention, and Causal-Attention in LLMs
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
27
December 2023
Ten Noteworthy AI Research Papers of 2023
This year has felt distinctly different. I've been working in, on, and with machine learning and AI for over a decade, yet I can't recall a time when…
Dec 30, 2023
•
Sebastian Raschka, PhD
344
Share this post
Ten Noteworthy AI Research Papers of 2023
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
32
Tackling Hallucinations, Boosting Reasoning Abilities, and New Insights into the Transformer Architecture
This month, I want to focus on three papers that address three distinct problem categories of Large Language Models (LLMs): Reducing hallucinations.
Dec 9, 2023
•
Sebastian Raschka, PhD
111
Share this post
Tackling Hallucinations, Boosting Reasoning Abilities, and New Insights into the Transformer Architecture
magazine.sebastianraschka.com
Copy link
Facebook
Email
Note
Other
13
Share
Copy link
Facebook
Email
Note
Other
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts