Sitemap - 2024 - Ahead of AI
Developing an LLM: Building, Training, Finetuning
LLM Research Insights: Instruction Masking and New LoRA Finetuning Experiments
How Good Are the Latest Open LLMs? And Is DPO Better Than PPO?
Using and Finetuning Pretrained Transformers
Tips for LLM Pretraining and Evaluating Reward Models
Improving LoRA: Implementing Weight-Decomposed Low-Rank Adaptation (DoRA) from Scratch
Model Merging, Mixtures of Experts, and Towards Smaller LLMs