Articles

Explore our latest machine learning and generative AI articles, including tutorials, news, and walkthroughs on the blog.

Article Filters

Conclusion

Current best practices for training LLMs from scratch Current best practices for training LLMs from scratch Conclusion Whether it’s OpenAI,...

< 1 min read

RLHF

REINFORCEMENT LEARNING THROUGH HUMAN FEEDBACK (RLHF) RLHFRLHF (Reinforcement Learning with Human Feedback) extends instruction tuning by incorporating human feedback after...

2 mins read

Instruction tuning

INSTRUCTION TUNING At this point, let’s assume we have a pre-trained, general-purpose LLM. If we did our job well, our...

2 mins read

Bias and toxicity

BIAS AND TOXICITY There are potential risks associated with large-scale, general-purpose language models trained on web text. Which is to...

2 mins read

Model evaluation

MODEL EVALUATION Typically, pre-trained models are evaluated on diverse language model datasets to assess their ability to perform logical reasoning,...

2 mins read

Pre-training steps

PRE-TRAINING STEPS Training a multi-billion parameter LLM is usually a highlyexperimental process with lots of trial and error. Normally, theteam...

5 mins read

Dataset pre-processing

DATASET PRE-PROCESSING In this section, we’ll cover both data adjustments (like deduplication and cleaning) and the pros and cons of...

2 mins read

Dataset collection

DATASET COLLECTION Bad data leads to bad models. But careful processing of high-quality, high-volume, diverse datasets directly contributes to model...

2 mins read

Hardware

HARDWARE It should come as no surprise that pre-training LLMs is a hardware-intensive effort. The following examples of current models...

3 mins read

Current best practices for training LLMs from scratch

Introduction Although we’re only a few years removed from the transformer breakthrough, LLMs have already grown massively in performance, cost,...

4 mins read

The scaling laws

THE SCALING LAWS Before you dive into training, it’s important to cover how LLMs scale. Understanding scaling lets you effectively...

3 mins read

What is Retrieval-Augmented Generation?

Retrieval-Augmented Generation (RAG) is a powerful technique in AI that combines large language models with real-time access to external data...

9 mins read