How-to Guides#

DIV Onclick Test

Write a Custom Training Script

Learn how to write a training script for your model from scratch using the Cerebras PyTorch API for training on the MNIST dataset

Train an LLM using Maximal Update Parameterization

Learn how to enable μP when training models on the Cerebras Wafer-Scale cluster

Train an LLM with a large or small context window

Learn how to use the CS-2 to train an LLM with a large or small context window


Fine-tune an LLM on a dataset using instructions

Learn how to fine-tune LLMs on datasets with instructions and corresponding responses

Train a model with weight sparsity

Learn how to train a model with weight sparsity to achieve a sparse model that requires fewer FLOPs to train and fewer parameters to store

Restart a dataloader

Learn how to resume training from the same point in the input-generating dataloader


Port a trained and fine-tuned model to Hugging Face

Learn how to port an LLM model trained in the Cerebras’s Wafer-Scale Cluster to Hugging Face to generate outputs

Control numerical precision level

Learn how to control the level of numerical precision used for training runs for large NLP models

Run Cerebras Model Zoo on a GPU

Learn how to run models in the Cerebras Model Zoo on GPUs and which packages to install