Training Utilities API Reference¶
free_transformer.train_utils
¶
Training utilities including FSDP and DeepSpeed support.
LRScheduler(optimizer, warmup_steps, max_lr, min_lr=0.0)
¶
Learning rate scheduler with warmup.
Source code in src/free_transformer/train_utils.py
step()
¶
Update learning rate.
Source code in src/free_transformer/train_utils.py
Trainer(model, optimizer, device, use_amp=False)
¶
Base trainer class.
Source code in src/free_transformer/train_utils.py
eval_step(batch)
¶
train_step(batch, grad_clip=None)
¶
Single training step.
Source code in src/free_transformer/train_utils.py
cleanup_distributed()
¶
count_parameters(model)
¶
load_checkpoint(model, optimizer, path)
¶
Load training checkpoint with better error handling.
Source code in src/free_transformer/train_utils.py
save_checkpoint(model, optimizer, step, path, metadata=None)
¶
Save training checkpoint.
Source code in src/free_transformer/train_utils.py
setup_distributed()
¶
Initialize distributed training.
Source code in src/free_transformer/train_utils.py
wrap_model_fsdp(model, mixed_precision=True, min_num_params=1000000)
¶
Wrap model with FSDP for distributed training.