rw-book-cover

Metadata

Highlights

  • Lately, we have been using Lightning Fabric, which brings together: • The flexibility of PyTorch. • And distributed training features that PyTorch Lightning provides. You only need to make 4 minor changes to your existing PyTorch code to easily scale it to the largest billion-parameter models/LLMs. (View Highlight)
  • While creating the Fabric object above, you can specify: • the accelerator and the number of devices • the parallelism strategy to use • the floating point precision, etc. Next, configure the model, the optimizer, and the dataloader (View Highlight)