Skip to content
This repository has been archived by the owner on Mar 8, 2024. It is now read-only.

Lightning-Universe/lightning-Galvatron

Repository files navigation

Lightning integration with Galvatron

Galvatron is a new system framework aiming at efficient Transformer training over multiple GPUs using automatic parallelism.

lightning PyPI - Version CI testing General checks Documentation Status pre-commit.ci status

It incorporates multiple popular parallelism dimensions (including data parallel, sharded data parallel, tensor parallel and pipeline parallel) and automatically finds the most efficient hybrid parallelism strategy.

Galvatron can be configured in the training script by specifying strategy arguments as follows:

from lightning import Trainer
from lightning_galvatron import GalvatronStrategy

trainer = Trainer(
    strategy=GalvatronStrategy(
        model_type="gpt",
        model_size="gpt-1.5b",
        pp_deg=2,
        global_tp_deg=2,
        fsdp=1
    ),
    accelerator="gpu",
    devices=4
)

Please see the official Galvatron repository for more model support and advanced features.

Cite

If you use Galvatron in a scientific publication, we would appreciate citations to the following paper:

@article{miao2023galvatron,
  title = {Galvatron: Efficient Transformer Training over Multiple GPUs Using Automatic Parallelism},
  author = {Miao, Xupeng and Wang, Yujie and Jiang, Youhe and Shi, Chunan and Nie, Xiaonan and Zhang, Hailin and Cui, Bin},
  journal = {Proc. {VLDB} Endow.},
  volume = {16},
  number = {3},
  pages = {470--479},
  year = {2023},
  doi = {10.14778/3570690.3570697},
  publisher = {VLDB Endowment},
}

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published