Seq1F1B: Efficient Sequence-Level Pipeline Parallelism for Large Language Model Training
June 05, 2024 Β· Entered Twilight Β· π North American Chapter of the Association for Computational Linguistics
Repo contents: .coveragerc, .github, .gitignore, .gitlab-ci.yml, CONTRIBUTING.md, LICENSE, README.md, docs, examples, exp2.sh, images, megatron, picture, pretrain_bert.py, pretrain_gpt.py, pretrain_gpt_core.py, pretrain_ict.py, pretrain_retro.py, pretrain_t5.py, pretrain_vision_classify.py, pretrain_vision_dino.py, pretrain_vision_inpaint.py, pyproject.toml, setup.py, tasks, tests, tools
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β Distributed Computing