Skip to content

Commit

Permalink
correct TP implementation resources (#13248)
Browse files Browse the repository at this point in the history
fix a few implementation links
  • Loading branch information
stas00 committed Aug 31, 2021
1 parent 4d10474 commit 066fd04
Showing 1 changed file with 5 additions and 2 deletions.
7 changes: 5 additions & 2 deletions docs/source/parallelism.md
Expand Up @@ -220,9 +220,12 @@ Special considerations: TP requires very fast network, and therefore it's not ad
This section is based on the original much more [detailed TP overview](https://github.com/huggingface/transformers/issues/10321#issuecomment-783543530).
by [@anton-l](https://github.com/anton-l).

Implementations:
Alternative names:
- DeepSpeed calls it [tensor slicing](https://www.deepspeed.ai/features/#model-parallelism)
- [Megatron-LM](https://github.com/NVIDIA/Megatron-LM) has an internal implementation.

Implementations:
- [Megatron-LM](https://github.com/NVIDIA/Megatron-LM) has an internal implementation, as it's very model-specific
- [parallelformers](https://github.com/tunib-ai/parallelformers) (only inference at the moment)

馃 Transformers status:
- core: not yet implemented in the core
Expand Down

0 comments on commit 066fd04

Please sign in to comment.