From 800ea8d39fc6e238d0ad8115e0219be2b256e045 Mon Sep 17 00:00:00 2001 From: gitnlp <36983436+gitnlp@users.noreply.github.com> Date: Sun, 27 Nov 2022 22:45:31 +0800 Subject: [PATCH] Update README.md --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 1165757..5b9511d 100644 --- a/README.md +++ b/README.md @@ -9,7 +9,7 @@ TorchScale is a PyTorch library that allows researchers and developers to scale It has the implementation of fundamental research to improve modeling generality and capability, as well as training stability and efficiency of scaling Transformers. - Stability - [**DeepNet**](https://arxiv.org/abs/2203.00555): scaling Transformers to 1,000 Layers and beyond -- Generality - [**Foundation Transformers (Magneto)**](https://arxiv.org/abs/2210.06423): towards the big convergence of model architectures across language, vision, speech, and multimodal +- Generality - [**Foundation Transformers (Magneto)**](https://arxiv.org/abs/2210.06423): towards true general-purpose modeling across tasks and modalities (including language, vision, speech, and multimodal) - Efficiency - [**X-MoE**](https://arxiv.org/abs/2204.09179): scalable & finetunable sparse Mixture-of-Experts (MoE) ## News