From 22438a85254640d41e1f1d8b37857a86c68c0ccc Mon Sep 17 00:00:00 2001 From: gitnlp <36983436+gitnlp@users.noreply.github.com> Date: Fri, 23 Dec 2022 08:26:08 +0800 Subject: [PATCH] Update README.md --- README.md | 1 + 1 file changed, 1 insertion(+) diff --git a/README.md b/README.md index 641be1f..9d469cc 100644 --- a/README.md +++ b/README.md @@ -10,6 +10,7 @@ It has the implementation of fundamental research to improve modeling generality - Stability - [**DeepNet**](https://arxiv.org/abs/2203.00555): scaling Transformers to 1,000 Layers and beyond - Generality - [**Foundation Transformers (Magneto)**](https://arxiv.org/abs/2210.06423): towards true general-purpose modeling across tasks and modalities (including language, vision, speech, and multimodal) +- Capability - A [**Length-Extrapolatable**](https://arxiv.org/abs/2212.10554) Transformer - Efficiency - [**X-MoE**](https://arxiv.org/abs/2204.09179): scalable & finetunable sparse Mixture-of-Experts (MoE) ## News