From 774003903e6ba7dfd6103e4ebc249db3450393e1 Mon Sep 17 00:00:00 2001 From: gitnlp <36983436+gitnlp@users.noreply.github.com> Date: Wed, 26 Jul 2023 18:38:49 +0800 Subject: [PATCH] Update README.md --- README.md | 6 ++++++ 1 file changed, 6 insertions(+) diff --git a/README.md b/README.md index d063581..c6de432 100644 --- a/README.md +++ b/README.md @@ -13,6 +13,12 @@ It has the implementation of fundamental research to improve modeling generality - Capability - A [**Length-Extrapolatable**](https://arxiv.org/abs/2212.10554) Transformer - Efficiency - [**X-MoE**](https://arxiv.org/abs/2204.09179): scalable & finetunable sparse Mixture-of-Experts (MoE) +### Revolutionizing Transformers for (M)LLMs and AI + +> [**RetNet**](https://arxiv.org/abs/2307.08621): Retentive Network: A Successor to Transformer for Large Language Models + +> [**LongNet**](https://arxiv.org/abs/2307.02486): Scaling Transformers to 1,000,000,000 Tokens + ## News - November, 2022: TorchScale 0.1.1 released [[Paper](https://arxiv.org/abs/2211.13184)] [[PyPI](https://pypi.org/project/torchscale/)]