Update README.md
This commit is contained in:
parent
774003903e
commit
f58c8247be
|
@ -14,10 +14,8 @@ It has the implementation of fundamental research to improve modeling generality
|
||||||
- Efficiency - [**X-MoE**](https://arxiv.org/abs/2204.09179): scalable & finetunable sparse Mixture-of-Experts (MoE)
|
- Efficiency - [**X-MoE**](https://arxiv.org/abs/2204.09179): scalable & finetunable sparse Mixture-of-Experts (MoE)
|
||||||
|
|
||||||
### Revolutionizing Transformers for (M)LLMs and AI
|
### Revolutionizing Transformers for (M)LLMs and AI
|
||||||
|
- [**RetNet**](https://arxiv.org/abs/2307.08621): Retentive Network: A Successor to Transformer for Large Language Models
|
||||||
> [**RetNet**](https://arxiv.org/abs/2307.08621): Retentive Network: A Successor to Transformer for Large Language Models
|
- [**LongNet**](https://arxiv.org/abs/2307.02486): Scaling Transformers to 1,000,000,000 Tokens
|
||||||
|
|
||||||
> [**LongNet**](https://arxiv.org/abs/2307.02486): Scaling Transformers to 1,000,000,000 Tokens
|
|
||||||
|
|
||||||
## News
|
## News
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue
Block a user