Update README.md
This commit is contained in:
parent
7c45a98f6e
commit
41f6ee5687
13
README.md
13
README.md
|
@ -1,14 +1,11 @@
|
||||||
# Project
|
# OpenScale - Transformers at (any) Scale
|
||||||
|
|
||||||
> This repo has been populated by an initial template to help get you started. Please
|
Fundamental research to improve modeling generality and capability, as well as training stability and efficiency of scaling Transformers.
|
||||||
> make sure to update the content to build a great experience for community-building.
|
|
||||||
|
|
||||||
As the maintainer of this project, please make a few updates:
|
- Stability - [**DeepNet**](https://arxiv.org/abs/2203.00555): scaling Transformers to 1,000 Layers and beyond
|
||||||
|
- Generality - [**Foundation Transformers (Magneto)**](https://arxiv.org/abs/2210.06423)
|
||||||
|
- Efficiency - [**X-MoE**](https://arxiv.org/abs/2204.09179): scalable & finetunable sparse Mixture-of-Experts (MoE)
|
||||||
|
|
||||||
- Improving this README.MD file to provide a great experience
|
|
||||||
- Updating SUPPORT.MD with content about this project's support experience
|
|
||||||
- Understanding the security reporting process in SECURITY.MD
|
|
||||||
- Remove this section from the README
|
|
||||||
|
|
||||||
## Contributing
|
## Contributing
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue
Block a user