We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 15c506c commit f69a87fCopy full SHA for f69a87f
README.md
@@ -14,7 +14,7 @@ Fundamental research to develop new architectures for foundation models and A(G)
14
- Efficiency - [**X-MoE**](https://arxiv.org/abs/2204.09179): scalable & finetunable sparse Mixture-of-Experts (MoE)
15
16
### Revolutionizing Transformers for (M)LLMs and AI
17
-> [**BitNet**](https://arxiv.org/abs/2310.11453): 1-bit Transformers for Large Language Models
+- [**BitNet**](https://arxiv.org/abs/2310.11453): 1-bit Transformers for Large Language Models
18
- [**RetNet**](https://arxiv.org/abs/2307.08621): Retentive Network: A Successor to Transformer for Large Language Models
19
- [**LongNet**](https://arxiv.org/abs/2307.02486): Scaling Transformers to 1,000,000,000 Tokens
20
0 commit comments