mirror of
https://github.com/kyegomez/OpenMythos.git
synced 2026-05-02 17:43:27 +02:00
Update README.md
This commit is contained in:
parent
f261645e5f
commit
8c68c1fcc7
@ -415,6 +415,7 @@ Theoretical analysis suggests 2-3x improvements in inference throughput. For a d
|
|||||||
- Relaxed Recursive Transformers — Effective Parameter Sharing with Layer-wise LoRA: https://arxiv.org/pdf/2410.20672
|
- Relaxed Recursive Transformers — Effective Parameter Sharing with Layer-wise LoRA: https://arxiv.org/pdf/2410.20672
|
||||||
- Mixture-of-Depths Attention: https://arxiv.org/abs/2603.15619
|
- Mixture-of-Depths Attention: https://arxiv.org/abs/2603.15619
|
||||||
- Hyperloop Transformers: https://arxiv.org/abs/2604.21254
|
- Hyperloop Transformers: https://arxiv.org/abs/2604.21254
|
||||||
|
- The Recurrent Transformer: Greater Effective Depth and Efficient Decoding: https://arxiv.org/abs/2604.21215
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user