mirror of
https://github.com/kyegomez/OpenMythos.git
synced 2026-05-02 17:43:27 +02:00
25 lines
1.4 KiB
Markdown
25 lines
1.4 KiB
Markdown
# OpenMythos
|
|
My very theoriticial interpretation of what claude mythos could be
|
|
|
|
|
|
|
|
# References
|
|
|
|
## Twitter / X
|
|
|
|
- Why Claude Mythos is so good — looped transformer theory (Sigrid Jin): https://x.com/realsigridjin/status/2044620031410266276
|
|
- LT implicit reasoning over parametric knowledge unlocks generalization (Yuekun Yao): https://x.com/yuekun_yao/status/2044229171627639004
|
|
- Looped transformer cyclic trajectories and input injection (rosinality): https://x.com/rosinality/status/2043953033428541853
|
|
- Parcae scaling laws for stable looped language models — thread (Hayden Prairie): https://x.com/hayden_prairie/status/2044453231913537927
|
|
- RoPE-like loop index embedding idea to differentiate functions across iterations (davidad): https://x.com/davidad/status/2044453231913537927
|
|
|
|
## Papers
|
|
|
|
- Loop, Think, & Generalize — Implicit Reasoning in Recurrent Depth Transformers: https://arxiv.org/pdf/2604.07822
|
|
- Parcae — Scaling Laws for Stable Looped Language Models: https://arxiv.org/abs/2604.12946
|
|
- Parcae blog: https://sandyresearch.github.io/parcae/
|
|
- Universal Transformers: https://arxiv.org/pdf/1807.03819
|
|
- Reasoning with Latent Thoughts — On the Power of Looped Transformers: https://arxiv.org/abs/2502.17416
|
|
- Training Large Language Models to Reason in a Continuous Latent Space: https://arxiv.org/abs/2412.06769
|
|
- Relaxed Recursive Transformers — Effective Parameter Sharing with Layer-wise LoRA: https://arxiv.org/pdf/2410.20672
|