# OpenMythos My very theoriticial interpretation of what claude mythos could be # References ## Twitter / X - Why Claude Mythos is so good — looped transformer theory (Sigrid Jin): https://x.com/realsigridjin/status/2044620031410266276 - LT implicit reasoning over parametric knowledge unlocks generalization (Yuekun Yao): https://x.com/yuekun_yao/status/2044229171627639004 - Looped transformer cyclic trajectories and input injection (rosinality): https://x.com/rosinality/status/2043953033428541853 - Parcae scaling laws for stable looped language models — thread (Hayden Prairie): https://x.com/hayden_prairie/status/2044453231913537927 - RoPE-like loop index embedding idea to differentiate functions across iterations (davidad): https://x.com/davidad/status/2044453231913537927 ## Papers - Loop, Think, & Generalize — Implicit Reasoning in Recurrent Depth Transformers: https://arxiv.org/pdf/2604.07822 - Parcae — Scaling Laws for Stable Looped Language Models: https://arxiv.org/abs/2604.12946 - Parcae blog: https://sandyresearch.github.io/parcae/ - Universal Transformers: https://arxiv.org/pdf/1807.03819 - Reasoning with Latent Thoughts — On the Power of Looped Transformers: https://arxiv.org/abs/2502.17416 - Training Large Language Models to Reason in a Continuous Latent Space: https://arxiv.org/abs/2412.06769 - Relaxed Recursive Transformers — Effective Parameter Sharing with Layer-wise LoRA: https://arxiv.org/pdf/2410.20672