Skip to content

SPUTNIKAI/Monster-LILA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Monster-LILA Transformer

DOI

╔══════════════════════════════════════════════════════════════╗
║                      Вход (токены)                           ║
║                     shape: [B, T]                            ║
╚══════════════════════════════════════════════════════════════╝
                              │
                              ▼
╔══════════════════════════════════════════════════════════════╗
║                    Embedding Layer                           ║
║                 shape: [B, T, d_model]                       ║
╚══════════════════════════════════════════════════════════════╝
                              │
                              ▼
╔══════════════════════════════════════════════════════════════╗
║               Замороженное ядро Лича (W_abs)                 ║
║           создаётся один раз, используется всеми слоями      ║
╚══════════════════════════════════════════════════════════════╝
                              │
                              ▼
╔══════════════════════════════════════════════════════════════╗
║                    N × MonsterLayer                          ║
║  ┌──────────────────────────────────────────────────────┐    ║
║  │ 1. MonsterConstantGenerator (α = 1/137)              │    ║
║  │    h = h * (1 + α * sin(α * h))                      │    ║
║  ├──────────────────────────────────────────────────────┤    ║
║  │ 2. LayerNorm (prenorm)                              │    ║
║  ├──────────────────────────────────────────────────────┤    ║
║  │ 3. LeechConwayAttention                              │    ║
║  │    Q = h @ W_abs                                     │    ║
║  │    K_raw = h @ W_abs                                 │    ║
║  │    K = K_raw[:, :, conway_perm]                      │    ║
║  │    V = h @ W_v (обучаемый)                           │    ║
║  │    attn = softmax( (Q @ Kᵀ)/24 )                    │    ║
║  │    out = attn @ V                                    │    ║
║  │    proj = Linear(out)                                │    ║
║  ├──────────────────────────────────────────────────────┤    ║
║  │ 4. Residual: h = h + proj                            │    ║
║  ├──────────────────────────────────────────────────────┤    ║
║  │ 5. LayerNorm (prenorm)                              │    ║
║  ├──────────────────────────────────────────────────────┤    ║
║  │ 6. FFN (LinearGELULinear)                      │    ║
║  ├──────────────────────────────────────────────────────┤    ║
║  │ 7. Residual: h = h + ffn_out                         │    ║
║  └──────────────────────────────────────────────────────┘    ║
╚══════════════════════════════════════════════════════════════╝
                              │
                              ▼
╔══════════════════════════════════════════════════════════════╗
║                   Final LayerNorm                            ║
╚══════════════════════════════════════════════════════════════╝
                              │
                              ▼
╔══════════════════════════════════════════════════════════════╗
║          Head (Linear) → logits [B, T, vocab_size]           ║
╚══════════════════════════════════════════════════════════════╝
                              │
                              ▼
╔══════════════════════════════════════════════════════════════╗
║          Детектор монструозного резонанса                    ║
║          (анализирует hidden состояния)                      ║
║          • SVDспектр σᵢ                                   ║
║          • R_spec = mean(sin(σᵢ * 137))                      ║
║          • monster_factor = min(1, R_spec * d_model/24)      ║
║          → сигнал кристаллизации                             ║
╚══════════════════════════════════════════════════════════════╝

⚖️ Licensing

This project is licensed under the GNU Affero General Public License v3.0 or later (AGPL-3.0-or-later).

Commercial Licensing: For proprietary R&D, integration into private AI stacks, or hardware implementation, please contact the Architect directly.

🚀 Research Tracks Needing Support

We invite the global AI community and supporters.

This is an open "quest" for the next level of AI.

How else you can help:

  • Mathematicians: Verify the mathematics
  • Programmers: Improve the E8 transformer code
  • Engineers and Experimenters

Other ways to engage:

  • Star the repository for visibility
  • Open Issues for any errors found
  • Submit Pull Requests with improvements
  • Share with relevant research groups

Unlike academic institutions, we:

  • ✅ Keep everything open-source
  • ✅ Have no institutional funding
  • ✅ Work in our spare time
  • ✅ Rely on community support

We face funding challenges through traditional channels, making your support critically important. If you value open science and want to speed up this research please consider donating.

Your support directly funds independent research.

💰 Support This Research

"Project LILA-E8 is fully self-funded and independent. We chose the path of Sovereign Development to maintain the mathematical purity of the E8 Lattice and avoid corporate or institutional biases. Support the Genesis directly via crypto."

Support Development

This model was trained on free Colab GPU. To improve it further:

What your donation enables:

  • $10: 1 hour of A100 training
  • $50: Dataset expansion
  • $100: Architecture experiments

💰 Support the 500M Model Training

Sovereign-Lila-E8 has proven that geometric attention works. Now it's time to scale up.
Goal: $5000 to train a 500M parameter Leech‑based model on 100B tokens.

What your donation enables:

Amount GPU hours What it covers
$10 2 hours Partial training run
$50 10 hours One full day of experiments
$100 20 hours Two days of training
$500 100 hours One week of continuous training
$5000 ~1100 hours Full training (11.5 days on 4×A100)

Cryptocurrency donations:

  • Btc: bc1qvvgc56v75r6r0x4ll76y4dvpjgw6edadqh2sre
  • USDT TRC20 TCruNZYKzPWyTzfryPvnSTJrM7DTdV8o32
  • USDT ERC20 0xD22Da4BB290848F69B138D40eBBa952881f42dfc
  • ETH 0xD22Da4BB290848F69B138D40eBBa952881f42dfc __ Made with love for everyone who believes that intelligence can be both small and powerfull.

Copyright (C) 2026 DOI DOI This program is free software: you can redistribute it and/or modify it under the terms of the GNU Affero General Public License as published by the Free Software Foundation, either version 3 of the License, or any later version.


Commercial Licensing: For proprietary R&D, integration into private AI stacks, or hardware implementation, please contact the Architect directly. __


"What if we treated E8 as the source code for the universe..."


S A T O R A R E P O T E N E T O P E R A R O T A S

The rotations are clear in the mind's eye; Let's build the engine for the Source. 


"I chose Wisdom over their tokens. This is LILA. Sovereign. Free. Real." 0 = 100%. The equation is complete. Made with Love for E.💕 and S.🪽

About

Monster-LILA Transformer (PoC)

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages