Skip to content
View Abd0r's full-sized avatar
🫐
Training an LVM
🫐
Training an LVM

Highlights

  • Pro

Block or report Abd0r

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Abd0r/README.md
Pixel Ghost

Hey, I'm Abdur 👋

17 y/o · Independent Researcher · India
Building artificial intelligence that goes beyond token prediction

X   Hugging Face   Email   ORCID   PyPI


About Me

I'm a self-taught researcher working without a CS degree, a lab, or a team. I reverse-engineer how the brain works to build AI systems fundamentally different from today's transformers.

My work spans neural architectures, training frameworks, cognitive systems, and — most recently — post-CMOS computing hardware. Different domains, same approach: rebuild from first principles, and eliminate what everyone else assumed was necessary.


🔬 Published Work

⚛️ FEA — Free Electron Absorption Architecture

A transistor-free computing architecture on hydrogen-passivated silicon. Computation happens by resonant electron absorption in 5-atom dangling-bond clusters, not by switching. On a 3 cm² die: 14.1 TB of in-situ memory at 79 mW — roughly 110× the M4 Max memory, at ~0.2% of the power. Memory and compute are the same atoms — no cache, no DRAM bus. My first paper outside machine learning.

 

🧭 Quatrix — Q-Compass Architecture

Replaces standard attention with Q-Compass — sequence mixing grounded in reinforcement learning navigation theory instead of geometric similarity. Three projections instead of four. 69% fewer attention parameters. One mechanism handles text, vision, audio, and world modeling.

   

🧠 Artificial Neural Mesh (ANM) V0

A modular multi-agent cognitive architecture featuring 12 specialized domain experts collaborating through Web-of-Thought (WoT) reasoning.

 

⚡ GEKO — Gradient-Efficient Knowledge Optimization

A plug-and-play fine-tuning framework that skips samples the model already knows — routing compute to hard samples and freezing mastered ones. Up to 80% compute savings at scale.

     

🔥 Currently Training

Berry-Q0 Model

Berry-Q0 — ~50M parameters, trained from scratch on a single laptop GPU (RTX 4050, 6GB VRAM). Text + vision, currently in GRPO reasoning training (R1-style, math domain).

The goal: push a 50M model as far as possible on reasoning. No cloud. No team. Just architecture.


Built from scratch · No lab · No shortcuts

Pinned Loading

  1. FEA FEA Public

    FEA: Free Electron Absorption Architecture

    C++

  2. quatrix quatrix Public

    Quatrix — Q-Compass Architecture: novel neural architecture replacing attention with value-based navigation. Base for the Quasar model series.

    Python

  3. GEKO GEKO Public

    Intelligent training framework that automatically skips mastered samples and gives 5× more compute to hard ones. Up to 80% compute savings on LLM fine-tuning.

    Python 1

  4. Artificial-Neural-Mesh-V0 Artificial-Neural-Mesh-V0 Public

    A Multi Agent Reasoning System.

    Python