Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
Open to Work
3
3
19
Tarun Reddi
PRO
Teen-Different
Follow
charvi020's profile picture
blazej-mako's profile picture
2 followers
·
17 following
https://redditarun.github.io/
_TeenDifferent
REDDITARUN
tarunreddi
AI & ML interests
Generative AI, Modular AI Systems, Reinforcement Learning
Recent Activity
posted
an
update
about 8 hours ago
Adaptive Attention at Inference Time: Does It Actually Work? A hypernetwork that rewires GPT's value heads on every forward pass. The answer: not a clean win — but not a failure either. Blog post: https://teendifferent.substack.com/p/adaptive-attention-at-inference-time Code: https://github.com/REDDITARUN/a-gpt Weights: https://huggingface.co/Teen-Different/adaptive-gpts What This Is Five small language model variants trained for 12k steps on a 300M token mixed corpus, answering one question: can the residual stream be used to slightly rewrite the model's own computation while it's running? Instead of a fixed W_v for every context, a TinyHeadTransformer hypernetwork generates low-rank (LoRA-style) updates to the value projection of each attention head — conditioned on the current residual stream. Each token gets a dynamically adapted value transformation. The Five Models Base GPT — 28.9M params, 139 tok/s, val loss ~3.82 Matched GPT (+2 layers) — 30.5M params, 204 tok/s, val loss ~3.80 Adaptive GPT — 30.5M params, 38.7 tok/s, val loss ~3.88–3.92 Diffusion GPT — 28.9M params, 110 tok/s, val loss ~5.0–5.2 Adaptive Diffusion GPT — 30.5M params, 40.4 tok/s, val loss ~5.0–5.2 Architecture: 4 layers, 4 heads, d_model=256, context=256, RoPE, GPT-2 tokenizer. How the Hypernetwork Works For each attention head, a TinyHeadTransformer encodes the head's residual stream slice, mean-pools it to a conditioning vector, then projects into low-rank factors A (d×r) and B (r×d) at rank=8. The dynamic value update follows LoRA conventions with alpha/r scaling. B is zero-initialized so the adaptive path starts inert and the model begins as a vanilla GPT — critical for training stability. The diffusion variant uses bidirectional attention, RMSNorm, squared ReLU, and a learned timestep embedding.
updated
a model
about 9 hours ago
Teen-Different/adaptive-gpts
published
a model
2 days ago
Teen-Different/adaptive-gpts
View all activity
Organizations
Teen-Different
's models
14
Sort: Recently updated
Teen-Different/adaptive-gpts
Updated
about 9 hours ago
Teen-Different/qwen2.5-1.5b-arc1-decoded-dft
Text Generation
•
2B
•
Updated
13 days ago
•
9
Teen-Different/qwen2.5-1.5b-arc1-encoded-dft
Updated
13 days ago
Teen-Different/arc-it
Updated
22 days ago
•
52
Teen-Different/CLIP-ViT-IJEPA-VLMs-1.5B
Image-Text-to-Text
•
Updated
25 days ago
Teen-Different/CLIP-ViT-IJEPA-VLMs-0.5B
Image-Text-to-Text
•
Updated
25 days ago
Teen-Different/smolvlm-256m-latex
Image-Text-to-Text
•
0.3B
•
Updated
Jan 9
•
1
Teen-Different/Qwen2.5-Coder-3B-KernelBook-Finetuned
3B
•
Updated
Aug 1, 2025
•
3
•
5
Teen-Different/TD-HallOumi-3B
Text Classification
•
3B
•
Updated
Apr 24, 2025
•
3
•
2
Teen-Different/Driver-Drowsiness-Detection
Updated
Mar 31, 2025
•
2
Teen-Different/F.E.A.S.T
Object Detection
•
Updated
Mar 30, 2025
Teen-Different/RxRovers_Roaming_for_Rapid_Relief
Reinforcement Learning
•
Updated
Mar 30, 2025
Teen-Different/squiral_maze
Reinforcement Learning
•
Updated
Mar 30, 2025
Teen-Different/Tabular_RL_For_Multi_Env
Reinforcement Learning
•
Updated
Mar 30, 2025