Text Generation
Portuguese

๐Ÿง  MiniAxion1.5-3M

Emergent reasoning in a 2.7M parameter model. A tiny Portuguese-first language model that learns how to think before it learns how to be correct.

๐Ÿš€ Overview

MiniAxion1.5-3M is an ultra-compact (~2.7M parameters) GPT-style language model designed to investigate reasoning emergence at extreme small scale.

Unlike typical small models optimized for fluency, MiniAxion is explicitly trained to produce:

Structured reasoning traces Step-by-step thinking () Deterministic answer formatting

It operates primarily in Portuguese, making it a rare example of a non-English reasoning-first nano model.

โšก Why This Model Is Interesting

Most models follow this trajectory:

Language โ†’ Knowledge โ†’ Reasoning

MiniAxion flips part of that:

Structure โ†’ Reasoning format โ†’ (still learning correctness)

๐Ÿ’ก Key insight:

The model demonstrates that reasoning structure can emerge independently of reasoning accuracy.

๐Ÿงช Evaluation Task Performance Task Accuracy Addition 10% Subtraction 10% Multiplication 0% Even/Odd 100% Comparison 5% Sequence Completion 0% Word Problems (Addition) 10% Word Problems (Subtraction) 0% Word Problems (Multiplication) 10% True/False 100% Chat/Greetings 100%

๐Ÿง  Reasoning Behavior Metrics Metric Score Thinking Rate 100% Step Format 100% Answer Completion 100%

โœ” The model always thinks โœ” The model always structures reasoning โœ” The model always produces an answer

๐Ÿ“Š Interpretation

MiniAxion exhibits a clear dissociation:

โœ… What it learned Reasoning format Step-by-step decomposition Logical task patterns (parity, boolean) โŒ What it did NOT learn Arithmetic correctness Numerical reasoning Multi-step computation

๐Ÿ”ฌ Core Finding

Reasoning โ‰  Correctness

MiniAxion shows that:

Models can internalize thinking patterns Without actually learning how to solve problems

This makes it a strong candidate for studying:

Emergent reasoning Tiny Recursive Models (TRMs) Reasoning distillation

๐Ÿ—๏ธ Architecture Type: GPT-style Transformer Parameters: ~2.7M Objective: Next-token prediction Language: Portuguese (primary) Specialization: Structured reasoning traces

๐Ÿง  Training Strategy

The model was trained with a reasoning-first approach:

Portuguese language grounding Structured reasoning data () Emphasis on: Deterministic formats Multi-step thinking Explicit reasoning tokens

๐Ÿšซ No RLHF ๐Ÿšซ No instruction tuning at scale ๐Ÿšซ No large model distillation (yet)

โš ๏ธ Limitations

  1. Arithmetic Collapse

Near-random performance in:

Addition

Subtraction

Multiplication

โ†’ Indicates lack of numerical representation learning

Strong dependence on:

Prompt format

Token patterns

Seen reasoning templates

๐Ÿ”ฎ Future Work

This model is just the beginning.

๐Ÿ“ˆ Scaling

5M / 10M / 20M versions

Track emergence of correctness

๐Ÿงช Distillation

Inject reasoning from larger models

Improve accuracy without scaling params

๐Ÿ” Self-Play / Synthetic Data

Generate reasoning loops

Reinforce correct chains

๐Ÿงฉ Hybrid Reasoning

Combine symbolic + neural learning

Fix arithmetic weakness

๐Ÿงพ Example Output

Identifico os nรบmeros Tento somar os valores Ajusto o resultado 74

โœ” Perfect reasoning structure โŒ Incorrect answer

๐Ÿ’ก Takeaway

MiniAxion1.5-3M proves something important:

Even a 2.7M model can learn to simulate thinking before it learns to actually think correctly.

๐Ÿค Use Cases

Research on emergent reasoning

Tiny model experimentation (CPU-friendly)

Educational demos of:

Chain-of-Thought

Reasoning failure modes

Base model for:

Distillation

NRM experiments

Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Dataset used to train AxionLab-Co/MiniAxion1.5-3M