MerlinSafety/Qwen3.5-4B-Safety-Thinking
Text Generation • 4B • Updated
• 1.31k • 4
Independent AI safety lab. Stockholm, Sweden. We test deployed LLM agents under adversarial conditions and measure behavioral alignment in production — not in controlled benchmarks.
Merlin Research is an independent AI safety and reasoning research organization focused on building practical, auditable, and robust open models.
We develop and evaluate models that are:
Our flagship releases are published under this organization with:
We welcome collaboration on:
For partnerships or research collaboration, contact us via Hugging Face discussions or linked channels in our repositories.
Merlin Research
Safe reasoning. Measurable alignment. Real-world robustness.