NRM: Nvyra Recursive Reasoning Model

Developed by Nvyra X — Fact-Checking and Disinformation Detection Service

Model Description

NRM (Nvyra Recursive Reasoning Model) is a state-of-the-art reasoning architecture that combines:

  • Mixture of Recursions (MoR) - Weight-tied transformer blocks applied recursively
  • Multi-Head Latent Attention (MLA) - 10× KV cache reduction (DeepSeek-V3)
  • ConvSwiGLU - Enhanced nonlinearity from URM paper
  • Aux-Loss-Free MoE - Bias-based expert load balancing
  • PonderNet - Adaptive computation time
  • Multi-Token Prediction - 4-ahead planning

Training

  • Budget: $115 ($25 Nebius + $90 Modal)
  • Hardware: H200 NVLink GPUs
  • Framework: PyTorch 2.9.1, Flash Attention 3, CUDA 12.8
  • Dataset: 300K+ reasoning examples (Sudoku, ARC, Logic, Object Tracking)

Usage

# This model uses a custom architecture - see repository for full code
from safetensors.torch import load_file
weights = load_file("model.safetensors")

Citation

If you use this model, please cite:

@misc{nrm2025,
  title={NRM: Nvyra Recursive Reasoning Model},
  author={Nvyra X Research Team},
  year={2025},
  url={https://huggingface.co/Feargal/nvyra-x-reasoning}
}

References

Downloads last month
37
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Papers for Feargal/nvyra-x-reasoning