GPT-OSS-120B
The open-weight Mixture-of-Experts model powering MathBeast's advanced mathematical reasoning. 120B parameters, optimized for STEM problem-solving and chain-of-thought reasoning.
120B Parameters
Mixture-of-Experts architecture with 120B total parameters, activating only 5.1B per pass.
Chain-of-Thought
Advanced reasoning capabilities with configurable depth for mathematical proofs.
Efficient Inference
Runs on a single H100 GPU with MXFP4 quantization for cost-effective deployment.
Apache 2.0 License
Fully open-weight model with permissive licensing for commercial use.
Open-Weight Mixture-of-Experts
Leveraging OpenAI's open-source 117B parameter model with only 5.1B active per pass, enabling sophisticated mathematical reasoning on a single H100 GPU.
8 of 256 experts activated for current mathematical reasoning task
Native 4-bit for single H100 deployment
Key Achievement: GPT-OSS-120B achieves 94.2% on GSM8K and solves 11/15 AIME 2024 problems, rivaling proprietary models at a fraction of the computational cost.
- MathQA (37K problems)
- GSM8K (8.5K problems)
- Competition Math (12.5K)
- AMC/AIME/USAMO/IMO
- Chain-of-Thought Prompting
- Multi-Step Verification
- Alternative Approaches
- Configurable Reasoning Levels
- Single H100 GPU
- Docker + NVIDIA Runtime
- Kubernetes Scaling
- Apache 2.0 License