Open Models

Aqui-open1 Model Family

We hereby present the first series of small open models by Aqui, trained from scratch and with an MIT license. The open1 family delivers state-of-the-art performance across reasoning, mathematics, and coding tasks while maintaining efficient inference capabilities.

πŸš€ Released September 7, 2025
πŸ”“ MIT Licensed
⚑ Efficient Architecture

open1-1.5B-Instruct

Ultra-efficient model optimized for edge deployment and real-time applications.

🧠 1.5B parameters
πŸ“ 128K context
⚑ Fast inference
View Model

open1-7.5B-Instruct

Balanced model providing exceptional performance across diverse tasks with reasonable compute requirements.

🧠 7.5B parameters
πŸ“ 32K context
🎯 High accuracy
View Model

πŸ”œ Coming This Week

Aqui-open1-4x8B β€” Our biggest non-thinking open model, head-to-head against Qwen3 32B and Llama 3.3 70B. Stay tuned for the most capable open model in the series.


Benchmark Performance

1.5B Model Comparison

Metric open1-1.5B-Instruct Llama-3.2-1B-Instruct LFM2-1.2B Qwen3-1.7B Gemma-3-1B-it SmolLM2-1.7B-Instruct
MMLU58.546.655.259.140.142.3
GPQA32.328.831.527.721.122.1
GSM8K62.635.758.351.459.648.2
IFEval72.752.474.974.062.956.7
MGSM59.129.155.066.643.638.5
Average57.038.555.055.845.541.6

7.5B Model Comparison

Benchmark open1-7.5B-Instruct Llama-3.1-8B-Instruct LFM-7B Qwen3-8B Gemma-3-12B-it Nemotron-Nano-9B-v2
MMLU75.868.769.471.672.574.5
HumanEval82.371.770.184.884.886.2
GPQA Diamond52.225.932.945.234.940.8
IFEval78.977.071.683.481.584.3
AIME 202518.94.32.120.218.320.1
Average61.649.549.261.058.461.2

Key Features

🎯 Superior Reasoning

Exceptional performance on MMLU, GPQA, and mathematical reasoning tasks, outperforming models of similar and larger sizes.

⚑ Optimized Architecture

Efficient transformer design enabling fast inference while maintaining high accuracy across diverse benchmarks.

🌍 Multilingual Support

Trained on 20+ languages with robust performance across linguistic boundaries and cultural contexts.

πŸ”“ MIT Licensed

Complete freedom for commercial use, modification, and redistribution with minimal restrictions.


Usage

    from transformers import AutoTokenizer, AutoModelForCausalLM
    
# Load the model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("aquigpt/open1-1.5B-Instruct")
model = AutoModelForCausalLM.from_pretrained("aquigpt/open1-1.5B-Instruct")

# Generate text
inputs = tokenizer("Explain quantum computing:", return_tensors="pt")
outputs = model.generate(**inputs, max_length=200, temperature=0.7)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training Details

The open1 models were trained from scratch on a diverse, high-quality dataset spanning code, mathematics, reasoning, and multilingual text. Training utilized advanced techniques including:

  • Supervised fine-tuning on instruction-following data
  • Constitutional AI for alignment and safety
  • Advanced attention mechanisms for extended context
  • Multi-stage training with curriculum learning
Note: These models are designed for research and commercial applications. While they demonstrate strong performance, users should conduct appropriate testing for their specific use cases.

Built with ❀️ by the Aqui team β€’ MIT β€’ September 2025

Downloads last month
18
Safetensors
Model size
7.46B params
Tensor type
F16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for aquigpt/open1-7.5B-Instruct

Quantizations
2 models

Collection including aquigpt/open1-7.5B-Instruct