willcb commited on
Commit
00eba36
·
verified ·
1 Parent(s): ecff4ea

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -0
README.md ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - prime-rl
5
+ - verifiers
6
+ - prime-intellect
7
+ license: mit
8
+ language:
9
+ - en
10
+ base_model:
11
+ - zai-org/GLM-4.5-Air-Base
12
+ pipeline_tag: text-generation
13
+ ---
14
+
15
+ # INTELLECT-3
16
+
17
+ **INTELLECT-3** is a 100B+ parameter Mixture-of-Experts reasoning model post-trained from [GLM-4.5-Air-Base](https://huggingface.co/zai-org/GLM-4.5-Air-Base) using supervised fine-tuning (SFT) followed by large-scale reinforcement learning (RL).
18
+
19
+ Training was performed with [prime-rl](https://github.com/PrimeIntellect-ai/prime-rl) using environments built with the [verifiers](https://github.com/PrimeIntellect-ai/verifiers) library. All training and evaluation environments are available on the [Environments Hub](https://app.primeintellect.ai/dashboard/environments).
20
+
21
+ The model, training frameworks, and environments are open-sourced under fully-permissive licenses (MIT and Apache 2.0).
22
+
23
+ For more details, see the [technical report](PAPER_LINK_PLACEHOLDER).
24
+
25
+ ## Evaluation
26
+
27
+ INTELLECT-3 achieves best-in-class performance on math, coding, and reasoning benchmarks:
28
+
29
+ | Benchmark | Score |
30
+ |-----------|-------|
31
+ | AIME 2025 | 88.0 |
32
+ | LiveCodeBench v6 | 69.3 |
33
+ | GPQA Diamond | 74.4 |
34
+ | HLE | 14.6 |
35
+
36
+ ## Model Variants
37
+
38
+ | Model | HuggingFace |
39
+ |-------|-------------|
40
+ | INTELLECT-3 | [PrimeIntellect/INTELLECT-3](https://huggingface.co/PrimeIntellect/INTELLECT-3) |
41
+ | INTELLECT-3-FP8 | [PrimeIntellect/INTELLECT-3-FP8](https://huggingface.co/PrimeIntellect/INTELLECT-3-FP8) |
42
+
43
+ ## Serving with vLLM
44
+
45
+ The BF16 version can be served on 2x H200s:
46
+ ```bash
47
+ vllm serve PrimeIntellect/INTELLECT-3 \
48
+ --tensor-parallel-size 2 \
49
+ --tool-call-parser qwen3_coder \
50
+ --reasoning-parser deepseek_r1
51
+ ```
52
+
53
+ The FP8 version can be served on a single H200:
54
+
55
+ ```bash
56
+ vllm serve PrimeIntellect/INTELLECT-3-FP8 \
57
+ --tool-call-parser qwen3_coder \
58
+ --reasoning-parser deepseek_r1
59
+ ```
60
+
61
+ ## Citation
62
+
63
+ ```bibtex
64
+ @misc{intellect3,
65
+ title={INTELLECT-3: Technical Report},
66
+ author={Prime Intellect Team},
67
+ year={2025},
68
+ url={https://huggingface.co/PrimeIntellect/INTELLECT-3}
69
+ }
70
+ ```