Skip to content

phi-4 reasoning models #2047

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 11 commits into from
May 27, 2025
Merged

phi-4 reasoning models #2047

merged 11 commits into from
May 27, 2025

Conversation

ysjprojects
Copy link
Collaborator

Models added:
Phi-4-mini-reasoning | 3.8B
Phi-4-reasoning | 14B
Phi-4-reasoning-plus | 14B

Relevant links:
https://arxiv.org/abs/2504.21233 (mini-reasoning)
https://arxiv.org/abs/2504.21318 (reasoning)
https://azure.microsoft.com/en-us/blog/one-year-of-phi-small-language-models-making-big-leaps-in-ai/

Highlight:

Despite their significantly smaller size, both models achieve better performance than OpenAI o1-mini and DeepSeek-R1-Distill-Llama-70B at most benchmarks, including mathematical reasoning and Ph.D. level science questions. They achieve performance better than the full DeepSeek-R1 model (with 671-billion parameters) on the AIME 2025 test, the 2025 qualifier for the USA Math Olympiad.

@Borda Borda enabled auto-merge (squash) May 22, 2025 12:11
@Borda Borda disabled auto-merge May 27, 2025 10:49
@Borda Borda merged commit 241bbd6 into Lightning-AI:main May 27, 2025
24 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants