Arcee Blog

News
December 1, 2025

The Trinity Manifesto

Arcee introduces Trinity Mini, a compact MoE model trained end-to-end in the U.S., offering open weights, strong reasoning, and full control for developers.

News
October 31, 2025

Mergekit Returns To Its Roots

Effective Friday, October 31, 2025, we are returning Mergekit to the GNU Lesser General Public License v3.

Open-Source Toolkits
October 6, 2025

IBM Research Uses Arcee MergeKit in Granite 4.0 Model Development

MergeKit helped evaluate merged checkpoints and select top performers, reflecting a broader enterprise shift to open models & reproducible tooling.

Research
September 30, 2025

Optimizing Arcee Foundation Models on Intel CPUs

Explore how to optimize small language models on Intel’s latest CPU, utilizing Arcee AI’s AFM-4.5B and Intel-optimized inference libraries.

Company
July 30, 2025

Arcee AI Secures Strategic Investment to Accelerate Enterprise-Grade AI Platform Built on AFM Foundation Models

Prosperity7 Ventures, M12, Hitachi Ventures, JC2, Wipro, Samsung, and Guidepoint are now backing Arcee AI.

Open-Source SLMs
July 29, 2025

Announcing the Official Launch of AFM-4.5B

Today, we’re excited to officially take AFM-4.5B out of preview and release the weights of both AFM-4.5B and AFM-4.5B-Base on HuggingFace. This marks a major milestone for our team at Arcee AI as we open up access to a new, enterprise-grade language model designed for both flexibility and performance across a wide range of deployment environments.

News
July 22, 2025

Seed Group teams up with Arcee AI to revolutionise enterprise AI innovation in Dubai and the wider MENA region

Seed Group, a company of The Private Office of Sheikh Saeed bin Ahmed Al Maktoum, has started a joint venture with Arcee AI to advance the deployment of enterprise-grade artificial intelligence across the UAE.

Open-Source SLMs
July 18, 2025

Arcee AI Models Excel Across Yupp.ai Leaderboards

Small and mighty!

Research
July 9, 2025

Is Running Language Models on CPU Really Viable?

Running language models on CPUs has been discussed for some time, but delivering accurate results with production-level performance remains unproven. So, is using CPUs for language models truly viable in production?