Try our new intelligent model routing solution, Arcee Conductor. Sign up today and get a $20 credit.
Company
The first release—AFM-4.5B—is a 4.5-billion-parameter model that delivers excellent accuracy, strict compliance, and very high cost-efficiency.
Today, we’re thrilled to unveil the Arcee Foundation Models, a new family of generative AI models built from the ground up for enterprise reality. The first release—AFM-4.5B—is a 4.5-billion-parameter frontier model that delivers excellent accuracy, strict compliance, and very high cost-efficiency. In short: enterprise-grade intelligence that can run anywhere—on a smartphone, at the edge, or in the cloud.
For a quick taste, you can test AFM-4.5B in our playground and on Together.ai.
For a deeper dive into the model’s training pipeline and benchmarks, details are available in our technical blog post.
In short, because our customers have told us they need it. Over the last 12 months we have met with more than 150 companies - from the Fortune 100 to AI startups like ourselves - to understand their challenges and evaluate how AI and small language models (SLMs) can solve them. Across hundreds of conversations and active collaborations with our customers, we repeatedly heard similar common roadblocks in their path to adopt generative AI.
Many of our enterprise customers adopted large language models (LLMs) from providers such as OpenAI, Anthropic, and DeepSeek due to their ease of use, speed of deployment, and broad general capabilities. However, these models present significant challenges: they are expensive to operate at scale and difficult, or prohibitively costly, to customize for use cases that require in-depth domain knowledge. They also raise substantial concerns around data privacy, IP liability, and regulatory compliance, as most, if not all, LLMs are tainted with copyrighted or paywalled data sources, which may expose the business to legal or reputational risk.
In response, some of our customers explored small, open-weight language models (SLMs) like Llama, Mistral, or Qwen. These models offer greater flexibility, lower inference costs, and the potential for fine-tuning on proprietary data, thereby addressing some of the challenges posed by proprietary large language models (LLMs). Unfortunately upon further inspection, these open models came with trade-offs: licensing restrictions, intellectual property concerns, and lingering doubts about safety and sovereignty, especially with models developed in China.
Our customers asked us to help them achieve cost efficiency without sacrificing performance, model customizability for their domain and data, and enterprise compliance without compromise. That's why we built Arcee Foundation Model.
We designed AFM-4.5B from the ground up to be a "no-trade-offs" model. We embedded cost efficiency, customizability, and compliance into the model, rather than adding them on afterward. Because our research team owns the full stack—from data sourcing to training infrastructure to deployment tools—we were able to iterate rapidly and make thoughtful, intentional choices at every stage.
The result is a model that delivers business performance comparable to much larger models at vastly lower hosting costs, while being efficient enough to run on low-RAM GPUs or even CPUs.
Thanks to its open architecture and open weights, you can deploy AFM-4.5B in any environment (on-prem, cloud, or hybrid), using popular open-source frameworks like Hugging Face Transformers, vLLM, or llama.cpp, and while maintaining full control and ownership.
Combined with built-in support for function calling and agentic reasoning, AFM-4.5B is ready to automate complex workflows immediately—no fragile prompt engineering required. From a business perspective, this means faster deployment, lower total cost of ownership (TCO), and measurable returns without compromising on quality, compliance, or sovereignty.
AFM-4.5B was trained on almost 7 trillion tokens of clean, rigorously filtered data. Tremendous effort was put towards excluding copyrighted books and material with unclear licensing. Training on this high-quality data ensures fewer hallucinations, better factual accuracy, and a much lower risk of intellectual property issues.
AFM-4.5B has been tested across a wide range of languages, including Arabic, English, French, German, Hindi, Italian, Korean, Mandarin, Portuguese, Russian, and Spanish—delivering strong performance in each. Thanks to our modular training and post-training architecture, adding support for additional languages or dialects is straightforward and can be achieved through lightweight customization per customer needs. Whether you're deploying in multilingual markets or building domain-specific solutions, AFM-4.5B is ready to meet you where you operate.
Licensing is transparent and flexible: non-commercial use will be available in the next few weeks on Hugging Face under a CC-BY-NC license, allowing you to test and experiment immediately at no cost. Commercial deployment is supported through direct licensing or white-label options, providing customers with full control, branding flexibility, and model upgrades when they become available. Please contact sales@arcee.ai for more information.
AFM-4.5B is part of a scalable family of Arcee Foundation Models, with future variants already on the roadmap. All models will share the same DNA and licensing model, enabling enterprises to scale up or down without requiring code rewriting, platform switching, or renegotiation of terms. AI teams will benefit from a consistent deployment and compliance story, whether they’re running ultra-compact models on edge devices or high-capacity versions on GPU clusters.
This model family is about time-to-value. Our domain-adaptation pipeline can produce high-quality, industry-specific checkpoints in days, not quarters, helping organizations deploy tailored models for verticals such as legal, life sciences, or industrial automation without lengthy R&D cycles.
In summary, the AFM models provide a production-grade foundation with the cost efficiency, compliance, and customizability that enterprises have been seeking. If AI is central to your business, why settle for trade-offs?
Give AFM-4.5B a try in our playground and let’s talk!
Building AFM was a company-wide effort, and we’d like to thank the extended Arcee AI team for their contribution: Fernando Fernandes, Varun Singh, Charles Goddard, Lucas Atkins, Mark McQuade, Maziyar Panahi, Conner Stewart, Colin Kealty, Raghav Ravishankar, Lucas Krauss, Anneketh Vij, Pranav Veldurthi, Abhishek Thakur, Julien Simon, Scott Zembsch, Benjamin Langer, Aleksiej Cecocho and Maitri Patel.