We break down the reactions to the LLama 4 release–analyzing the missteps but also highlighting strengths that may have been overlooked amidst the initial noise.
Our Chief Evangelist, Julien Simon, explores the advantages and practical applications of running SLM inference on Arm CPUs.
Facing growing, unpredictable AI budgets? Arcee Conductor intelligently routes prompts to the optimal AI model based on complexity, cutting costs by up to 99% per prompt without sacrificing quality. Beyond a simple LLM router, it offers a comprehensive catalog of both SLMs and LLMs.
Why does the model or (models) used in your agentic AI workflows matter? In this article we explain why the choice of models is key for a successful agentic AI strategy–and why small language models (SLMs) are the right choice.
Today we bring you exciting updates on two small language models (SLMs) we've been working on: our first reasoning model, Arcee-Maestro-7B-Preview, and a fast and efficient Mistral-based DeepSeek distillation we call Arcee-Blitz.
AngelQ and Arcee AI launch KidRails, the first open-source framework for training LLMs to deliver safe, age-appropriate responses for children aged 5-12.
With Arcee Orchestra, you’ll transform complexity into simplicity, see faster collaboration across departments, and ultimately drive business growth—one automated workflow at a time.
MergeKit changed the game when it came to model merging, and today we're excited to bring you some game-changing updates to MergeKit–with what we're calling MergeKit v0.1. Starting today, you'll be able to unlock the power of model merging more than ever, with enterprise hosting, premium features, and expert support.
Discover how knowledge distillation makes AI models faster, more efficient, and cost-effective without sacrificing performance. Learn how Arcee AI leverages this technique to optimize models like Virtuoso Lite and Virtuoso-Medium-v2, delivering powerful AI solutions with lower computational demands. Explore the benefits, use cases, and how your organization can implement knowledge distillation to scale AI performance while reducing costs.