At Arcee AI, we're committed to democratizing access to cutting-edge AI through open-source collaboration.
Our goal is to empower developers and organizations to build, adapt, and improve state-of-the-art AI models that deliver immediate ROI.
Our open-source models provide a robust foundation for enterprise customers to build and tailor AI solutions specific to their business needs.
A 14B parameter model designed for general-purpose use cases
Learn moreAn 8-billion parameter model distilled from our flagship SuperNova model
Learn moreA 7 billion parameter model designed for general -purpose use cases
Learn moreA 7 billion parameter model expertly designed for both Arabic and English
Learn moreA 7 billion parameter model expertly designed for function-calling and tool-use
Learn moreA compact 1.5 billion parameter language model for resource-constrained environments
Learn moreA 72 billion parameter language model for general-purpose use cases
Learn moreA versatile chat model that specializes in creative writing and functions as a writing assistant
Learn moreA 70 billion parameter domain-specific model for SEC data analysis
Learn more10B-parameter model and excels in a variety of tasks, demonstrating advanced reasoning, code generation, and mathematical problem-solving capabilities.
Learn more32B distillation of DeepSeek-V3, which has delivered our highest scores yet on public benchmarks and performs well in advanced chatbots and enterprise data analysis.
Learn moreYou can deploy and test our models on these platforms
Our open-source libraries provide easy-to-use tools for researching, exploring, and enhancing the adoption of advanced model training techniques.
A framework for model distillation, allowing smaller models to learn from larger, more knowledgeable ones.
A library for evolutionary optimization, facilitating the discovery of better models through iterative refinement.
A pre-training methodology for models, enhancing their performance and adaptability.
Check out the latest publications by the Arcee AI team
Learn more about MergeKit, an open-source library enabling efficient model merging to create multitask models without additional training, preserving original capabilities while enhancing AI performance and versatility.
This is a preprint technical report with thorough evaluations to understand the entire process of domain adaptation, continual pre-training, and model merging for enhancing language model performance on financial regulatory data.
This paper examines model merging techniques, from simple averaging methods like Model Soups to advanced approaches like DARE and TIES-Merging, and introduces Differentiable Adaptive Merging (DAM) for efficient model integration.