Frequently Asked Questions

Explore how Arcee.ai can transform project development and boost business success.

What's the difference between an SLM and an LLM?

The main difference between an SLM and an LLM is the size. Arcee’s views SLM’s ranging in 7B - 70B parameters in size, with our sweet spot sitting between 7B-13B parameters.

Why shouldn’t we just use a closed-source model, like GPT-4 or Claude?

Closed-source models have several drawbacks for companies that want to build their own use-case specific models.

  • Even though closed-source general purpose LLMs are very powerful, hitting them at scale can be very costly.
  • You are at the mercy of closed-source model providers. When their API goes down, or has an increase in traffic, your performance is negatively affected.
  • Closed-source model training and training datasets are typically a black box. This means you have very little insight into how they were trained or what data they were trained on.
  • Training LLMs with closed-course models exposes your data to third-party platforms, which compromises data security.
  • You never have ownership of a closed-source model. Your model should be your asset. Just as data was the new gold, LLMs should be your new gold. These assets should be owned and controlled by you.

In most cases, smaller, more specialized language models running on your own cloud is a more suitable business case. This is what Arcee is built for.

We have built our own model in-house already, can we only incorporate some of your features?

The short answer is yes. We are open-source model agnostic for the most part so you can utilize your model in our SLM adaptation system. Arcee’s SLM adaptation has been built with a modular focus.

What is so unique about Arcee’s product?

Arcee has built a way to take the power of LLMs and bring them down into smaller, specialized and scalable models. We are able to do this thanks to our unique SLM Adaptation System, where we offer customers pre-training, alignment, and continuous retrieval-augmented generation all in one place. Furthermore, our focus on doing everything 100% in your own cloud via virtual private cloud (VPC) deployment means both your data and your models are 100% owned by your from start to finish.

Do I own my model? Do I own my stack?

Yes, your model will be 100% yours. It runs inside your cloud, inside your VPC and you can wield it anyway you see fit.

Do you pretrain our model from scratch?

We do not go all the way back to pretraining from scratch, but take the domain adaptive continual pre-training approach. This means we start with a great open-source model, such as Mistral or Llama 2, and extend the pretraining of the existing model by injecting new data into it. The end result is a model that holds all the great general capability of the base model, and all the domain knowledge of your injected corpus. This makes for a much more powerful model than pre-training from scratch, at a fraction of the price.

Why do you lean into smaller parameter models over larger parameter models?

If you go to such depths of domain adaptation as we do, then you can be smaller with your model size. We also believe the data and knowledge in large general models are largely unnecessary in many business use cases, and are overkill for the ROI that you are trying to achieve.

How does your pricing work?

Arcee has different tiers for API calls in production with a sliding scale for consumption and support tailored to your use case and needs. Please contact sales@arcee.ai for more details.

Can we train a model with Arcee, then deploy on our own?

Yes. Arcee is all about the company owning the model. We have different packages for pre-training only as well as end-to-end deployment. To get a sense of what you can get from Arcee throughout your journey with us, click here for a visualization of our end-to-end deployment and maintenance lifecycle.

How can we trust that our data will be secure?

One of the core pillars of our product is 100% in-VPC (virtual private cloud) deployment. This means that when you build your SLMs with Arcee, all the way from pre-training to deployment, everything stays in your own cloud. This is the case for both your data AND your SLMs.

Which cloud providers do you work with?

Currently we are focused on serving customers who use AWS. In the near future, we plan on expanding to other cloud providers as well. As we expand we are looking for test cases in other clouds so please reach out if you have a use case outside of AWS.

Do you offer multi-modal models?

For now, we do not offer multi-modal models. However, this is an area that we are actively exploring, so stay tuned for more information in the coming months!

Do you offer multilingual model training?

We do offer multilingual model training, and have done so with some of our customers already. Get in touch with our team with more information about your needs to get a quote for multilingual models.

Start building your SLMs today

Book a call with our team to discuss how Arcee can help you train, deploy, maintain, and continuously improve your smaller, secure, specialized, and scalable language models.

Book a Demo