Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 6 Current »

While creating a customized agent, you will need to select a model (also known as LLM - Large Language Model), which will take care of a given task, such as retrieving information from the web or a data source to generate an answer relevant to the user’s query, analyzing data, summarizing content, creating content and much more.

Therefore it is crucial to choose the right LLM for your use case. This page will give you some insights to help you select a model from the ones we offer.

Available LLMs

Below is the list of available LLMs with their main characteristics :

Model name

Provider

Context window

(Max number of tokens)

GPT 4o (EU)

Azure

128K

GPT 4o (East US)

Azure

128k

GPT 4o (South US)

Azure

128k

GPT 4o mini (EU)

Azure

128k

GPT 4o mini (East US)

Azure

128k

GPT 4o mini (South US)

Azure

128k

Mistral Mini (EU)

Azure

128k

Mistral Large (EU)

Azure

128k

Mistral Mini (US)

Azure

128k

Mistral Large (US)

Azure

128k

Claude Haiku (US)

Anthropic

200k

Claude Sonnet (US)

Anthropic

200k

Claude Haiku (EU)

AWS

200k

Claude Sonnet (EU)

AWS

200k

Claude Haiku (US)

AWS

200k

Claude Sonnet (US)

AWS

200k

Gemini Flash (US)

Google

250k

Gemini Pro (US)

Google

250k

Llama 3.1 70B (US)

AWS

128k

Llama 3.1 8B (US)

AWS

128k

Mistral Mini (EU)

Mistral

128k

Mistral Large (EU)

Mistral

128k

Mistral Small (EU)

Mistral

128k

Selection criteria

need more info from comparison tests (Alban will do them when he has time)

Selecting an LLM for an agent depends on the following criteria.

  • the type of task done by the agent. Some LLMs excel in specific tasks like …. for example.

  • the context window of the LLM. This corresponds to the total number of tokens in input and output that can be processed by the LLM.

Some LLMs have larger context windows than others, which might be interesting if you intend to process substantial documents. On the opposite, if you want to process smaller documents, you might want to choose an LLM with a smaller context window.

The size of the context window also matters if you want your agent to retain more information from earlier in the conversation.

  • the price for the processed tokens. Some LLMs are more expensive than others.

  • the response time of the LLM. Some LLMs answer faster than others.

Best performing models

Below are our insights on the best-performing models. (might be outdated)

Overall,

  • GPT 4o is the best-performing model.

  • GPT 4o mini has good performance and is very fast and cheap.

  • Mistral Large is slightly worse than GPT 4o but slightly cheaper.

  • The Groq models are ideal when the text to generate is long.

  • No labels