Skip to main content
Skip table of contents

Choosing your LLM model, context-window and algorithm

Chathive offers the possibility to change your LLM model, the context window and the algorithm of how this model is used. This guide will help you decide what configuration is best for your use case.

Default configuration

Whenever you create a project in Chathive your project will be set to these settings:

Setting

Default value

Model

GPT 4o-mini

Context size

8k (growth tier and higher) or 4k (free trial or basic tier)

If you need higher contextsize contact support@chathive.co.

Algorithm

Default

Although these are usually good defaults, you can greatly improve performance by choosing these settings carefully based on your goals.

How to choose an LLM model

All LLM models have its strengths and weaknesses, you will always make a trade-off based on your choice. However, your choice should be decided by these 5 factors:

  1. Accuracy: Models with more parameters and a larger training dataset (like GPT-5) tend to be a lot more accurate than less complex and smaller models (like GPT-5-mini or gpt-5-nano). If accuracy and reasoning is important, choose a more complex model.

  2. Speed of generating responses: More complex models tend to require more compute and are as a result slower in generating text. For example, GPT-5-nano is quicker than GPT-5.

  3. Steerability: More complex models generally follow instructions better than less complex ones. If strict adherence to instructions is important, choose a more steerable model. Although more accurate models are typically more steerable, this is not always the case.

  4. Tool usage: Some models excel at using tools, with more complex and newer models generally performing better.

  5. Price: More complex models also tend to be more expensive to run, and require more message credits to run as a result.

    image-20250815-103659.png

Model comparison

To help you decide, we have put all models into a table for you to compare and make your choice.

Model Group

Model

Credits

๐ŸŽฏ Accuracy

๐Ÿงญ Steerability

๐Ÿ›  Tool Usage

โšก Speed

Reasoning

GPT-5

GPT-5

5

๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ 5

๐Ÿงญ๐Ÿงญ๐Ÿงญ๐Ÿงญ๐Ÿงญ 5

๐Ÿ› ๐Ÿ› ๐Ÿ› ๐Ÿ› ๐Ÿ›  5

โšกโšกโšก 3

โœ“ (minimal by default)

GPT-5 Mini

3

๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ 4

๐Ÿงญ๐Ÿงญ๐Ÿงญ๐Ÿงญ 4

๐Ÿ› ๐Ÿ› ๐Ÿ› ๐Ÿ›  4

โšกโšกโšกโšก 4

โœ“ (minimal by default)

GPT-5 Nano

1

๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ 3

๐Ÿงญ๐Ÿงญ๐Ÿงญ 3

๐Ÿ› ๐Ÿ› ๐Ÿ› ๐Ÿ›  4

โšกโšกโšกโšกโšก 5

โœ“ (minimal by default)

GPT-4o

GPT-4o

5

๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ 4

๐Ÿงญ๐Ÿงญ๐Ÿงญ 3

๐Ÿ› ๐Ÿ› ๐Ÿ›  3

โšกโšก 2

โœ—

GPT-4o Mini

1

๐ŸŽฏ๐ŸŽฏ 2

๐Ÿงญ๐Ÿงญ 2

๐Ÿ› ๐Ÿ›  2

โšกโšกโšกโšก 4

โœ—

GPT-4.1

GPT-4.1

4

๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ 3

๐Ÿงญ๐Ÿงญ๐Ÿงญ๐Ÿงญ 4

๐Ÿ› ๐Ÿ› ๐Ÿ› ๐Ÿ›  4

โšกโšก 2

โœ—

GPT-4.1 Mini

3

๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ 3

๐Ÿงญ๐Ÿงญ๐Ÿงญ 3

๐Ÿ› ๐Ÿ› ๐Ÿ›  3

โšกโšกโšก 3

โœ—

GPT-4.1 Nano

0.5

๐ŸŽฏ๐ŸŽฏ 2

๐Ÿงญ๐Ÿงญ 2

๐Ÿ› ๐Ÿ› ๐Ÿ›  3

โšกโšกโšกโšก 4

โœ—

Reasoning

o3 Mini

12

๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ๐ŸŽฏ 4

๐Ÿงญ๐Ÿงญ 2

๐Ÿ› ๐Ÿ› ๐Ÿ›  3

โšก 1

โœ“ (medium by default)

Deprecated

GPT-4

60

๐ŸŽฏ 1

๐Ÿงญ 1

๐Ÿ›  1

โšก 1

โœ—

GPT-4 Turbo

10

๐ŸŽฏ๐ŸŽฏ 2

๐Ÿงญ 1

๐Ÿ› ๐Ÿ›  2

โšกโšก 2

โœ—

We generally recommend to use GPT-5 for most use cases that require a high level of accuracy. And recommend GPT-5 nano if you need quick replies and/or if you have a very large volume and need to cut costs.

Choosing your context window

This can only be changed by Chathive team members, request via support if you want to change this.

Chathive allows you to choose how large of a context window to use for each message. A larger context window allows a larger amount of text from your sources for each message. It also allows longer questions to and longer replies from the AI assistant.

Larger context window for higher accuracy

We mostly recommend increasing context window for increasing accuracy of your AI assistant. With a larger context window we can provide larger snippets of text from your sources in each response. This means the model has more input to work with and tends to improve accuracy.

These are the ways a larger context window is more accurate:

  • More complete snippets from your sources: Smaller context windows only allow us to include small snippets of your training data as sources. Larger context windows allow us to include larger snippets, thus improving the chance that the full needed context is there for creating the answer.

  • Using more sources: Smaller context windows also limit the amount of sources we can include. And as a result have more chance of not including the right sources needed.

  • Long questions, hogging up space for sources: The longer the question of the user the less space we have for questions. If your questions are large, it will greatly reduce performance in smaller context sizes.

  • More conversation history: To fit as many sources as we can, we cut the conversation history short if we need that space for the sources. This results in the model knowing less of the past conversation when answering the question. Increasing context size will let the AI assistant retain more memory of the conversation.

  • Longer responses: Sometimes you need a long explanation for complex cases, more context allows the AI assitant. to create longer responses.

Changing context size does have pricing implications!

Choosing your algorithm

The last step is to decide on the algorithm to use. Algorithms are ways the model and context window is used to find sources and generate responses.

Currently, we offer 2 distinct algorithms:

  • Default: Use the question to find sources and feed those into the model.

  • Fusion algorithm (recommended): Rephrases the question a few times and uses all those rephrasings to search the database. This algorithm greatly improves the search results if the user uses different words or phrasings than are used in your AI database. It slightly increases the time it takes for the AI to start responding, as it first has to rephrase and only then can start responding.

Screenshot 2024-04-15 at 16.38.26.png

You can view the pricing implications at our mesage credits documentation as well.

JavaScript errors detected

Please note, these errors can depend on your browser setup.

If this problem persists, please contact our support.