Search Docs…

Guide

ChatGPT & Llama

ChatGPT & Llama2

Odyssey supports two different large language models: ChatGPT and Llama. ChatGPT calls OpenAI's API and has a cost associated with each API call. Llama can be run directly on your Mac for free. This section will take a closer look at how to use the different models.

Both LLM nodes are a work in progress.

GPT

To use ChatGPT within Odyssey, you will first need to add your OpenAI API key. You can find your API key in your OpenAI account at this link.

Once you add your API key, drag the GPT node and the language prompt node onto your canvas. Odyssey will regularly update the ChatGPT API to support the latest models.

ChatGPT support in Odyssey

Keep in mind that each API call to OpenAI incurs a cost so using the GPT node can get expensive. Find more details on advanced settings below.

Llama

Llama is Meta's open source LLM that you can run locally on your computer. This means that there are no costs associated with Llama and any data you send to Llama will be completely private.

Here's an overview of the different models Odyssey currently supports

Llama 2 - 7B
Mode: Chat
Purpose: A chat-based model designed for building an assistant or chatbot
Parameters: 7B
Context length: 4k

Alpaca Cielo 2
Mode: Chat
Purpose: A chat-based model designed for creative tasks, such as storytelling and role-play
Parameters: 13B
Context length: 4k

Llama 2 (Instruct)
Mode: Instructions
Purpose: Processing large amounts of data and following instructions, such as providing summaries and answering questions.
Parameters: 7B
Context length: 32k

Marcoroni
Mode: Instructions
Purpose: A range of tasks including language modeling, text generation, and text augmentation
Parameters: 7B
Context length: 32k

To get started, drag the Llama node onto the Odyssey canvas. On the right hand panel, there are a few settings you can choose from:

  • Mode dictates whether you want to run the model in Chat mode or Instructions mode. Chat mode is best for conversational interactions while instructions mode can help do things like analyze long strings of text or generally help you be more productive. Depending on which mode you select, there will be different options for models

  • Model lets you choose from a range of open sourced models. The default models are Llama2 - however you can download additional models by navigating to Settings > Downloads.

Chat mode

The Odyssey canvas with a Llama2 node

Chat mode has slightly different configurable settings and UI patterns than Instructions mode. To chat with an AI model, simply connect a prompt node to the Llama node.

In the instructions box input how you would like the chatbot to behave. We start with a default set of instructions that guide the model to a positive interaction. Let's look at what happens when the instructions change to having the model behave like a pirate.

The Odyssey canvas with a Llama2 workflow

You can thread chats and have the model remember your history. This will allow you to have long-form, continuous conversations with any chatbot you create.

Instructions

Instructions mode behaves slightly differently than Chat mode. Rather than have a set of system instructions, Instructions mode takes a single input as context.

In future versions of Odyssey, we'll begin adding nodes that enable you to input large amounts of data into the Llama node.

Here's an example of when we put Shirley Jackson's The Lottery in as context and ask Llama to summarize it.

Llama2 summarizes Shirley Jackson's seminal short story The Lottery

One thing to keep in mind is that models are running on your Mac for free. We've optimized them to run as fast as they can but there might be some latency when you first start chatting or when you insert large amounts of context.

Advanced settings

Both chat mode and instructions mode have a shared set of Advanced settings.

Length

  • Context window - A token is the basic unit of text that a large language model can process. Humans read text word by word, but language models break text up into tokens. 1 token is about 3/4 of an English word. The context window is the maximum number of tokens a model can process at a time. You can think of it as how much the model can remember. Every time you provide a prompt, Odyssey takes the chat history (i.e. all the prompts and responses produced thus far) and adds the prompt to the history before giving the history to the model. However, if the number of tokens in the history exceeds the context window, Odyssey needs to edit the chat history by removing some of the earlier tokens, thereby effectively editing the model's memory.

    If you don't want the model to remember anything, you can disable Remember history. It's disabled by default for Instruction.

  • Enable infinite length does not restrain how long the response from the model will be. If you turn this setting off, you will be able to specify the maximum length of the model's response

Sampling

  • Preset dictates how the model will respond and influence the temperature, top P, Top K, and Tail-free sampling. You can pick from balanced, creative, precise, and custom

  • Temperature determines the randomness of the response. If you turn the

    temperature up, you'll get more random and unexpected responses. If you turn it down, you'll get more predictable and focused responses. A temperature of 0 will result in identical output across successive runs

  • Top P - An LLM produces a new word by looking at a sentence or partial sentence and determining which word would make the most sense as the next word in the sentence.

    This calculation involves choosing the best option from a number of guesses or probable next words. The Top P value instructs the LLM to limit those guesses to a percentage of its total available guesses. For example, a value of 0.1 means the LLM should only pick from the top 10% of its available guesses. In other words, of the 1,000 most likely words, choose one word from the 100 at the top of the list

  • Top K is similar to Top P, but, instead of using a percentage, it uses a fixed number.

    In other words, if the LLM has 100 possible words from which to choose the next word in a sentence, a Top K number of 5 means it has to choose 1 word from the 5 best words in its list of 100 words

  • Tail-free sampling (TFS) is a technique for generating text that reduces the likelihood that less likely tokens (or words) will be included in the output. To determine the most probable words from which to choose, TFS filters out low-probability words by using the second derivative of their respective probability values. Once the sum of the tokens' second derivatives reaches the value represented by TFS, the model stops adding words to the list of probable words from which it will choose its next word. A value of 1 is the same as including all words and, therefore, will disable TFS

Repetition

  • Repeat Last N determines the number of tokens (or words) to use when penalizing repetitions. The larger the value, the further back in the generated text the model will look to prevent repetitions. A value of zero effectively disables the repetition penalty

  • Repetition Penalty helps prevent repetitive or monotonous output. Higher values will penalize repetitions more severely, while lower values will be less strict

  • Presence Penalty discourages the model from reusing words that have already appeared in the generated text. It encourages the model to explore new topics and makes it less repetitive. For instance, if we're discussing excellent writers, a higher presence penalty would reduce the chance of multiple mentions of, say, Stephen King and instead encourage the model to mention Terry Pratchett, Cormac McCarthy, and Carl Hiaasen as well

  • Frequency Penalty discourages the model from frequently reusing words by looking at the how many times a word or phrase has already appeared in the generated text. For example, when asked to write an essay and given a high frequency penalty, the model will be less inclined to reach for a phrase such as for example if it has already used it in the generated text

  • Penalize New Lines allows the model to consider new-line characters (I.e. ) when penalizing repetitions

ChatGPT & Llama2

Odyssey supports two different large language models: ChatGPT and Llama. ChatGPT calls OpenAI's API and has a cost associated with each API call. Llama can be run directly on your Mac for free. This section will take a closer look at how to use the different models.

Both LLM nodes are a work in progress.

GPT

To use ChatGPT within Odyssey, you will first need to add your OpenAI API key. You can find your API key in your OpenAI account at this link.

Once you add your API key, drag the GPT node and the language prompt node onto your canvas. Odyssey will regularly update the ChatGPT API to support the latest models.

ChatGPT support in Odyssey

Keep in mind that each API call to OpenAI incurs a cost so using the GPT node can get expensive. Find more details on advanced settings below.

Llama

Llama is Meta's open source LLM that you can run locally on your computer. This means that there are no costs associated with Llama and any data you send to Llama will be completely private.

Here's an overview of the different models Odyssey currently supports

Llama 2 - 7B
Mode: Chat
Purpose: A chat-based model designed for building an assistant or chatbot
Parameters: 7B
Context length: 4k

Alpaca Cielo 2
Mode: Chat
Purpose: A chat-based model designed for creative tasks, such as storytelling and role-play
Parameters: 13B
Context length: 4k

Llama 2 (Instruct)
Mode: Instructions
Purpose: Processing large amounts of data and following instructions, such as providing summaries and answering questions.
Parameters: 7B
Context length: 32k

Marcoroni
Mode: Instructions
Purpose: A range of tasks including language modeling, text generation, and text augmentation
Parameters: 7B
Context length: 32k

To get started, drag the Llama node onto the Odyssey canvas. On the right hand panel, there are a few settings you can choose from:

  • Mode dictates whether you want to run the model in Chat mode or Instructions mode. Chat mode is best for conversational interactions while instructions mode can help do things like analyze long strings of text or generally help you be more productive. Depending on which mode you select, there will be different options for models

  • Model lets you choose from a range of open sourced models. The default models are Llama2 - however you can download additional models by navigating to Settings > Downloads.

Chat mode

The Odyssey canvas with a Llama2 node

Chat mode has slightly different configurable settings and UI patterns than Instructions mode. To chat with an AI model, simply connect a prompt node to the Llama node.

In the instructions box input how you would like the chatbot to behave. We start with a default set of instructions that guide the model to a positive interaction. Let's look at what happens when the instructions change to having the model behave like a pirate.

The Odyssey canvas with a Llama2 workflow

You can thread chats and have the model remember your history. This will allow you to have long-form, continuous conversations with any chatbot you create.

Instructions

Instructions mode behaves slightly differently than Chat mode. Rather than have a set of system instructions, Instructions mode takes a single input as context.

In future versions of Odyssey, we'll begin adding nodes that enable you to input large amounts of data into the Llama node.

Here's an example of when we put Shirley Jackson's The Lottery in as context and ask Llama to summarize it.

Llama2 summarizes Shirley Jackson's seminal short story The Lottery

One thing to keep in mind is that models are running on your Mac for free. We've optimized them to run as fast as they can but there might be some latency when you first start chatting or when you insert large amounts of context.

Advanced settings

Both chat mode and instructions mode have a shared set of Advanced settings.

Length

  • Context window - A token is the basic unit of text that a large language model can process. Humans read text word by word, but language models break text up into tokens. 1 token is about 3/4 of an English word. The context window is the maximum number of tokens a model can process at a time. You can think of it as how much the model can remember. Every time you provide a prompt, Odyssey takes the chat history (i.e. all the prompts and responses produced thus far) and adds the prompt to the history before giving the history to the model. However, if the number of tokens in the history exceeds the context window, Odyssey needs to edit the chat history by removing some of the earlier tokens, thereby effectively editing the model's memory.

    If you don't want the model to remember anything, you can disable Remember history. It's disabled by default for Instruction.

  • Enable infinite length does not restrain how long the response from the model will be. If you turn this setting off, you will be able to specify the maximum length of the model's response

Sampling

  • Preset dictates how the model will respond and influence the temperature, top P, Top K, and Tail-free sampling. You can pick from balanced, creative, precise, and custom

  • Temperature determines the randomness of the response. If you turn the

    temperature up, you'll get more random and unexpected responses. If you turn it down, you'll get more predictable and focused responses. A temperature of 0 will result in identical output across successive runs

  • Top P - An LLM produces a new word by looking at a sentence or partial sentence and determining which word would make the most sense as the next word in the sentence.

    This calculation involves choosing the best option from a number of guesses or probable next words. The Top P value instructs the LLM to limit those guesses to a percentage of its total available guesses. For example, a value of 0.1 means the LLM should only pick from the top 10% of its available guesses. In other words, of the 1,000 most likely words, choose one word from the 100 at the top of the list

  • Top K is similar to Top P, but, instead of using a percentage, it uses a fixed number.

    In other words, if the LLM has 100 possible words from which to choose the next word in a sentence, a Top K number of 5 means it has to choose 1 word from the 5 best words in its list of 100 words

  • Tail-free sampling (TFS) is a technique for generating text that reduces the likelihood that less likely tokens (or words) will be included in the output. To determine the most probable words from which to choose, TFS filters out low-probability words by using the second derivative of their respective probability values. Once the sum of the tokens' second derivatives reaches the value represented by TFS, the model stops adding words to the list of probable words from which it will choose its next word. A value of 1 is the same as including all words and, therefore, will disable TFS

Repetition

  • Repeat Last N determines the number of tokens (or words) to use when penalizing repetitions. The larger the value, the further back in the generated text the model will look to prevent repetitions. A value of zero effectively disables the repetition penalty

  • Repetition Penalty helps prevent repetitive or monotonous output. Higher values will penalize repetitions more severely, while lower values will be less strict

  • Presence Penalty discourages the model from reusing words that have already appeared in the generated text. It encourages the model to explore new topics and makes it less repetitive. For instance, if we're discussing excellent writers, a higher presence penalty would reduce the chance of multiple mentions of, say, Stephen King and instead encourage the model to mention Terry Pratchett, Cormac McCarthy, and Carl Hiaasen as well

  • Frequency Penalty discourages the model from frequently reusing words by looking at the how many times a word or phrase has already appeared in the generated text. For example, when asked to write an essay and given a high frequency penalty, the model will be less inclined to reach for a phrase such as for example if it has already used it in the generated text

  • Penalize New Lines allows the model to consider new-line characters (I.e. ) when penalizing repetitions

ChatGPT & Llama2

Odyssey supports two different large language models: ChatGPT and Llama. ChatGPT calls OpenAI's API and has a cost associated with each API call. Llama can be run directly on your Mac for free. This section will take a closer look at how to use the different models.

Both LLM nodes are a work in progress.

GPT

To use ChatGPT within Odyssey, you will first need to add your OpenAI API key. You can find your API key in your OpenAI account at this link.

Once you add your API key, drag the GPT node and the language prompt node onto your canvas. Odyssey will regularly update the ChatGPT API to support the latest models.

ChatGPT support in Odyssey

Keep in mind that each API call to OpenAI incurs a cost so using the GPT node can get expensive. Find more details on advanced settings below.

Llama

Llama is Meta's open source LLM that you can run locally on your computer. This means that there are no costs associated with Llama and any data you send to Llama will be completely private.

Here's an overview of the different models Odyssey currently supports

Llama 2 - 7B
Mode: Chat
Purpose: A chat-based model designed for building an assistant or chatbot
Parameters: 7B
Context length: 4k

Alpaca Cielo 2
Mode: Chat
Purpose: A chat-based model designed for creative tasks, such as storytelling and role-play
Parameters: 13B
Context length: 4k

Llama 2 (Instruct)
Mode: Instructions
Purpose: Processing large amounts of data and following instructions, such as providing summaries and answering questions.
Parameters: 7B
Context length: 32k

Marcoroni
Mode: Instructions
Purpose: A range of tasks including language modeling, text generation, and text augmentation
Parameters: 7B
Context length: 32k

To get started, drag the Llama node onto the Odyssey canvas. On the right hand panel, there are a few settings you can choose from:

  • Mode dictates whether you want to run the model in Chat mode or Instructions mode. Chat mode is best for conversational interactions while instructions mode can help do things like analyze long strings of text or generally help you be more productive. Depending on which mode you select, there will be different options for models

  • Model lets you choose from a range of open sourced models. The default models are Llama2 - however you can download additional models by navigating to Settings > Downloads.

Chat mode

The Odyssey canvas with a Llama2 node

Chat mode has slightly different configurable settings and UI patterns than Instructions mode. To chat with an AI model, simply connect a prompt node to the Llama node.

In the instructions box input how you would like the chatbot to behave. We start with a default set of instructions that guide the model to a positive interaction. Let's look at what happens when the instructions change to having the model behave like a pirate.

The Odyssey canvas with a Llama2 workflow

You can thread chats and have the model remember your history. This will allow you to have long-form, continuous conversations with any chatbot you create.

Instructions

Instructions mode behaves slightly differently than Chat mode. Rather than have a set of system instructions, Instructions mode takes a single input as context.

In future versions of Odyssey, we'll begin adding nodes that enable you to input large amounts of data into the Llama node.

Here's an example of when we put Shirley Jackson's The Lottery in as context and ask Llama to summarize it.

Llama2 summarizes Shirley Jackson's seminal short story The Lottery

One thing to keep in mind is that models are running on your Mac for free. We've optimized them to run as fast as they can but there might be some latency when you first start chatting or when you insert large amounts of context.

Advanced settings

Both chat mode and instructions mode have a shared set of Advanced settings.

Length

  • Context window - A token is the basic unit of text that a large language model can process. Humans read text word by word, but language models break text up into tokens. 1 token is about 3/4 of an English word. The context window is the maximum number of tokens a model can process at a time. You can think of it as how much the model can remember. Every time you provide a prompt, Odyssey takes the chat history (i.e. all the prompts and responses produced thus far) and adds the prompt to the history before giving the history to the model. However, if the number of tokens in the history exceeds the context window, Odyssey needs to edit the chat history by removing some of the earlier tokens, thereby effectively editing the model's memory.

    If you don't want the model to remember anything, you can disable Remember history. It's disabled by default for Instruction.

  • Enable infinite length does not restrain how long the response from the model will be. If you turn this setting off, you will be able to specify the maximum length of the model's response

Sampling

  • Preset dictates how the model will respond and influence the temperature, top P, Top K, and Tail-free sampling. You can pick from balanced, creative, precise, and custom

  • Temperature determines the randomness of the response. If you turn the

    temperature up, you'll get more random and unexpected responses. If you turn it down, you'll get more predictable and focused responses. A temperature of 0 will result in identical output across successive runs

  • Top P - An LLM produces a new word by looking at a sentence or partial sentence and determining which word would make the most sense as the next word in the sentence.

    This calculation involves choosing the best option from a number of guesses or probable next words. The Top P value instructs the LLM to limit those guesses to a percentage of its total available guesses. For example, a value of 0.1 means the LLM should only pick from the top 10% of its available guesses. In other words, of the 1,000 most likely words, choose one word from the 100 at the top of the list

  • Top K is similar to Top P, but, instead of using a percentage, it uses a fixed number.

    In other words, if the LLM has 100 possible words from which to choose the next word in a sentence, a Top K number of 5 means it has to choose 1 word from the 5 best words in its list of 100 words

  • Tail-free sampling (TFS) is a technique for generating text that reduces the likelihood that less likely tokens (or words) will be included in the output. To determine the most probable words from which to choose, TFS filters out low-probability words by using the second derivative of their respective probability values. Once the sum of the tokens' second derivatives reaches the value represented by TFS, the model stops adding words to the list of probable words from which it will choose its next word. A value of 1 is the same as including all words and, therefore, will disable TFS

Repetition

  • Repeat Last N determines the number of tokens (or words) to use when penalizing repetitions. The larger the value, the further back in the generated text the model will look to prevent repetitions. A value of zero effectively disables the repetition penalty

  • Repetition Penalty helps prevent repetitive or monotonous output. Higher values will penalize repetitions more severely, while lower values will be less strict

  • Presence Penalty discourages the model from reusing words that have already appeared in the generated text. It encourages the model to explore new topics and makes it less repetitive. For instance, if we're discussing excellent writers, a higher presence penalty would reduce the chance of multiple mentions of, say, Stephen King and instead encourage the model to mention Terry Pratchett, Cormac McCarthy, and Carl Hiaasen as well

  • Frequency Penalty discourages the model from frequently reusing words by looking at the how many times a word or phrase has already appeared in the generated text. For example, when asked to write an essay and given a high frequency penalty, the model will be less inclined to reach for a phrase such as for example if it has already used it in the generated text

  • Penalize New Lines allows the model to consider new-line characters (I.e. ) when penalizing repetitions