Skip to content

Latest commit

 

History

History
473 lines (300 loc) · 12.3 KB

models.md

File metadata and controls

473 lines (300 loc) · 12.3 KB

Models

Table of Contents

  1. Model Aliases
  2. Model Alias Values

Model Aliases

To simplify using LLMInterface.sendMessage(), you can use the following model aliases:

  • default
  • large
  • small
  • agent

If no model is passed, the system will use the default model for the LLM provider. If you'd prefer to specify your model by size instead of name, pass large or small.

Aliases can simplify working with multiple LLM providers letting you call different providers with the same model names out of the box.

const response = await LLMInterface.sendMessage("openai", "Explain the importance of low latency LLMs", { model: "small" });
const geminiResult = await LLMInterface.sendMessage("gemini", "Explain the importance of low latency LLMs", { model: "small" });

Changing the aliases is easy:

LLMInterface.setModelAlias("openai", "default", "gpt-4o-mini");

Model Alias Values

ai21

  • default: jamba-instruct
  • large: jamba-instruct
  • small: jamba-instruct
  • agent: jamba-instruct

  • default: Llama-2-70b
  • large: Qwen/Qwen1.5-72B-Chat
  • small: alpaca-7b
  • agent: Llama-2-70b

aimlapi

  • default: gpt-3.5-turbo-16k
  • large: Qwen/Qwen1.5-72B-Chat
  • small: Qwen/Qwen1.5-0.5B-Chat
  • agent: gpt-4-32k-0613

anyscale

  • default: mistralai/Mixtral-8x22B-Instruct-v0.1
  • large: meta-llama/Llama-3-70b-chat-hf
  • small: mistralai/Mistral-7B-Instruct-v0.1
  • agent: mistralai/Mixtral-8x22B-Instruct-v0.1

anthropic

  • default: claude-3-sonnet-20240229
  • large: claude-3-opus-20240229
  • small: claude-3-haiku-20240307
  • agent: claude-3-sonnet-20240229

cloudflareai

  • default: @cf/meta/llama-3-8b-instruct
  • large: @hf/thebloke/llama-2-13b-chat-awq
  • small: @cf/tinyllama/tinyllama-1.1b-chat-v1.0
  • agent: @cf/meta/llama-3-8b-instruct

cohere

  • default: command-r
  • large: command-r-plus
  • small: command-light
  • agent: command-r-plus

corcel

  • default: gpt-4-turbo-2024-04-09
  • large: gpt-4o
  • small: cortext-lite
  • agent: gemini-pro

deepinfra

  • default: openchat/openchat-3.6-8b
  • large: nvidia/Nemotron-4-340B-Instruct
  • small: microsoft/WizardLM-2-7B
  • agent: Qwen/Qwen2-7B-Instruct

deepseek

  • default: deepseek-chat
  • large: deepseek-chat
  • small: deepseek-chat
  • agent: deepseek-chat

  • default: accounts/fireworks/models/llama-v3-8b-instruct
  • large: accounts/fireworks/models/llama-v3-70b-instruct
  • small: accounts/fireworks/models/phi-3-mini-128k-instruct
  • agent: accounts/fireworks/models/llama-v3-8b-instruct

forefront

  • default: forefront/Mistral-7B-Instruct-v0.2-chatml
  • large: forefront/Mistral-7B-Instruct-v0.2-chatml
  • small: forefront/Mistral-7B-Instruct-v0.2-chatml
  • agent:

  • default: mixtral-8x7b-instruct-v0-1
  • large: meta-llama-3-70b-instruct
  • small: meta-llama-3-8b-instruct
  • agent: gemma-7b-it

  • default: gemini-1.5-flash
  • large: gemini-1.5-pro
  • small: gemini-1.5-flash
  • agent: gemini-1.5-pro

gooseai

  • default: gpt-neo-20b
  • large: gpt-neo-20b
  • small: gpt-neo-125m
  • agent: gpt-j-6b

  • default: llama3-8b-8192
  • large: llama3-70b-8192
  • small: gemma-7b-it
  • agent: llama3-8b-8192

  • default: meta-llama/Meta-Llama-3-8B-Instruct
  • large: meta-llama/Meta-Llama-3-8B-Instruct
  • small: microsoft/Phi-3-mini-4k-instruct
  • agent: meta-llama/Meta-Llama-3-8B-Instruct

  • default: hive
  • large: gpt-4o
  • small: gemini-1.5-flash
  • agent: gpt-4o

lamini

  • default: meta-llama/Meta-Llama-3-8B-Instruct
  • large: meta-llama/Meta-Llama-3-8B-Instruct
  • small: microsoft/phi-2
  • agent: meta-llama/Meta-Llama-3-8B-Instruct

  • default: gpt-3.5-turbo
  • large: gpt-3.5-turbo
  • small: gpt-3.5-turbo
  • agent: openhermes

mistralai

  • default: mistral-large-latest
  • large: mistral-large-latest
  • small: mistral-small-latest
  • agent: mistral-large-latest

monsterapi

  • default: meta-llama/Meta-Llama-3-8B-Instruct
  • large: google/gemma-2-9b-it
  • small: microsoft/Phi-3-mini-4k-instruct
  • agent: google/gemma-2-9b-it

neetsai

  • default: Neets-7B
  • large: mistralai/Mixtral-8X7B-Instruct-v0.1
  • small: Neets-7B
  • agent:

  • default: meta-llama/llama-3-8b-instruct
  • large: meta-llama/llama-3-70b-instruct
  • small: meta-llama/llama-3-8b-instruct
  • agent: meta-llama/llama-3-70b-instruct

  • default: nvidia/llama3-chatqa-1.5-8b
  • large: nvidia/nemotron-4-340b-instruct
  • small: microsoft/phi-3-mini-128k-instruct
  • agent: nvidia/llama3-chatqa-1.5-8b

  • default: mistral-7b-instruct
  • large: mixtral-8x22b-instruct
  • small: mistral-7b-instruct
  • agent: mixtral-8x22b-instruct

  • default: llama3
  • large: llama3
  • small: llama3
  • agent:

  • default: gpt-4o-mini
  • large: gpt-4o
  • small: gpt-4o-mini
  • agent: gpt-4o

perplexity

  • default: llama-3.1-sonar-large-128k-chat
  • large: llama-3.1-sonar-large-128k-chat
  • small: llama-3.1-sonar-small-128k-chat
  • agent: llama-3.1-70b-instruct

rekaai

  • default: reka-core
  • large: reka-core
  • small: reka-edge
  • agent: reka-core

replicate

  • default: mistralai/mistral-7b-instruct-v0.2
  • large: meta/meta-llama-3-70b-instruct
  • small: mistralai/mistral-7b-instruct-v0.2
  • agent: meta/meta-llama-3-70b-instruct

shuttleai

  • default: shuttle-2-turbo
  • large: shuttle-2-turbo
  • small: shuttle-2-turbo
  • agent: shuttle-2-turbo

  • default: gpt-4-turbo
  • large: llama-3-70b-chat
  • small: llama-2-7b-chat
  • agent: gpt-4-turbo

togetherai

  • default: google/gemma-7b
  • large: mistralai/Mixtral-8x22B
  • small: google/gemma-2b
  • agent: Qwen/Qwen1.5-14B

  • default: ibm/granite-13b-chat-v2
  • large: meta-llama/llama-3-70b-instruct
  • small: google/flan-t5-xxl
  • agent: meta-llama/llama-3-70b-instruct

writer

  • default: palmyra-x-002-32k
  • large: palmyra-x-002-32k
  • small: palmyra-x-002-32k
  • agent:

  • default: glm-4-airx
  • large: glm-4
  • small: glm-4-flash
  • agent: glm-4