Skip to main content

LLM Configuration

Before you start with Rubra, configure the models you want Rubra to access by editing the llm-config.yaml file.

The models currently supported:

  • OpenAI
    • GPT-4-turbo (gpt-4-1106-preview)
  • Anthropic
    • claude-2.1
  • Local Models

For OpenAI and Anthropic models, you must provide your API Key for the respective provider.

Your config file should look like this:

model_list:
- model_name: my-gpt-4-turbo
litellm_params:
model: gpt-4-1106-preview
api_key: "OPENAI_API_KEY" # Replace with your OpenAI API Key
custom_llm_provider: "openai"

- model_name: name-this-what-you-want-claude-2.1
litellm_params:
model: claude-2.1
api_key: "CLAUDE_API_KEY" # Replace with your Anthropic API Key

# The following is for locally running LLMs. Do not modify
- model_name: custom
litellm_params:
model: openai/custom
api_base: "http://host.docker.internal:1234/v1" # host.docker.internal allows docker to use your local machine's IP address (localhost)
api_key: "None"
custom_llm_provider: "openai"

Edit the model list to include the models you want to use. You need to specify at least one model to use Rubra.

Architecture Diagram We currently use LiteLLM as the chat completions server. This may change in the future.