mirror of
https://github.com/lobehub/lobehub
synced 2026-04-21 17:47:27 +00:00
846 lines
28 KiB
Text
846 lines
28 KiB
Text
---
|
||
title: LobeHub Model Service Providers - Environment Variables and Configuration
|
||
description: >-
|
||
Learn about the environment variables and configuration settings for various
|
||
model service providers like OpenAI, Google AI, AWS Bedrock, Ollama,
|
||
Perplexity AI, Anthropic AI, Mistral AI, Groq AI, OpenRouter AI, and 01.AI.
|
||
tags:
|
||
- Model Service Providers
|
||
- Environment Variables
|
||
---
|
||
|
||
# Model Providers
|
||
|
||
When deploying LobeHub, a rich set of environment variables related to model service providers is provided, allowing you to easily define the model service providers to be enabled in LobeHub.
|
||
|
||
## OpenAI
|
||
|
||
### `ENABLED_OPENAI`
|
||
|
||
- Type:Optional
|
||
- Description:Enables OpenAI as a model provider by default, turns off the OpenAI service when set to `0`
|
||
- Default:`1`
|
||
- Example:`0`
|
||
|
||
### `OPENAI_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for on the OpenAI account page, you can check it out [here](https://platform.openai.com/api-keys)
|
||
- Default: -
|
||
- Example: `sk-xxxxxx...xxxxxx`
|
||
|
||
### `OPENAI_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: If you manually configure the OpenAI interface proxy, you can use this configuration item to override the default OpenAI API request base URL
|
||
- Default: `https://api.openai.com/v1`
|
||
- Example: `https://api.chatanywhere.cn` or `https://aihubmix.com/v1`
|
||
|
||
<Callout type={'warning'}>
|
||
Please check the request suffix of your proxy service provider. Some proxy service providers may
|
||
add `/v1` to the request suffix, while others may not. If you find that the AI returns an empty
|
||
message during testing, try adding the `/v1` suffix and retry.
|
||
</Callout>
|
||
|
||
<Callout type={'info'}>
|
||
Whether to fill in `/v1` is closely related to the model service provider. For example, the
|
||
default address of openai is `api.openai.com/v1`. If your proxy forwards the `/v1` interface, you
|
||
can simply fill in `proxy.com`. However, if the model service provider directly forwards the
|
||
`api.openai.com` domain, then you need to add `/v1` to the URL yourself.
|
||
</Callout>
|
||
|
||
Related discussions:
|
||
|
||
- [Why is the return value blank after installing Docker, configuring environment variables?](https://github.com/lobehub/lobehub/discussions/623)
|
||
- [Reasons for errors when using third-party interfaces](https://github.com/lobehub/lobehub/discussions/734)
|
||
- [No response in chat after filling in the proxy server address](https://github.com/lobehub/lobehub/discussions/1065)
|
||
|
||
### `OPENAI_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `+qwen-7b-chat,+glm-6b,-gpt-3.5-turbo,gpt-4-0125-preview=gpt-4-turbo`
|
||
|
||
The above example adds `qwen-7b-chat` and `glm-6b` to the model list, removes `gpt-3.5-turbo` from the list, and displays the name of `gpt-4-0125-preview` as `gpt-4-turbo`. If you want to disable all models first and then enable specific models, you can use `-all,+gpt-3.5-turbo`, which means only `gpt-3.5-turbo` will be enabled.
|
||
|
||
You can find all current model names in [modelProviders](https://github.com/lobehub/lobehub/tree/main/src/config/modelProviders).
|
||
|
||
## Azure OpenAI
|
||
|
||
If you need to use Azure OpenAI to provide model services, you can refer to the [Deploying with Azure OpenAI](/docs/self-hosting/examples/azure-openai) section for detailed steps. Here, we will list the environment variables related to Azure OpenAI.
|
||
|
||
### `AZURE_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for on the Azure OpenAI account page
|
||
- Default: -
|
||
- Example: `c55168be3874490ef0565d9779ecd5a6`
|
||
|
||
### `AZURE_ENDPOINT`
|
||
|
||
- Type: Required
|
||
- Description: The endpoint of Azure OpenAI, you can find it in the Azure OpenAI account page
|
||
- Default: `-`
|
||
- Example: `https://docs-test-001.openai.azure.com`
|
||
|
||
### `AZURE_API_VERSION`
|
||
|
||
- Type: Required
|
||
- Description: The API version of Azure, following the format YYYY-MM-DD
|
||
- Default: `-`
|
||
- Example: `2023-02-01`, refer to [latest version](https://docs.microsoft.com/zh-cn/azure/developer/javascript/api-reference/es-modules/azure-sdk/ai-translation/translationconfiguration?view=azure-node-latest#api-version)
|
||
|
||
### `AZURE_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `id->deploymentName=displayName` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `gpt-35-turbo->my-deploy=GPT 3.5 Turbo` 或 `gpt-4-turbo->my-gpt4=GPT 4 Turbo<128000:vision:fc>`
|
||
|
||
## Google AI
|
||
|
||
### `GOOGLE_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the Google AI Platform to access Google AI services
|
||
- Default: -
|
||
- Example: `AIraDyDwcw254kwJaGjI9wwaHcdDCS__Vt3xQE`
|
||
|
||
### `GOOGLE_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: If you manually configure the Google API proxy, you can use this configuration item to override the default Google API request base URL
|
||
- Default: `https://generativelanguage.googleapis.com`
|
||
- Example: `https://api.genai.gd.edu.kg/google`
|
||
|
||
### `GOOGLE_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+gemini-1.5-flash-latest,+gemini-1.5-pro-latest`
|
||
|
||
## Vertex AI
|
||
|
||
### `VERTEXAI_CREDENTIALS`
|
||
|
||
- Type: Required
|
||
- Description: A JSON string of your Google Cloud service account key, you can get the key from [here](/docs/usage/providers/vertexai).
|
||
- Default: -
|
||
- Example: `{"type": "service_account", "project_id": "your-gcp-project-id", ...}`
|
||
|
||
### `VERTEXAI_PROJECT`
|
||
|
||
- Type: Optional
|
||
- Description: Your Google Cloud project ID. If not set, it will be obtained from the `project_id` field in `VERTEXAI_CREDENTIALS`.
|
||
- Default: -
|
||
- Example: `your-gcp-project-id`
|
||
|
||
### `VERTEXAI_LOCATION`
|
||
|
||
- Type: Optional
|
||
- Description: The region where your Vertex AI model is located.
|
||
- Default: `global`
|
||
- Example: `us-central1`
|
||
|
||
### `VERTEXAI_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+gemini-1.5-flash-latest,+gemini-1.5-pro-latest`
|
||
|
||
## Anthropic AI
|
||
|
||
### `ANTHROPIC_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied from Anthropic AI
|
||
- Default: -
|
||
- Example: `sk-ant-apixx-xxxxxxxxx-xxxxxxxxxxxxxxxxx-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx-xxxxxxxx`
|
||
|
||
### `ANTHROPIC_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: If you manually configure the Anthropic API proxy, you can use this configuration item to override the default Anthropic API request base URL
|
||
- Default: `https://api.anthropic.com`
|
||
- Example: `https://my-anthropic-proxy.com`
|
||
|
||
## AWS Bedrock
|
||
|
||
### `ENABLED_AWS_BEDROCK`
|
||
|
||
- Type:Optional
|
||
- Description:Enables AWS Bedrock as a model provider by default, turns on the AWS Bedrock service when set to `1`
|
||
- Default:`0`
|
||
- Example:`1`
|
||
|
||
### `AWS_ACCESS_KEY_ID`
|
||
|
||
- Type: Required
|
||
- Description: Access key ID for AWS service authentication
|
||
- Default: -
|
||
- Example: `AKIA5STVRLFSB4S9HWBR`
|
||
|
||
### `AWS_SECRET_ACCESS_KEY`
|
||
|
||
- Type: Required
|
||
- Description: Key for AWS service authentication
|
||
- Default: -
|
||
- Example: `Th3vXxLYpuKcv2BARktPSTPxx+jbSiFT6/0w7oEC`
|
||
|
||
### `AWS_REGION`
|
||
|
||
- Type: Optional
|
||
- Description: Region setting for AWS services
|
||
- Default: `us-east-1`
|
||
- Example: `us-east-1`
|
||
|
||
## DeepSeek AI
|
||
|
||
### `DEEPSEEK_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: If you manually configure the DeepSeek API proxy, you can use this configuration item to override the default DeepSeek API request base URL
|
||
- Default: `https://api.deepseek.com`
|
||
- Example: `https://my-deepseek-proxy.com`
|
||
|
||
### `DEEPSEEK_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the DeepSeek AI service
|
||
- Default: -
|
||
- Example: `sk-xxxxxx...xxxxxx`
|
||
|
||
### `DEEPSEEK_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=displayName` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+deepseek-reasoner`
|
||
|
||
## XAI
|
||
|
||
### `XAI_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the XAI service
|
||
- Default: -
|
||
- Example: `xai-xxxxxx...xxxxxx`
|
||
|
||
## Wenxin
|
||
|
||
### `WENXIN_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the Baidu Wenxin service
|
||
- Default: -
|
||
- Example: `xxxxxx...xxxxxx`
|
||
|
||
## OpenRouter AI
|
||
|
||
### `OPENROUTER_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the OpenRouter AI service
|
||
- Default: -
|
||
- Example: `sk-or-v1-xxxxxx...xxxxxx=`
|
||
|
||
### `OPENROUTER_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to specify a custom OpenRouter model list. Model definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+01-ai/yi-34b-chat,+huggingfaceh4/zephyr-7b-beta`
|
||
|
||
## PPIO
|
||
|
||
### `PPIO_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This your PPIO API Key.
|
||
- Default: -
|
||
- Example: `sk_xxxxxxxxxx`
|
||
|
||
### `PPIO_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+deepseek/deepseek-v3/community,+deepseek/deepseek-r1-distill-llama-70b`
|
||
|
||
## Github
|
||
|
||
### `GITHUB_TOKEN`
|
||
|
||
- Type: Required
|
||
- Description: This is your Personal access tokens you got for in the Github
|
||
- Default: `-`
|
||
- Example:`ghp_xxxxxx...xxxxxx=`
|
||
|
||
### `GITHUB_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+gpt-4o,+gpt-4o-mini`
|
||
|
||
## TogetherAI
|
||
|
||
### `TOGETHERAI_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the TogetherAI service
|
||
- Default: -
|
||
- Example: `xxxxxx...xxxxxx`
|
||
|
||
### `TOGETHERAI_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to specify a custom TogetherAI model list. Model definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo,+meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo`
|
||
|
||
## Fireworks AI
|
||
|
||
### `FIREWORKSAI_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the Fireworks AI service
|
||
- Default: `-`
|
||
- Example:`xxxxxx...xxxxxx`
|
||
|
||
### `FIREWORKSAI_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+accounts/fireworks/models/firefunction-v2,+accounts/fireworks/models/firefunction-v1`
|
||
|
||
## Ollama
|
||
|
||
### `ENABLED_OLLAMA`
|
||
|
||
- Type:Optional
|
||
- Description:Enables Ollama as a model provider by default, turns off the Ollama service when set to `0`
|
||
- Default:`1`
|
||
- Example:`0`
|
||
|
||
### `OLLAMA_PROXY_URL`
|
||
|
||
- Type: Required
|
||
- Description: Used to enable the Ollama service, setting this will display optional open-source language models in the language model list and can also specify custom language models
|
||
- Default: -
|
||
- Example: `http://127.0.0.1:11434`
|
||
|
||
### `OLLAMA_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to specify a custom Ollama language model. Model definition syntax rules see [model-list][model-list]
|
||
- Default: -
|
||
- Example: `llama2:7B`
|
||
|
||
## Moonshot AI
|
||
|
||
### `MOONSHOT_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the Moonshot AI service
|
||
- Default: -
|
||
- Example: `Y2xpdGhpMzNhZXNoYjVtdnZjMWc6bXNrLWIxQlk3aDNPaXpBWnc0V1RaMDhSRmRFVlpZUWY=`
|
||
|
||
### `MOONSHOT_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: If you manually configure the Moonshot API proxy, you can use this configuration item to override the default Moonshot API request base URL
|
||
- Default: `https://api.moonshot.cn/v1`
|
||
- Example: `https://my-moonshot-proxy.com/v1`
|
||
|
||
## Perplexity AI
|
||
|
||
### `PERPLEXITY_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied from Perplexity AI
|
||
- Default: -
|
||
- Example: `pplx-xxxxxx...xxxxxx`
|
||
|
||
### `PERPLEXITY_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: If you manually configure the Perplexity API proxy, you can use this configuration item to override the default Perplexity API request base URL
|
||
- Default: `https://api.Perplexity.ai`
|
||
- Example: `https://my-Perplexity-proxy.com`
|
||
|
||
### `PERPLEXITY_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see
|
||
- Default: `-`
|
||
- Example: `-all,+llama-3.1-sonar-small-128k-online,+llama-3.1-sonar-small-128k-chat`
|
||
|
||
## Minimax AI
|
||
|
||
### `MINIMAX_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the Minimax AI service
|
||
- Default: -
|
||
- Example: `xxxxxx...xxxxxx`
|
||
|
||
## Mistral AI
|
||
|
||
### `MISTRAL_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the Mistral AI service
|
||
- Default: -
|
||
- Example: `xxxxxx...xxxxxx=`
|
||
|
||
## Groq AI
|
||
|
||
### `GROQ_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied from Groq AI
|
||
- Default: -
|
||
- Example: `gsk_xxxxxx...xxxxxx`
|
||
|
||
### `GROQ_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: If you manually configure the Groq API proxy, you can use this configuration item to override the default Groq API request base URL
|
||
- Default: `https://api.groq.com/openai/v1`
|
||
- Example: `https://my-groq-proxy.com/v1`
|
||
|
||
### `GROQ_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+gemma2-9b-it,+llama-3.1-8b-instant`
|
||
|
||
## ZHIPU AI
|
||
|
||
### `ZHIPU_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the ZHIPU AI service
|
||
- Default: -
|
||
- Example: `4582d332441a313f5c2ed9824d1798ca.rC8EcTAhgbOuAuVT`
|
||
|
||
### `ZHIPU_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+glm-4-alltools,+glm-4-plus`
|
||
|
||
## 01.AI
|
||
|
||
### `ZEROONE_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the 01.AI service
|
||
- Default: -
|
||
- Example:`xxxxxx...xxxxxx`
|
||
|
||
### `ZEROONE_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+yi-large,+yi-large-rag`
|
||
|
||
## Qiniu
|
||
|
||
### `QINIU_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you can obtain from Qiniu AI service
|
||
- Default: -
|
||
- Example:`sk-xxxxx...xxxxx`
|
||
|
||
### `QINIU_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+deepseek-r1,+deepseek-v3`
|
||
|
||
### `QINIU_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: If you manually configure the Qiniu API proxy, you can use this configuration item to override the default Qiniu API request base URL
|
||
- Default: `https://api.qnaigc.com/v1`
|
||
- Example: `https://my-qnaigc.com/v1`
|
||
|
||
## Qwen
|
||
|
||
### `QWEN_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the DashScope API key you can obtain from Alibaba Cloud
|
||
- Default: -
|
||
- Example:`sk-xxxxx...xxxxx`
|
||
|
||
### `QWEN_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name->deploymentName=display_name` to customize the display name of a model, separated by commas. The deploymentName `->deploymentName` can be omitted, and it defaults to the latest model version. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+qwen-turbo,+qwen-plus,+qwen-max->qwen-max-2025-01-25`
|
||
|
||
### `QWEN_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: If you manually configure the Qwen API proxy, you can use this configuration item to override the default Qwen API request base URL
|
||
- Default: `https://dashscope.aliyuncs.com/compatible-mode/v1`
|
||
- Example: `https://my-qwen-proxy.com/v1`
|
||
|
||
## Stepfun AI
|
||
|
||
### `STEPFUN_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the DashScope API key you can obtain from Stepfun AI service
|
||
- Default: -
|
||
- Example:`sk-xxxxx...xxxxx`
|
||
|
||
## Novita AI
|
||
|
||
### `NOVITA_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the Novita AI service
|
||
- Default: -
|
||
- Example:`xxxxxx...xxxxxx`
|
||
|
||
### `NOVITA_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+meta-llama/llama-3.1-8b-instruct,+meta-llama/llama-3.1-70b-instruct`
|
||
|
||
## BAICHUAN
|
||
|
||
### `BAICHUAN_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the BAICHUAN service
|
||
- Default: -
|
||
- Example:`xxxxxx...xxxxxx`
|
||
|
||
## TAICHU
|
||
|
||
### `TAICHU_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the TAICHU service
|
||
- Default: -
|
||
- Example:`xxxxxx...xxxxxx`
|
||
|
||
## 360 AI
|
||
|
||
### `AI360_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the 360 AI service
|
||
- Default: -
|
||
- Example:`xxxxxx...xxxxxx`
|
||
|
||
## Siliconflow
|
||
|
||
### `SILICONCLOUD_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied from Siliconflow service
|
||
- Default: -
|
||
- Example: `xxxxxx...xxxxxx`
|
||
|
||
### `SILICONCLOUD_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: If you manually configure the Siliconflow API proxy, you can use this configuration item to override the default Siliconflow API request base URL
|
||
- Default: `https://api.siliconflow.cn/v1`
|
||
- Example: `https://my-siliconflow-proxy.com/v1`
|
||
|
||
### `SILICONCLOUD_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+deepseek-ai/DeepSeek-V2.5,+Qwen/Qwen2.5-7B-Instruct`
|
||
|
||
## Upstage AI
|
||
|
||
### `UPSTAGE_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied from Upstage AI service
|
||
- Default: -
|
||
- Example: `xxxxxx...xxxxxx`
|
||
|
||
## Spark AI
|
||
|
||
### `SPARK_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied from Spark AI service
|
||
- Default: -
|
||
- Example: `xxxxxx...xxxxxx`
|
||
|
||
## A21 AI
|
||
|
||
### `AI21_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied from `AI21_API_KEY` service
|
||
- Default: -
|
||
- Example: `xxxxxx...xxxxxx`
|
||
|
||
## HUNYUAN
|
||
|
||
### `HUNYUAN_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied from HUNYUAN service
|
||
- Default: -
|
||
- Example: `xxxxxx...xxxxxx`
|
||
|
||
### `HUNYUAN_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+hunyuan-lite,+hunyuan-standard`
|
||
|
||
## Volcengine
|
||
|
||
### `VOLCENGINE_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied from Volcengine service, you can check it out [here](https://console.volcengine.com/ark/region:ark+cn-beijing/endpoint)
|
||
- Default: -
|
||
- Example: `ad925b3d-aaaa-bbbb-cccc-51fde2f50547`
|
||
|
||
### `VOLCENGINE_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name->deploymentName=display_name` to customize the display name of a model, separated by commas. The deploymentName `->deploymentName` can be omitted, and it defaults to the latest model version. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+deepseek-r1,+deepseek-v3->deepseek-v3-250324,+doubao-1.5-pro-256k,+doubao-1.5-pro-32k->doubao-1-5-pro-32k-250115,+doubao-1.5-lite-32k`
|
||
|
||
### `VOLCENGINE_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: If you manually configure the Volcengine API proxy, you can use this configuration item to override the default Volcengine API request base URL
|
||
- Default: `https://ark.cn-beijing.volces.com/api/v3`
|
||
- Example: `https://my-volcengine-proxy.com/v1`
|
||
|
||
## InfiniAI
|
||
|
||
### `INFINIAI_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied from Infini-AI, you can check it out [here](https://cloud.infini-ai.com)
|
||
- Default: -
|
||
- Example: `sk-xxxxxx...xxxxxx`
|
||
|
||
### `INFINIAI_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the model list, use `+` to add a model, use `-` to hide a model, use `model_name=display_name` to customize the display name of a model, separated by commas. Definition syntax rules see [model-list][model-list]
|
||
- Default: `-`
|
||
- Example: `-all,+qwq-32b,+deepseek-r1`
|
||
|
||
## FAL
|
||
|
||
### `ENABLED_FAL`
|
||
|
||
- Type: Optional
|
||
- Description: Enables FAL as a model provider by default. Set to `0` to disable the FAL service.
|
||
- Default: `1`
|
||
- Example: `0`
|
||
|
||
### `FAL_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the FAL service.
|
||
- Default: -
|
||
- Example: `fal-xxxxxx...xxxxxx`
|
||
|
||
### `FAL_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the FAL model list. Use `+` to add a model, `-` to hide a model, and `model_name=display_name` to customize the display name of a model. Separate multiple entries with commas. The definition syntax follows the same rules as other providers' model lists.
|
||
- Default: `-`
|
||
- Example: `-all,+fal-ai/flux/schnell,+fal-ai/flux-pro/kontext=FLUX.1 Kontext [pro]`
|
||
|
||
The above example disables all models first, then enables `fal-ai/flux/schnell` and `fal-ai/flux-pro/kontext` (displayed as `FLUX.1 Kontext [pro]`).
|
||
|
||
## ComfyUI
|
||
|
||
### `ENABLED_COMFYUI`
|
||
|
||
- Type: Optional
|
||
- Description: Enables ComfyUI as a model provider by default. Set to `0` to disable the ComfyUI service.
|
||
- Default: `1`
|
||
- Example: `0`
|
||
|
||
### `COMFYUI_BASE_URL`
|
||
|
||
- Type: Optional
|
||
- Description: The base URL address of the ComfyUI service
|
||
- Default: `http://localhost:8188`
|
||
- Example: `http://192.168.1.100:8188` or `https://my-comfyui-server.com`
|
||
|
||
### `COMFYUI_AUTH_TYPE`
|
||
|
||
- Type: Optional
|
||
- Description: The authentication type for ComfyUI, supporting 4 authentication methods
|
||
- `none`: No authentication (default)
|
||
- `basic`: Basic authentication (username + password)
|
||
- `bearer`: Bearer Token authentication (API key)
|
||
- `custom`: Custom request header authentication
|
||
- Default: `none`
|
||
- Example: `basic`
|
||
|
||
### `COMFYUI_API_KEY`
|
||
|
||
- Type: Optional
|
||
- Description: The API key used when the authentication type is `bearer`
|
||
- Default: -
|
||
- Example: `sk-xxxxxx...xxxxxx`
|
||
|
||
### `COMFYUI_USERNAME`
|
||
|
||
- Type: Optional
|
||
- Description: The username used when the authentication type is `basic`
|
||
- Default: -
|
||
- Example: `admin`
|
||
|
||
### `COMFYUI_PASSWORD`
|
||
|
||
- Type: Optional
|
||
- Description: The password used when the authentication type is `basic`
|
||
- Default: -
|
||
- Example: `password123`
|
||
|
||
### `COMFYUI_CUSTOM_HEADERS`
|
||
|
||
- Type: Optional
|
||
- Description: Custom request headers used when the authentication type is `custom`, requires JSON format string
|
||
- Default: -
|
||
- Example: `{"X-Auth-Token": "your-token", "X-Custom-Header": "value"}`
|
||
|
||
<Callout type={'info'}>
|
||
ComfyUI supports multiple authentication methods. Please choose the appropriate authentication
|
||
type and corresponding authentication parameters according to your ComfyUI service configuration.
|
||
If your ComfyUI service has no authentication set up, you can skip configuring
|
||
authentication-related environment variables.
|
||
</Callout>
|
||
|
||
## BFL
|
||
|
||
### `BFL_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the BFL service.
|
||
- Default: -
|
||
- Example: `12345678-1234-1234-1234-123456789abc`
|
||
|
||
### `BFL_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the BFL model list. Use `+` to add a model, `-` to hide a model, and `model_name=display_name` to customize the display name of a model. Separate multiple entries with commas. The definition syntax follows the same rules as other providers' model lists.
|
||
- Default: `-`
|
||
- Example: `-all,+flux-pro-1.1,+flux-kontext-pro=FLUX.1 Kontext [pro]`
|
||
|
||
The above example disables all models first, then enables `flux-pro-1.1` and `flux-kontext-pro` (displayed as `FLUX.1 Kontext [pro]`).
|
||
|
||
## NewAPI
|
||
|
||
### `NEWAPI_API_KEY`
|
||
|
||
- Type: Optional
|
||
- Description: This is the API key for your NewAPI service instance. NewAPI is a multi-provider model aggregation service that provides unified access to various AI model APIs.
|
||
- Default: -
|
||
- Example: `sk-xxxxxx...xxxxxx`
|
||
|
||
### `NEWAPI_PROXY_URL`
|
||
|
||
- Type: Optional
|
||
- Description: The base URL for your NewAPI server instance. This should point to your deployed NewAPI service endpoint.
|
||
- Default: -
|
||
- Example: `https://your-newapi-server.com/`
|
||
|
||
NewAPI is a multi-provider model aggregation service that supports automatic model routing based on provider detection. It offers cost management features and provides a single endpoint for accessing models from multiple providers including OpenAI, Anthropic, Google, and more. Learn more about NewAPI at [https://github.com/Calcium-Ion/new-api](https://github.com/Calcium-Ion/new-api).
|
||
|
||
## Vercel AI Gateway
|
||
|
||
### `VERCELAIGATEWAY_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the Vercel AI Gateway service.
|
||
- Default: -
|
||
- Example: `vck_xxxxxx...xxxxxx`
|
||
|
||
### `VERCELAIGATEWAY_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the Vercel AI Gateway model list. Use `+` to add a model, `-` to hide a model, and `model_name=display_name` to customize the display name of a model. Separate multiple entries with commas. The definition syntax follows the same rules as other providers' model lists.
|
||
- Default: `-`
|
||
- Example: `-all,+vercel-model-1,+vercel-model-2=vercel-special`
|
||
|
||
## Cerebras
|
||
|
||
### `CEREBRAS_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the Cerebras service.
|
||
- Default: -
|
||
- Example: `csk-xxxxxx...xxxxxx`
|
||
|
||
### `CEREBRAS_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the Cerebras model list. Use `+` to add a model, `-` to hide a model, and `model_name=display_name` to customize the display name of a model. Separate multiple entries with commas. The definition syntax follows the same rules as other providers' model lists.
|
||
- Default: `-`
|
||
- Example: `-all,+cerebras-model-1,+cerebras-model-2=cerebras-special`
|
||
|
||
## AiHubMix
|
||
|
||
### `AIHUBMIX_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the AiHubMix service.
|
||
- Default: -
|
||
- Example: `sk-xxxxxx...xxxxxx`
|
||
|
||
### `AIHUBMIX_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the AiHubMix model list. Use `+` to add a model, `-` to hide a model, and `model_name=display_name` to customize the display name of a model. Separate multiple entries with commas. The definition syntax follows the same rules as other providers' model lists.
|
||
- Default: `-`
|
||
- Example: `-all,+claude-opus-4-1-20250805,+claude-opus-4-20250514=claude-opus-4`
|
||
|
||
## Straico
|
||
|
||
### `STRAICO_API_KEY`
|
||
|
||
- Type: Required
|
||
- Description: This is the API key you applied for in the Straico service.
|
||
- Default: -
|
||
- Example: `sk-xxxxxx...xxxxxx`
|
||
|
||
### `STRAICO_MODEL_LIST`
|
||
|
||
- Type: Optional
|
||
- Description: Used to control the Straico model list. Use `+` to add a model, `-` to hide a model, and `model_name=display_name` to customize the display name of a model. Separate multiple entries with commas. The definition syntax follows the same rules as other providers' model lists.
|
||
- Default: `-`
|
||
- Example: `-all,+straico-model-1,+straico-model-2=straico-special`
|
||
|
||
[model-list]: /docs/self-hosting/advanced/model-list
|