mirror of
https://github.com/danny-avila/LibreChat.git
synced 2025-09-22 06:00:56 +02:00
🤗 feat: Known Endpoints: HuggingFace (#2646)
* endpoints: huggingface * Update ai_endpoints.md * huggingface: update icon
This commit is contained in:
parent
83bae9e9d9
commit
d73ea8e1f2
7 changed files with 137 additions and 35 deletions
|
@ -68,6 +68,7 @@ PROXY=
|
||||||
# APIPIE_API_KEY=
|
# APIPIE_API_KEY=
|
||||||
# FIREWORKS_API_KEY=
|
# FIREWORKS_API_KEY=
|
||||||
# GROQ_API_KEY=
|
# GROQ_API_KEY=
|
||||||
|
# HUGGINGFACE_TOKEN=
|
||||||
# MISTRAL_API_KEY=
|
# MISTRAL_API_KEY=
|
||||||
# OPENROUTER_KEY=
|
# OPENROUTER_KEY=
|
||||||
# PERPLEXITY_API_KEY=
|
# PERPLEXITY_API_KEY=
|
||||||
|
|
8
client/public/assets/huggingface.svg
Normal file
8
client/public/assets/huggingface.svg
Normal file
File diff suppressed because one or more lines are too long
After Width: | Height: | Size: 34 KiB |
|
@ -8,6 +8,7 @@ const knownEndpointAssets = {
|
||||||
[KnownEndpoints.cohere]: '/assets/cohere.png',
|
[KnownEndpoints.cohere]: '/assets/cohere.png',
|
||||||
[KnownEndpoints.fireworks]: '/assets/fireworks.png',
|
[KnownEndpoints.fireworks]: '/assets/fireworks.png',
|
||||||
[KnownEndpoints.groq]: '/assets/groq.png',
|
[KnownEndpoints.groq]: '/assets/groq.png',
|
||||||
|
[KnownEndpoints.huggingface]: '/assets/huggingface.svg',
|
||||||
[KnownEndpoints.mistral]: '/assets/mistral.png',
|
[KnownEndpoints.mistral]: '/assets/mistral.png',
|
||||||
[KnownEndpoints.mlx]: '/assets/mlx.png',
|
[KnownEndpoints.mlx]: '/assets/mlx.png',
|
||||||
[KnownEndpoints.ollama]: '/assets/ollama.png',
|
[KnownEndpoints.ollama]: '/assets/ollama.png',
|
||||||
|
|
|
@ -123,6 +123,39 @@ Some of the endpoints are marked as **Known,** which means they might have speci
|
||||||
|
|
||||||

|

|
||||||
|
|
||||||
|
## Apple MLX
|
||||||
|
> MLX API key: ignored - [MLX OpenAI Compatibility](https://github.com/ml-explore/mlx-examples/blob/main/llms/mlx_lm/SERVER.md)
|
||||||
|
|
||||||
|
**Notes:**
|
||||||
|
|
||||||
|
- **Known:** icon provided.
|
||||||
|
|
||||||
|
- API is mostly strict with unrecognized parameters.
|
||||||
|
- Support only one model at a time, otherwise you'll need to run a different endpoint with a different `baseURL`.
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
- name: "MLX"
|
||||||
|
apiKey: "mlx"
|
||||||
|
baseURL: "http://localhost:8080/v1/"
|
||||||
|
models:
|
||||||
|
default: [
|
||||||
|
"Meta-Llama-3-8B-Instruct-4bit"
|
||||||
|
]
|
||||||
|
fetch: false # fetching list of models is not supported
|
||||||
|
titleConvo: true
|
||||||
|
titleModel: "current_model"
|
||||||
|
summarize: false
|
||||||
|
summaryModel: "current_model"
|
||||||
|
forcePrompt: false
|
||||||
|
modelDisplayLabel: "Apple MLX"
|
||||||
|
addParams:
|
||||||
|
max_tokens: 2000
|
||||||
|
"stop": [
|
||||||
|
"<|eot_id|>"
|
||||||
|
]
|
||||||
|
```
|
||||||
|
|
||||||
|

|
||||||
|
|
||||||
## Cohere
|
## Cohere
|
||||||
> Cohere API key: [dashboard.cohere.com](https://dashboard.cohere.com/)
|
> Cohere API key: [dashboard.cohere.com](https://dashboard.cohere.com/)
|
||||||
|
@ -204,7 +237,6 @@ Some of the endpoints are marked as **Known,** which means they might have speci
|
||||||
default: [
|
default: [
|
||||||
"llama3-70b-8192",
|
"llama3-70b-8192",
|
||||||
"llama3-8b-8192",
|
"llama3-8b-8192",
|
||||||
"llama2-70b-4096",
|
|
||||||
"mixtral-8x7b-32768",
|
"mixtral-8x7b-32768",
|
||||||
"gemma-7b-it",
|
"gemma-7b-it",
|
||||||
]
|
]
|
||||||
|
@ -216,6 +248,97 @@ Some of the endpoints are marked as **Known,** which means they might have speci
|
||||||
|
|
||||||

|

|
||||||
|
|
||||||
|
## Huggingface
|
||||||
|
> groq API key: [wow.groq.com](https://wow.groq.com/)
|
||||||
|
|
||||||
|
**Notes:**
|
||||||
|
|
||||||
|
- **Known:** icon provided.
|
||||||
|
|
||||||
|
- The provided models are free but rate limited
|
||||||
|
|
||||||
|
- The use of [`dropParams`](./custom_config.md#dropparams) to drop "top_p" params is required.
|
||||||
|
- Fetching models isn't supported
|
||||||
|
- Note: Some models currently work better than others, answers are very short (at least when using the free tier).
|
||||||
|
|
||||||
|
- The example includes a model list, which was last updated on May 09, 2024, for your convenience.
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
- name: 'HuggingFace'
|
||||||
|
apiKey: '${HUGGINGFACE_TOKEN}'
|
||||||
|
baseURL: 'https://api-inference.huggingface.co/v1'
|
||||||
|
models:
|
||||||
|
default: [
|
||||||
|
"codellama/CodeLlama-34b-Instruct-hf",
|
||||||
|
"google/gemma-1.1-2b-it",
|
||||||
|
"google/gemma-1.1-7b-it",
|
||||||
|
"HuggingFaceH4/starchat2-15b-v0.1",
|
||||||
|
"HuggingFaceH4/zephyr-7b-beta",
|
||||||
|
"meta-llama/Meta-Llama-3-8B-Instruct",
|
||||||
|
"microsoft/Phi-3-mini-4k-instruct",
|
||||||
|
"mistralai/Mistral-7B-Instruct-v0.1",
|
||||||
|
"mistralai/Mistral-7B-Instruct-v0.2",
|
||||||
|
"mistralai/Mixtral-8x7B-Instruct-v0.1",
|
||||||
|
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
|
||||||
|
]
|
||||||
|
fetch: true
|
||||||
|
titleConvo: true
|
||||||
|
titleModel: "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO"
|
||||||
|
dropParams: ["top_p"]
|
||||||
|
modelDisplayLabel: "HuggingFace"
|
||||||
|
```
|
||||||
|
|
||||||
|
??? warning "Other Model Errors"
|
||||||
|
|
||||||
|
Here’s a list of the other models that were tested along with their corresponding errors
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
models:
|
||||||
|
default: [
|
||||||
|
"CohereForAI/c4ai-command-r-plus", # Model requires a Pro subscription
|
||||||
|
"HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1", # Model requires a Pro subscription
|
||||||
|
"meta-llama/Llama-2-7b-hf", # Model requires a Pro subscription
|
||||||
|
"meta-llama/Meta-Llama-3-70B-Instruct", # Model requires a Pro subscription
|
||||||
|
"meta-llama/Llama-2-13b-chat-hf", # Model requires a Pro subscription
|
||||||
|
"meta-llama/Llama-2-13b-hf", # Model requires a Pro subscription
|
||||||
|
"meta-llama/Llama-2-70b-chat-hf", # Model requires a Pro subscription
|
||||||
|
"meta-llama/Llama-2-7b-chat-hf", # Model requires a Pro subscription
|
||||||
|
"------",
|
||||||
|
"bigcode/octocoder", # template not found
|
||||||
|
"bigcode/santacoder", # template not found
|
||||||
|
"bigcode/starcoder2-15b", # template not found
|
||||||
|
"bigcode/starcoder2-3b", # template not found
|
||||||
|
"codellama/CodeLlama-13b-hf", # template not found
|
||||||
|
"codellama/CodeLlama-7b-hf", # template not found
|
||||||
|
"google/gemma-2b", # template not found
|
||||||
|
"google/gemma-7b", # template not found
|
||||||
|
"HuggingFaceH4/starchat-beta", # template not found
|
||||||
|
"HuggingFaceM4/idefics-80b-instruct", # template not found
|
||||||
|
"HuggingFaceM4/idefics-9b-instruct", # template not found
|
||||||
|
"HuggingFaceM4/idefics2-8b", # template not found
|
||||||
|
"kashif/stack-llama-2", # template not found
|
||||||
|
"lvwerra/starcoderbase-gsm8k", # template not found
|
||||||
|
"tiiuae/falcon-7b", # template not found
|
||||||
|
"timdettmers/guanaco-33b-merged", # template not found
|
||||||
|
"------",
|
||||||
|
"bigscience/bloom", # 404 status code (no body)
|
||||||
|
"------",
|
||||||
|
"google/gemma-2b-it", # stream` is not supported for this model / unknown error
|
||||||
|
"------",
|
||||||
|
"google/gemma-7b-it", # AI Response error likely caused by Google censor/filter
|
||||||
|
"------",
|
||||||
|
"bigcode/starcoder", # Service Unavailable
|
||||||
|
"google/flan-t5-xxl", # Service Unavailable
|
||||||
|
"HuggingFaceH4/zephyr-7b-alpha", # Service Unavailable
|
||||||
|
"mistralai/Mistral-7B-v0.1", # Service Unavailable
|
||||||
|
"OpenAssistant/oasst-sft-1-pythia-12b", # Service Unavailable
|
||||||
|
"OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5", # Service Unavailable
|
||||||
|
]
|
||||||
|
```
|
||||||
|
|
||||||
|
|
||||||
|

|
||||||
|
|
||||||
|
|
||||||
## LiteLLM
|
## LiteLLM
|
||||||
> LiteLLM API key: master_key value [LiteLLM](./litellm.md)
|
> LiteLLM API key: master_key value [LiteLLM](./litellm.md)
|
||||||
|
@ -271,39 +394,6 @@ Some of the endpoints are marked as **Known,** which means they might have speci
|
||||||
|
|
||||||

|

|
||||||
|
|
||||||
## Apple MLX
|
|
||||||
> MLX API key: ignored - [MLX OpenAI Compatibility](https://github.com/ml-explore/mlx-examples/blob/main/llms/mlx_lm/SERVER.md)
|
|
||||||
|
|
||||||
**Notes:**
|
|
||||||
|
|
||||||
- **Known:** icon provided.
|
|
||||||
|
|
||||||
- API is mostly strict with unrecognized parameters.
|
|
||||||
- Support only one model at a time, otherwise you'll need to run a different endpoint with a different `baseURL`.
|
|
||||||
|
|
||||||
```yaml
|
|
||||||
- name: "MLX"
|
|
||||||
apiKey: "mlx"
|
|
||||||
baseURL: "http://localhost:8080/v1/"
|
|
||||||
models:
|
|
||||||
default: [
|
|
||||||
"Meta-Llama-3-8B-Instruct-4bit"
|
|
||||||
]
|
|
||||||
fetch: false # fetching list of models is not supported
|
|
||||||
titleConvo: true
|
|
||||||
titleModel: "current_model"
|
|
||||||
summarize: false
|
|
||||||
summaryModel: "current_model"
|
|
||||||
forcePrompt: false
|
|
||||||
modelDisplayLabel: "Apple MLX"
|
|
||||||
addParams:
|
|
||||||
max_tokens: 2000
|
|
||||||
"stop": [
|
|
||||||
"<|eot_id|>"
|
|
||||||
]
|
|
||||||
```
|
|
||||||
|
|
||||||

|
|
||||||
|
|
||||||
## Ollama
|
## Ollama
|
||||||
> Ollama API key: Required but ignored - [Ollama OpenAI Compatibility](https://github.com/ollama/ollama/blob/main/docs/openai.md)
|
> Ollama API key: Required but ignored - [Ollama OpenAI Compatibility](https://github.com/ollama/ollama/blob/main/docs/openai.md)
|
||||||
|
|
|
@ -166,6 +166,7 @@ ANYSCALE_API_KEY=
|
||||||
APIPIE_API_KEY=
|
APIPIE_API_KEY=
|
||||||
FIREWORKS_API_KEY=
|
FIREWORKS_API_KEY=
|
||||||
GROQ_API_KEY=
|
GROQ_API_KEY=
|
||||||
|
HUGGINGFACE_TOKEN=
|
||||||
MISTRAL_API_KEY=
|
MISTRAL_API_KEY=
|
||||||
OPENROUTER_KEY=
|
OPENROUTER_KEY=
|
||||||
PERPLEXITY_API_KEY=
|
PERPLEXITY_API_KEY=
|
||||||
|
|
|
@ -2,7 +2,7 @@
|
||||||
# https://docs.librechat.ai/install/configuration/custom_config.html
|
# https://docs.librechat.ai/install/configuration/custom_config.html
|
||||||
|
|
||||||
# Configuration version (required)
|
# Configuration version (required)
|
||||||
version: 1.0.6
|
version: 1.0.9
|
||||||
|
|
||||||
# Cache settings: Set to true to enable caching
|
# Cache settings: Set to true to enable caching
|
||||||
cache: true
|
cache: true
|
||||||
|
|
|
@ -306,6 +306,7 @@ export enum KnownEndpoints {
|
||||||
cohere = 'cohere',
|
cohere = 'cohere',
|
||||||
fireworks = 'fireworks',
|
fireworks = 'fireworks',
|
||||||
groq = 'groq',
|
groq = 'groq',
|
||||||
|
huggingface = 'huggingface',
|
||||||
mistral = 'mistral',
|
mistral = 'mistral',
|
||||||
mlx = 'mlx',
|
mlx = 'mlx',
|
||||||
ollama = 'ollama',
|
ollama = 'ollama',
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue