mirror of
https://github.com/danny-avila/LibreChat.git
synced 2025-12-23 11:50:14 +01:00
🚀 feat: Support for GPT-3.5 Turbo/0125 Model (#1704)
* 🚀 feat: Support for GPT-3.5 Turbo/0125 Model
* ci: fix tx test
This commit is contained in:
parent
30e143e96d
commit
8479ac7293
7 changed files with 15 additions and 4 deletions
|
|
@ -101,7 +101,7 @@ GOOGLE_KEY=user_provided
|
||||||
#============#
|
#============#
|
||||||
|
|
||||||
OPENAI_API_KEY=user_provided
|
OPENAI_API_KEY=user_provided
|
||||||
# OPENAI_MODELS=gpt-3.5-turbo-0301,gpt-3.5-turbo,gpt-4,gpt-4-0613,gpt-4-vision-preview,gpt-3.5-turbo-0613,gpt-3.5-turbo-16k-0613,gpt-4-0125-preview,gpt-4-turbo-preview,gpt-4-1106-preview,gpt-3.5-turbo-1106,gpt-3.5-turbo-instruct,gpt-3.5-turbo-instruct-0914,gpt-3.5-turbo-16k
|
# OPENAI_MODELS=gpt-3.5-turbo-0125,gpt-3.5-turbo-0301,gpt-3.5-turbo,gpt-4,gpt-4-0613,gpt-4-vision-preview,gpt-3.5-turbo-0613,gpt-3.5-turbo-16k-0613,gpt-4-0125-preview,gpt-4-turbo-preview,gpt-4-1106-preview,gpt-3.5-turbo-1106,gpt-3.5-turbo-instruct,gpt-3.5-turbo-instruct-0914,gpt-3.5-turbo-16k
|
||||||
|
|
||||||
DEBUG_OPENAI=false
|
DEBUG_OPENAI=false
|
||||||
|
|
||||||
|
|
@ -127,7 +127,7 @@ DEBUG_OPENAI=false
|
||||||
# Plugins #
|
# Plugins #
|
||||||
#============#
|
#============#
|
||||||
|
|
||||||
# PLUGIN_MODELS=gpt-4,gpt-4-turbo-preview,gpt-4-0125-preview,gpt-4-1106-preview,gpt-4-0613,gpt-3.5-turbo,gpt-3.5-turbo-1106,gpt-3.5-turbo-0613
|
# PLUGIN_MODELS=gpt-4,gpt-4-turbo-preview,gpt-4-0125-preview,gpt-4-1106-preview,gpt-4-0613,gpt-3.5-turbo,gpt-3.5-turbo-0125,gpt-3.5-turbo-1106,gpt-3.5-turbo-0613
|
||||||
|
|
||||||
DEBUG_PLUGINS=true
|
DEBUG_PLUGINS=true
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -12,6 +12,7 @@ const tokenValues = {
|
||||||
'16k': { prompt: 3, completion: 4 },
|
'16k': { prompt: 3, completion: 4 },
|
||||||
'gpt-3.5-turbo-1106': { prompt: 1, completion: 2 },
|
'gpt-3.5-turbo-1106': { prompt: 1, completion: 2 },
|
||||||
'gpt-4-1106': { prompt: 10, completion: 30 },
|
'gpt-4-1106': { prompt: 10, completion: 30 },
|
||||||
|
'gpt-3.5-turbo-0125': { prompt: 0.5, completion: 1.5 },
|
||||||
};
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
@ -29,6 +30,8 @@ const getValueKey = (model, endpoint) => {
|
||||||
|
|
||||||
if (modelName.includes('gpt-3.5-turbo-16k')) {
|
if (modelName.includes('gpt-3.5-turbo-16k')) {
|
||||||
return '16k';
|
return '16k';
|
||||||
|
} else if (modelName.includes('gpt-3.5-turbo-0125')) {
|
||||||
|
return 'gpt-3.5-turbo-0125';
|
||||||
} else if (modelName.includes('gpt-3.5-turbo-1106')) {
|
} else if (modelName.includes('gpt-3.5-turbo-1106')) {
|
||||||
return 'gpt-3.5-turbo-1106';
|
return 'gpt-3.5-turbo-1106';
|
||||||
} else if (modelName.includes('gpt-3.5')) {
|
} else if (modelName.includes('gpt-3.5')) {
|
||||||
|
|
|
||||||
|
|
@ -90,6 +90,9 @@ describe('getMultiplier', () => {
|
||||||
expect(getMultiplier({ tokenType: 'completion', model: 'gpt-4-turbo-vision-preview' })).toBe(
|
expect(getMultiplier({ tokenType: 'completion', model: 'gpt-4-turbo-vision-preview' })).toBe(
|
||||||
tokenValues['gpt-4-1106'].completion,
|
tokenValues['gpt-4-1106'].completion,
|
||||||
);
|
);
|
||||||
|
expect(getMultiplier({ tokenType: 'completion', model: 'gpt-3.5-turbo-0125' })).toBe(
|
||||||
|
tokenValues['gpt-3.5-turbo-0125'].completion,
|
||||||
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should return defaultRate if derived valueKey does not match any known patterns', () => {
|
it('should return defaultRate if derived valueKey does not match any known patterns', () => {
|
||||||
|
|
|
||||||
|
|
@ -55,6 +55,7 @@ const openAIModels = {
|
||||||
'gpt-3.5-turbo-16k': 16375, // -10 from max
|
'gpt-3.5-turbo-16k': 16375, // -10 from max
|
||||||
'gpt-3.5-turbo-16k-0613': 16375, // -10 from max
|
'gpt-3.5-turbo-16k-0613': 16375, // -10 from max
|
||||||
'gpt-3.5-turbo-1106': 16375, // -10 from max
|
'gpt-3.5-turbo-1106': 16375, // -10 from max
|
||||||
|
'gpt-3.5-turbo-0125': 16375, // -10 from max
|
||||||
'mistral-': 31990, // -10 from max
|
'mistral-': 31990, // -10 from max
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -92,6 +92,9 @@ describe('getModelMaxTokens', () => {
|
||||||
expect(getModelMaxTokens('gpt-4-0125-preview')).toBe(
|
expect(getModelMaxTokens('gpt-4-0125-preview')).toBe(
|
||||||
maxTokensMap[EModelEndpoint.openAI]['gpt-4-0125'],
|
maxTokensMap[EModelEndpoint.openAI]['gpt-4-0125'],
|
||||||
);
|
);
|
||||||
|
expect(getModelMaxTokens('gpt-3.5-turbo-0125')).toBe(
|
||||||
|
maxTokensMap[EModelEndpoint.openAI]['gpt-3.5-turbo-0125'],
|
||||||
|
);
|
||||||
});
|
});
|
||||||
|
|
||||||
test('should return correct tokens for Anthropic models', () => {
|
test('should return correct tokens for Anthropic models', () => {
|
||||||
|
|
|
||||||
|
|
@ -317,7 +317,7 @@ DEBUG_OPENAI=false
|
||||||
- Leave it blank or commented out to use internal settings.
|
- Leave it blank or commented out to use internal settings.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
OPENAI_MODELS=gpt-3.5-turbo-0301,gpt-3.5-turbo,gpt-4,gpt-4-0613,gpt-4-vision-preview,gpt-3.5-turbo-0613,gpt-3.5-turbo-16k-0613,gpt-4-0125-preview,gpt-4-turbo-preview,gpt-4-1106-preview,gpt-3.5-turbo-1106,gpt-3.5-turbo-instruct,gpt-3.5-turbo-instruct-0914,gpt-3.5-turbo-16k
|
OPENAI_MODELS=gpt-3.5-turbo-0125,gpt-3.5-turbo-0301,gpt-3.5-turbo,gpt-4,gpt-4-0613,gpt-4-vision-preview,gpt-3.5-turbo-0613,gpt-3.5-turbo-16k-0613,gpt-4-0125-preview,gpt-4-turbo-preview,gpt-4-1106-preview,gpt-3.5-turbo-1106,gpt-3.5-turbo-instruct,gpt-3.5-turbo-instruct-0914,gpt-3.5-turbo-16k
|
||||||
```
|
```
|
||||||
|
|
||||||
- Titling is enabled by default when initiating a conversation.
|
- Titling is enabled by default when initiating a conversation.
|
||||||
|
|
@ -383,7 +383,7 @@ Here are some useful documentation about plugins:
|
||||||
- Identify the available models, separated by commas **without spaces**. The first model in the list will be set as default. Leave it blank or commented out to use internal settings.
|
- Identify the available models, separated by commas **without spaces**. The first model in the list will be set as default. Leave it blank or commented out to use internal settings.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
PLUGIN_MODELS=gpt-4,gpt-4-turbo-preview,gpt-4-0125-preview,gpt-4-1106-preview,gpt-4-0613,gpt-3.5-turbo,gpt-3.5-turbo-1106,gpt-3.5-turbo-0613
|
PLUGIN_MODELS=gpt-4,gpt-4-turbo-preview,gpt-4-0125-preview,gpt-4-1106-preview,gpt-4-0613,gpt-3.5-turbo,gpt-3.5-turbo-0125,gpt-3.5-turbo-1106,gpt-3.5-turbo-0613
|
||||||
```
|
```
|
||||||
|
|
||||||
- Set to false or comment out to disable debug mode for plugins
|
- Set to false or comment out to disable debug mode for plugins
|
||||||
|
|
|
||||||
|
|
@ -91,6 +91,7 @@ export const defaultModels = {
|
||||||
'claude-instant-1-100k',
|
'claude-instant-1-100k',
|
||||||
],
|
],
|
||||||
[EModelEndpoint.openAI]: [
|
[EModelEndpoint.openAI]: [
|
||||||
|
'gpt-3.5-turbo-0125',
|
||||||
'gpt-3.5-turbo-16k-0613',
|
'gpt-3.5-turbo-16k-0613',
|
||||||
'gpt-3.5-turbo-16k',
|
'gpt-3.5-turbo-16k',
|
||||||
'gpt-4-turbo-preview',
|
'gpt-4-turbo-preview',
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue