mirror of
https://github.com/danny-avila/LibreChat.git
synced 2025-12-18 09:20:15 +01:00
🗣️ feat: Azure OpenAI speech (#2985)
* feat: Azure STT * feat: Azure TTS * refactor: use enums * fix: frontend tests * fix(config): wrong key provider
This commit is contained in:
parent
87bdbda10a
commit
a05e2c1dcc
4 changed files with 201 additions and 79 deletions
|
|
@ -1,3 +1,4 @@
|
||||||
|
const { TTSProviders } = require('librechat-data-provider');
|
||||||
const getCustomConfig = require('~/server/services/Config/getCustomConfig');
|
const getCustomConfig = require('~/server/services/Config/getCustomConfig');
|
||||||
const { getProvider } = require('./textToSpeech');
|
const { getProvider } = require('./textToSpeech');
|
||||||
|
|
||||||
|
|
@ -24,13 +25,16 @@ async function getVoices(req, res) {
|
||||||
let voices;
|
let voices;
|
||||||
|
|
||||||
switch (provider) {
|
switch (provider) {
|
||||||
case 'openai':
|
case TTSProviders.OPENAI:
|
||||||
voices = ttsSchema.openai?.voices;
|
voices = ttsSchema.openai?.voices;
|
||||||
break;
|
break;
|
||||||
case 'elevenlabs':
|
case TTSProviders.AZURE_OPENAI:
|
||||||
|
voices = ttsSchema.azureOpenAI?.voices;
|
||||||
|
break;
|
||||||
|
case TTSProviders.ELEVENLABS:
|
||||||
voices = ttsSchema.elevenlabs?.voices;
|
voices = ttsSchema.elevenlabs?.voices;
|
||||||
break;
|
break;
|
||||||
case 'localai':
|
case TTSProviders.LOCALAI:
|
||||||
voices = ttsSchema.localai?.voices;
|
voices = ttsSchema.localai?.voices;
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
|
|
|
||||||
|
|
@ -1,8 +1,9 @@
|
||||||
const axios = require('axios');
|
|
||||||
const { Readable } = require('stream');
|
const { Readable } = require('stream');
|
||||||
const { logger } = require('~/config');
|
const axios = require('axios');
|
||||||
|
const { extractEnvVariable, STTProviders } = require('librechat-data-provider');
|
||||||
const getCustomConfig = require('~/server/services/Config/getCustomConfig');
|
const getCustomConfig = require('~/server/services/Config/getCustomConfig');
|
||||||
const { extractEnvVariable } = require('librechat-data-provider');
|
const { genAzureEndpoint } = require('~/utils');
|
||||||
|
const { logger } = require('~/config');
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Handle the response from the STT API
|
* Handle the response from the STT API
|
||||||
|
|
@ -24,12 +25,34 @@ async function handleResponse(response) {
|
||||||
return response.data.text.trim();
|
return response.data.text.trim();
|
||||||
}
|
}
|
||||||
|
|
||||||
function getProvider(sttSchema) {
|
/**
|
||||||
if (sttSchema?.openai) {
|
* getProviderSchema function
|
||||||
return 'openai';
|
* This function takes the customConfig object and returns the name of the provider and its schema
|
||||||
|
* If more than one provider is set or no provider is set, it throws an error
|
||||||
|
*
|
||||||
|
* @param {Object} customConfig - The custom configuration containing the STT schema
|
||||||
|
* @returns {Promise<[string, Object]>} The name of the provider and its schema
|
||||||
|
* @throws {Error} Throws an error if multiple providers are set or no provider is set
|
||||||
|
*/
|
||||||
|
async function getProviderSchema(customConfig) {
|
||||||
|
const sttSchema = customConfig.speech.stt;
|
||||||
|
|
||||||
|
if (!sttSchema) {
|
||||||
|
throw new Error(`No STT schema is set. Did you configure STT in the custom config (librechat.yaml)?
|
||||||
|
|
||||||
|
https://www.librechat.ai/docs/configuration/stt_tts#stt`);
|
||||||
}
|
}
|
||||||
|
|
||||||
throw new Error('Invalid provider');
|
const providers = Object.entries(sttSchema).filter(([, value]) => Object.keys(value).length > 0);
|
||||||
|
|
||||||
|
if (providers.length > 1) {
|
||||||
|
throw new Error('Multiple providers are set. Please set only one provider.');
|
||||||
|
} else if (providers.length === 0) {
|
||||||
|
throw new Error('No provider is set. Please set a provider.');
|
||||||
|
} else {
|
||||||
|
const provider = providers[0][0];
|
||||||
|
return [provider, sttSchema[provider]];
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
function removeUndefined(obj) {
|
function removeUndefined(obj) {
|
||||||
|
|
@ -83,72 +106,63 @@ function openAIProvider(sttSchema, audioReadStream) {
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This function prepares the necessary data and headers for making a request to the Azure API
|
* Prepares the necessary data and headers for making a request to the Azure API.
|
||||||
* It uses the provided request and audio stream to create the request
|
* It uses the provided Speech-to-Text (STT) schema and audio file to create the request.
|
||||||
*
|
*
|
||||||
* @param {Object} req - The request object, which should contain the endpoint in its body
|
* @param {Object} sttSchema - The STT schema object, which should contain instanceName, deploymentName, apiVersion, and apiKey.
|
||||||
* @param {Stream} audioReadStream - The audio data to be transcribed
|
* @param {Buffer} audioBuffer - The audio data to be transcribed
|
||||||
|
* @param {Object} audioFile - The audio file object, which should contain originalname, mimetype, and size.
|
||||||
*
|
*
|
||||||
* @returns {Array} An array containing the URL for the API request, the data to be sent, and the headers for the request
|
* @returns {Array} An array containing the URL for the API request, the data to be sent, and the headers for the request.
|
||||||
* If an error occurs, it returns an array with three null values and logs the error with logger
|
* If an error occurs, it logs the error with logger and returns an array with three null values.
|
||||||
*/
|
*/
|
||||||
function azureProvider(req, audioReadStream) {
|
function azureOpenAIProvider(sttSchema, audioBuffer, audioFile) {
|
||||||
try {
|
try {
|
||||||
const { endpoint } = req.body;
|
const instanceName = sttSchema?.instanceName;
|
||||||
const azureConfig = req.app.locals[endpoint];
|
const deploymentName = sttSchema?.deploymentName;
|
||||||
|
const apiVersion = sttSchema?.apiVersion;
|
||||||
|
|
||||||
if (!azureConfig) {
|
const url =
|
||||||
throw new Error(`No configuration found for endpoint: ${endpoint}`);
|
genAzureEndpoint({
|
||||||
|
azureOpenAIApiInstanceName: instanceName,
|
||||||
|
azureOpenAIApiDeploymentName: deploymentName,
|
||||||
|
}) +
|
||||||
|
'/audio/transcriptions?api-version=' +
|
||||||
|
apiVersion;
|
||||||
|
|
||||||
|
const apiKey = sttSchema.apiKey ? extractEnvVariable(sttSchema.apiKey) : '';
|
||||||
|
|
||||||
|
if (audioBuffer.byteLength > 25 * 1024 * 1024) {
|
||||||
|
throw new Error('The audio file size exceeds the limit of 25MB');
|
||||||
|
}
|
||||||
|
const acceptedFormats = ['flac', 'mp3', 'mp4', 'mpeg', 'mpga', 'm4a', 'ogg', 'wav', 'webm'];
|
||||||
|
const fileFormat = audioFile.mimetype.split('/')[1];
|
||||||
|
if (!acceptedFormats.includes(fileFormat)) {
|
||||||
|
throw new Error(`The audio file format ${fileFormat} is not accepted`);
|
||||||
}
|
}
|
||||||
|
|
||||||
const { apiKey, instanceName, whisperModel, apiVersion } = Object.entries(
|
const formData = new FormData();
|
||||||
azureConfig.groupMap,
|
|
||||||
).reduce((acc, [, value]) => {
|
|
||||||
if (acc) {
|
|
||||||
return acc;
|
|
||||||
}
|
|
||||||
|
|
||||||
const whisperKey = Object.keys(value.models).find((modelKey) =>
|
const audioBlob = new Blob([audioBuffer], { type: audioFile.mimetype });
|
||||||
modelKey.startsWith('whisper'),
|
|
||||||
);
|
|
||||||
|
|
||||||
if (whisperKey) {
|
formData.append('file', audioBlob, audioFile.originalname);
|
||||||
return {
|
|
||||||
apiVersion: value.version,
|
|
||||||
apiKey: value.apiKey,
|
|
||||||
instanceName: value.instanceName,
|
|
||||||
whisperModel: value.models[whisperKey]['deploymentName'],
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
return null;
|
let data = formData;
|
||||||
}, null);
|
|
||||||
|
|
||||||
if (!apiKey || !instanceName || !whisperModel || !apiVersion) {
|
let headers = {
|
||||||
throw new Error('Required Azure configuration values are missing');
|
|
||||||
}
|
|
||||||
|
|
||||||
const baseURL = `https://${instanceName}.openai.azure.com`;
|
|
||||||
|
|
||||||
const url = `${baseURL}/openai/deployments/${whisperModel}/audio/transcriptions?api-version=${apiVersion}`;
|
|
||||||
|
|
||||||
let data = {
|
|
||||||
file: audioReadStream,
|
|
||||||
filename: 'audio.wav',
|
|
||||||
contentType: 'audio/wav',
|
|
||||||
knownLength: audioReadStream.length,
|
|
||||||
};
|
|
||||||
|
|
||||||
const headers = {
|
|
||||||
...data.getHeaders(),
|
|
||||||
'Content-Type': 'multipart/form-data',
|
'Content-Type': 'multipart/form-data',
|
||||||
'api-key': apiKey,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
|
[headers].forEach(removeUndefined);
|
||||||
|
|
||||||
|
if (apiKey) {
|
||||||
|
headers['api-key'] = apiKey;
|
||||||
|
}
|
||||||
|
|
||||||
return [url, data, headers];
|
return [url, data, headers];
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
logger.error('An error occurred while preparing the Azure API STT request: ', error);
|
logger.error('An error occurred while preparing the Azure OpenAI API STT request: ', error);
|
||||||
return [null, null, null];
|
throw error;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -176,16 +190,16 @@ async function speechToText(req, res) {
|
||||||
const audioReadStream = Readable.from(audioBuffer);
|
const audioReadStream = Readable.from(audioBuffer);
|
||||||
audioReadStream.path = 'audio.wav';
|
audioReadStream.path = 'audio.wav';
|
||||||
|
|
||||||
const provider = getProvider(customConfig.speech.stt);
|
const [provider, sttSchema] = await getProviderSchema(customConfig);
|
||||||
|
|
||||||
let [url, data, headers] = [];
|
let [url, data, headers] = [];
|
||||||
|
|
||||||
switch (provider) {
|
switch (provider) {
|
||||||
case 'openai':
|
case STTProviders.OPENAI:
|
||||||
[url, data, headers] = openAIProvider(customConfig.speech.stt, audioReadStream);
|
[url, data, headers] = openAIProvider(sttSchema, audioReadStream);
|
||||||
break;
|
break;
|
||||||
case 'azure':
|
case STTProviders.AZURE_OPENAI:
|
||||||
[url, data, headers] = azureProvider(req, audioReadStream);
|
[url, data, headers] = azureOpenAIProvider(sttSchema, audioBuffer, req.file);
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
throw new Error('Invalid provider');
|
throw new Error('Invalid provider');
|
||||||
|
|
|
||||||
|
|
@ -1,8 +1,9 @@
|
||||||
const axios = require('axios');
|
const axios = require('axios');
|
||||||
const getCustomConfig = require('~/server/services/Config/getCustomConfig');
|
const { extractEnvVariable, TTSProviders } = require('librechat-data-provider');
|
||||||
const { getRandomVoiceId, createChunkProcessor, splitTextIntoChunks } = require('./streamAudio');
|
|
||||||
const { extractEnvVariable } = require('librechat-data-provider');
|
|
||||||
const { logger } = require('~/config');
|
const { logger } = require('~/config');
|
||||||
|
const getCustomConfig = require('~/server/services/Config/getCustomConfig');
|
||||||
|
const { genAzureEndpoint } = require('~/utils');
|
||||||
|
const { getRandomVoiceId, createChunkProcessor, splitTextIntoChunks } = require('./streamAudio');
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* getProvider function
|
* getProvider function
|
||||||
|
|
@ -91,6 +92,59 @@ function openAIProvider(ttsSchema, input, voice) {
|
||||||
return [url, data, headers];
|
return [url, data, headers];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Generates the necessary parameters for making a request to Azure's OpenAI Text-to-Speech API.
|
||||||
|
*
|
||||||
|
* @param {TCustomConfig['tts']['azureOpenAI']} ttsSchema - The TTS schema containing the AzureOpenAI configuration
|
||||||
|
* @param {string} input - The text to be converted to speech
|
||||||
|
* @param {string} voice - The voice to be used for the speech
|
||||||
|
*
|
||||||
|
* @returns {Array} An array containing the URL for the API request, the data to be sent, and the headers for the request
|
||||||
|
* If an error occurs, it throws an error with a message indicating that the selected voice is not available
|
||||||
|
*/
|
||||||
|
function azureOpenAIProvider(ttsSchema, input, voice) {
|
||||||
|
const instanceName = ttsSchema?.instanceName;
|
||||||
|
const deploymentName = ttsSchema?.deploymentName;
|
||||||
|
const apiVersion = ttsSchema?.apiVersion;
|
||||||
|
|
||||||
|
const url =
|
||||||
|
genAzureEndpoint({
|
||||||
|
azureOpenAIApiInstanceName: instanceName,
|
||||||
|
azureOpenAIApiDeploymentName: deploymentName,
|
||||||
|
}) +
|
||||||
|
'/audio/speech?api-version=' +
|
||||||
|
apiVersion;
|
||||||
|
|
||||||
|
const apiKey = ttsSchema.apiKey ? extractEnvVariable(ttsSchema.apiKey) : '';
|
||||||
|
|
||||||
|
if (
|
||||||
|
ttsSchema?.voices &&
|
||||||
|
ttsSchema.voices.length > 0 &&
|
||||||
|
!ttsSchema.voices.includes(voice) &&
|
||||||
|
!ttsSchema.voices.includes('ALL')
|
||||||
|
) {
|
||||||
|
throw new Error(`Voice ${voice} is not available.`);
|
||||||
|
}
|
||||||
|
|
||||||
|
let data = {
|
||||||
|
model: ttsSchema?.model,
|
||||||
|
input,
|
||||||
|
voice: ttsSchema?.voices && ttsSchema.voices.length > 0 ? voice : undefined,
|
||||||
|
};
|
||||||
|
|
||||||
|
let headers = {
|
||||||
|
'Content-Type': 'application/json',
|
||||||
|
};
|
||||||
|
|
||||||
|
[data, headers].forEach(removeUndefined);
|
||||||
|
|
||||||
|
if (apiKey) {
|
||||||
|
headers['api-key'] = apiKey;
|
||||||
|
}
|
||||||
|
|
||||||
|
return [url, data, headers];
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* elevenLabsProvider function
|
* elevenLabsProvider function
|
||||||
* This function prepares the necessary data and headers for making a request to the Eleven Labs TTS
|
* This function prepares the necessary data and headers for making a request to the Eleven Labs TTS
|
||||||
|
|
@ -225,13 +279,16 @@ async function getVoice(providerSchema, requestVoice) {
|
||||||
async function ttsRequest(provider, ttsSchema, { input, voice, stream = true } = { stream: true }) {
|
async function ttsRequest(provider, ttsSchema, { input, voice, stream = true } = { stream: true }) {
|
||||||
let [url, data, headers] = [];
|
let [url, data, headers] = [];
|
||||||
switch (provider) {
|
switch (provider) {
|
||||||
case 'openai':
|
case TTSProviders.OPENAI:
|
||||||
[url, data, headers] = openAIProvider(ttsSchema, input, voice);
|
[url, data, headers] = openAIProvider(ttsSchema, input, voice);
|
||||||
break;
|
break;
|
||||||
case 'elevenlabs':
|
case TTSProviders.AZURE_OPENAI:
|
||||||
|
[url, data, headers] = azureOpenAIProvider(ttsSchema, input, voice);
|
||||||
|
break;
|
||||||
|
case TTSProviders.ELEVENLABS:
|
||||||
[url, data, headers] = elevenLabsProvider(ttsSchema, input, voice, stream);
|
[url, data, headers] = elevenLabsProvider(ttsSchema, input, voice, stream);
|
||||||
break;
|
break;
|
||||||
case 'localai':
|
case TTSProviders.LOCALAI:
|
||||||
[url, data, headers] = localAIProvider(ttsSchema, input, voice);
|
[url, data, headers] = localAIProvider(ttsSchema, input, voice);
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
|
|
|
||||||
|
|
@ -6,7 +6,6 @@ import { fileConfigSchema } from './file-config';
|
||||||
import { specsConfigSchema } from './models';
|
import { specsConfigSchema } from './models';
|
||||||
import { FileSources } from './types/files';
|
import { FileSources } from './types/files';
|
||||||
import { TModelsConfig } from './types';
|
import { TModelsConfig } from './types';
|
||||||
import { speech } from './api-endpoints';
|
|
||||||
|
|
||||||
export const defaultSocialLogins = ['google', 'facebook', 'openid', 'github', 'discord'];
|
export const defaultSocialLogins = ['google', 'facebook', 'openid', 'github', 'discord'];
|
||||||
|
|
||||||
|
|
@ -234,6 +233,15 @@ const ttsOpenaiSchema = z.object({
|
||||||
voices: z.array(z.string()),
|
voices: z.array(z.string()),
|
||||||
});
|
});
|
||||||
|
|
||||||
|
const ttsAzureOpenAISchema = z.object({
|
||||||
|
instanceName: z.string(),
|
||||||
|
apiKey: z.string(),
|
||||||
|
deploymentName: z.string(),
|
||||||
|
apiVersion: z.string(),
|
||||||
|
model: z.string(),
|
||||||
|
voices: z.array(z.string()),
|
||||||
|
});
|
||||||
|
|
||||||
const ttsElevenLabsSchema = z.object({
|
const ttsElevenLabsSchema = z.object({
|
||||||
url: z.string().optional(),
|
url: z.string().optional(),
|
||||||
websocketUrl: z.string().optional(),
|
websocketUrl: z.string().optional(),
|
||||||
|
|
@ -260,18 +268,27 @@ const ttsLocalaiSchema = z.object({
|
||||||
|
|
||||||
const ttsSchema = z.object({
|
const ttsSchema = z.object({
|
||||||
openai: ttsOpenaiSchema.optional(),
|
openai: ttsOpenaiSchema.optional(),
|
||||||
|
azureOpenAI: ttsAzureOpenAISchema.optional(),
|
||||||
elevenLabs: ttsElevenLabsSchema.optional(),
|
elevenLabs: ttsElevenLabsSchema.optional(),
|
||||||
localai: ttsLocalaiSchema.optional(),
|
localai: ttsLocalaiSchema.optional(),
|
||||||
});
|
});
|
||||||
|
|
||||||
const sttSchema = z.object({
|
const sttOpenaiSchema = z.object({
|
||||||
openai: z
|
|
||||||
.object({
|
|
||||||
url: z.string().optional(),
|
url: z.string().optional(),
|
||||||
apiKey: z.string().optional(),
|
apiKey: z.string(),
|
||||||
model: z.string().optional(),
|
model: z.string(),
|
||||||
})
|
});
|
||||||
.optional(),
|
|
||||||
|
const sttAzureOpenAISchema = z.object({
|
||||||
|
instanceName: z.string(),
|
||||||
|
apiKey: z.string(),
|
||||||
|
deploymentName: z.string(),
|
||||||
|
apiVersion: z.string(),
|
||||||
|
});
|
||||||
|
|
||||||
|
const sttSchema = z.object({
|
||||||
|
openai: sttOpenaiSchema.optional(),
|
||||||
|
azureOpenAI: sttAzureOpenAISchema.optional(),
|
||||||
});
|
});
|
||||||
|
|
||||||
const speechTab = z
|
const speechTab = z
|
||||||
|
|
@ -846,6 +863,36 @@ export enum SettingsTabValues {
|
||||||
ACCOUNT = 'account',
|
ACCOUNT = 'account',
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export enum STTProviders {
|
||||||
|
/**
|
||||||
|
* Provider for OpenAI STT
|
||||||
|
*/
|
||||||
|
OPENAI = 'openai',
|
||||||
|
/**
|
||||||
|
* Provider for Microsoft Azure STT
|
||||||
|
*/
|
||||||
|
AZURE_OPENAI = 'azureOpenAI',
|
||||||
|
}
|
||||||
|
|
||||||
|
export enum TTSProviders {
|
||||||
|
/**
|
||||||
|
* Provider for OpenAI TTS
|
||||||
|
*/
|
||||||
|
OPENAI = 'openai',
|
||||||
|
/**
|
||||||
|
* Provider for Microsoft Azure OpenAI TTS
|
||||||
|
*/
|
||||||
|
AZURE_OPENAI = 'azureOpenAI',
|
||||||
|
/**
|
||||||
|
* Provider for ElevenLabs TTS
|
||||||
|
*/
|
||||||
|
ELEVENLABS = 'elevenlabs',
|
||||||
|
/**
|
||||||
|
* Provider for LocalAI TTS
|
||||||
|
*/
|
||||||
|
LOCALAI = 'localai',
|
||||||
|
}
|
||||||
|
|
||||||
/** Enum for app-wide constants */
|
/** Enum for app-wide constants */
|
||||||
export enum Constants {
|
export enum Constants {
|
||||||
/** Key for the app's version. */
|
/** Key for the app's version. */
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue