ci(backend-review.yml): add linter step to the backend review workflow (#625)

* ci(backend-review.yml): add linter step to the backend review workflow

* chore(backend-review.yml): remove prettier from lint-action configuration

* chore: apply new linting workflow

* chore(lint-staged.config.js): reorder lint-staged tasks for JavaScript and TypeScript files

* chore(eslint): update ignorePatterns in .eslintrc.js
chore(lint-action): remove prettier option in backend-review.yml
chore(package.json): add lint and lint:fix scripts

* chore(lint-staged.config.js): remove prettier --write command for js, jsx, ts, tsx files

* chore(titleConvo.js): remove unnecessary console.log statement
chore(titleConvo.js): add missing comma in options object

* chore: apply linting to all files

* chore(lint-staged.config.js): update lint-staged configuration to include prettier formatting
This commit is contained in:
Danny Avila 2023-07-14 09:36:49 -04:00 committed by GitHub
parent 637bb6bc11
commit e5336039fc
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
231 changed files with 1688 additions and 1526 deletions

View file

@ -3,7 +3,7 @@ const { google } = require('googleapis');
const { Agent, ProxyAgent } = require('undici');
const {
encoding_for_model: encodingForModel,
get_encoding: getEncoding
get_encoding: getEncoding,
} = require('@dqbd/tiktoken');
const tokenizersCache = {};
@ -43,20 +43,20 @@ class GoogleClient extends BaseClient {
// nested options aren't spread properly, so we need to do this manually
this.options.modelOptions = {
...this.options.modelOptions,
...options.modelOptions
...options.modelOptions,
};
delete options.modelOptions;
// now we can merge options
this.options = {
...this.options,
...options
...options,
};
} else {
this.options = options;
}
this.options.examples = this.options.examples.filter(
(obj) => obj.input.content !== '' && obj.output.content !== ''
(obj) => obj.input.content !== '' && obj.output.content !== '',
);
const modelOptions = this.options.modelOptions || {};
@ -66,7 +66,7 @@ class GoogleClient extends BaseClient {
model: modelOptions.model || 'chat-bison',
temperature: typeof modelOptions.temperature === 'undefined' ? 0.2 : modelOptions.temperature, // 0 - 1, 0.2 is recommended
topP: typeof modelOptions.topP === 'undefined' ? 0.95 : modelOptions.topP, // 0 - 1, default: 0.95
topK: typeof modelOptions.topK === 'undefined' ? 40 : modelOptions.topK // 1-40, default: 40
topK: typeof modelOptions.topK === 'undefined' ? 40 : modelOptions.topK, // 1-40, default: 40
// stop: modelOptions.stop // no stop method for now
};
@ -86,7 +86,7 @@ class GoogleClient extends BaseClient {
throw new Error(
`maxPromptTokens + maxOutputTokens (${this.maxPromptTokens} + ${this.maxResponseTokens} = ${
this.maxPromptTokens + this.maxResponseTokens
}) must be less than or equal to maxContextTokens (${this.maxContextTokens})`
}) must be less than or equal to maxContextTokens (${this.maxContextTokens})`,
);
}
@ -105,7 +105,7 @@ class GoogleClient extends BaseClient {
this.endToken = '<|im_end|>';
this.gptEncoder = this.constructor.getTokenizer('text-davinci-003', true, {
'<|im_start|>': 100264,
'<|im_end|>': 100265
'<|im_end|>': 100265,
});
} else {
// Previously I was trying to use "<|endoftext|>" but there seems to be some bug with OpenAI's token counting
@ -143,7 +143,7 @@ class GoogleClient extends BaseClient {
getMessageMapMethod() {
return ((message) => ({
author: message?.author ?? (message.isCreatedByUser ? this.userLabel : this.modelLabel),
content: message?.content ?? message.text
content: message?.content ?? message.text,
})).bind(this);
}
@ -153,9 +153,9 @@ class GoogleClient extends BaseClient {
instances: [
{
messages: formattedMessages,
}
},
],
parameters: this.options.modelOptions
parameters: this.options.modelOptions,
};
if (this.options.promptPrefix) {
@ -170,8 +170,8 @@ class GoogleClient extends BaseClient {
if (this.isTextModel) {
payload.instances = [
{
prompt: messages[messages.length -1].content
}
prompt: messages[messages.length - 1].content,
},
];
}
@ -199,9 +199,9 @@ class GoogleClient extends BaseClient {
method: 'POST',
agent: new Agent({
bodyTimeout: 0,
headersTimeout: 0
headersTimeout: 0,
}),
signal: abortController.signal
signal: abortController.signal,
};
if (this.options.proxy) {
@ -218,7 +218,7 @@ class GoogleClient extends BaseClient {
return {
promptPrefix: this.options.promptPrefix,
modelLabel: this.options.modelLabel,
...this.modelOptions
...this.modelOptions,
};
}
@ -239,7 +239,7 @@ class GoogleClient extends BaseClient {
'';
if (blocked === true) {
reply = `Google blocked a proper response to your message:\n${JSON.stringify(
result.predictions[0].safetyAttributes
result.predictions[0].safetyAttributes,
)}${reply.length > 0 ? `\nAI Response:\n${reply}` : ''}`;
}
if (this.options.debug) {