40 lines
1.4 KiB
TypeScript
40 lines
1.4 KiB
TypeScript
import type { ClientOptions } from 'openai'
|
|
import { ChatOpenAI as LangchainChatOpenAI, OpenAIChatInput, LegacyOpenAIInput, AzureOpenAIInput } from '@langchain/openai'
|
|
import { BaseChatModelParams } from '@langchain/core/language_models/chat_models'
|
|
import { IMultiModalOption, IVisionChatModal } from '../../../src'
|
|
|
|
export class ChatOpenAI extends LangchainChatOpenAI implements IVisionChatModal {
|
|
configuredModel: string
|
|
configuredMaxToken: number
|
|
multiModalOption: IMultiModalOption
|
|
id: string
|
|
|
|
constructor(
|
|
id: string,
|
|
fields?: Partial<OpenAIChatInput> &
|
|
Partial<AzureOpenAIInput> &
|
|
BaseChatModelParams & { configuration?: ClientOptions & LegacyOpenAIInput },
|
|
/** @deprecated */
|
|
configuration?: ClientOptions & LegacyOpenAIInput
|
|
) {
|
|
super(fields, configuration)
|
|
this.id = id
|
|
this.configuredModel = fields?.modelName ?? 'gpt-3.5-turbo'
|
|
this.configuredMaxToken = fields?.maxTokens ?? 256
|
|
}
|
|
|
|
revertToOriginalModel(): void {
|
|
super.modelName = this.configuredModel
|
|
super.maxTokens = this.configuredMaxToken
|
|
}
|
|
|
|
setMultiModalOption(multiModalOption: IMultiModalOption): void {
|
|
this.multiModalOption = multiModalOption
|
|
}
|
|
|
|
setVisionModel(): void {
|
|
super.modelName = 'gpt-4-vision-preview'
|
|
super.maxTokens = 1024
|
|
}
|
|
}
|