import { FlowiseSummaryMemory, IMessage, IDatabaseEntity, INode, INodeData, INodeParams, MemoryMethods, ICommonObject } from '../../../src/Interface' import { getBaseClasses, mapChatMessageToBaseMessage } from '../../../src/utils' import { BaseLanguageModel } from '@langchain/core/language_models/base' import { BaseMessage, HumanMessage, SystemMessage } from '@langchain/core/messages' import { ConversationSummaryMemory, ConversationSummaryMemoryInput } from 'langchain/memory' import { DataSource } from 'typeorm' import { ChatAnthropic } from '../../chatmodels/ChatAnthropic/FlowiseChatAnthropic' class ConversationSummaryMemory_Memory implements INode { label: string name: string version: number description: string type: string icon: string category: string baseClasses: string[] inputs: INodeParams[] constructor() { this.label = 'Conversation Summary Memory' this.name = 'conversationSummaryMemory' this.version = 2.0 this.type = 'ConversationSummaryMemory' this.icon = 'memory.svg' this.category = 'Memory' this.description = 'Summarizes the conversation and stores the current summary in memory' this.baseClasses = [this.type, ...getBaseClasses(ConversationSummaryMemory)] this.inputs = [ { label: 'Chat Model', name: 'model', type: 'BaseChatModel' }, { label: 'Session Id', name: 'sessionId', type: 'string', description: 'If not specified, a random id will be used. Learn more', default: '', optional: true, additionalParams: true }, { label: 'Memory Key', name: 'memoryKey', type: 'string', default: 'chat_history', additionalParams: true } ] } async init(nodeData: INodeData, _: string, options: ICommonObject): Promise { const model = nodeData.inputs?.model as BaseLanguageModel const sessionId = nodeData.inputs?.sessionId as string const memoryKey = (nodeData.inputs?.memoryKey as string) ?? 'chat_history' const appDataSource = options.appDataSource as DataSource const databaseEntities = options.databaseEntities as IDatabaseEntity const chatflowid = options.chatflowid as string const obj: ConversationSummaryMemoryInput & BufferMemoryExtendedInput = { llm: model, memoryKey, returnMessages: true, sessionId, appDataSource, databaseEntities, chatflowid } return new ConversationSummaryMemoryExtended(obj) } } interface BufferMemoryExtendedInput { sessionId: string appDataSource: DataSource databaseEntities: IDatabaseEntity chatflowid: string } class ConversationSummaryMemoryExtended extends FlowiseSummaryMemory implements MemoryMethods { appDataSource: DataSource databaseEntities: IDatabaseEntity chatflowid: string sessionId = '' constructor(fields: ConversationSummaryMemoryInput & BufferMemoryExtendedInput) { super(fields) this.sessionId = fields.sessionId this.appDataSource = fields.appDataSource this.databaseEntities = fields.databaseEntities this.chatflowid = fields.chatflowid } async getChatMessages( overrideSessionId = '', returnBaseMessages = false, prependMessages?: IMessage[] ): Promise { const id = overrideSessionId ? overrideSessionId : this.sessionId if (!id) return [] this.buffer = '' let chatMessage = await this.appDataSource.getRepository(this.databaseEntities['ChatMessage']).find({ where: { sessionId: id, chatflowid: this.chatflowid }, order: { createdDate: 'ASC' } }) if (prependMessages?.length) { chatMessage.unshift(...prependMessages) } const baseMessages = await mapChatMessageToBaseMessage(chatMessage) // Get summary if (this.llm && typeof this.llm !== 'string') { this.buffer = baseMessages.length ? await this.predictNewSummary(baseMessages.slice(-2), this.buffer) : '' } if (returnBaseMessages) { // Anthropic doesn't support multiple system messages if (this.llm instanceof ChatAnthropic) { return [new HumanMessage(`Below is the summarized conversation:\n\n${this.buffer}`)] } else { return [new SystemMessage(this.buffer)] } } if (this.buffer) { return [ { message: this.buffer, type: 'apiMessage' } ] } let returnIMessages: IMessage[] = [] for (const m of chatMessage) { returnIMessages.push({ message: m.content as string, type: m.role }) } return returnIMessages } async addChatMessages(): Promise { // adding chat messages is done on server level return } async clearChatMessages(): Promise { // clearing chat messages is done on server level return } } module.exports = { nodeClass: ConversationSummaryMemory_Memory }