106 lines
3.7 KiB
TypeScript
106 lines
3.7 KiB
TypeScript
import { FlowiseSummaryMemory, IMessage, INode, INodeData, INodeParams, MemoryMethods, MessageType } from '../../../src/Interface'
|
|
import { convertBaseMessagetoIMessage, getBaseClasses } from '../../../src/utils'
|
|
import { ConversationSummaryMemory, ConversationSummaryMemoryInput } from 'langchain/memory'
|
|
import { BaseLanguageModel } from 'langchain/base_language'
|
|
import { BaseMessage } from 'langchain/schema'
|
|
|
|
class ConversationSummaryMemory_Memory implements INode {
|
|
label: string
|
|
name: string
|
|
version: number
|
|
description: string
|
|
type: string
|
|
icon: string
|
|
category: string
|
|
baseClasses: string[]
|
|
inputs: INodeParams[]
|
|
|
|
constructor() {
|
|
this.label = 'Conversation Summary Memory'
|
|
this.name = 'conversationSummaryMemory'
|
|
this.version = 1.0
|
|
this.type = 'ConversationSummaryMemory'
|
|
this.icon = 'memory.svg'
|
|
this.category = 'Memory'
|
|
this.description = 'Summarizes the conversation and stores the current summary in memory'
|
|
this.baseClasses = [this.type, ...getBaseClasses(ConversationSummaryMemory)]
|
|
this.inputs = [
|
|
{
|
|
label: 'Chat Model',
|
|
name: 'model',
|
|
type: 'BaseChatModel'
|
|
},
|
|
{
|
|
label: 'Memory Key',
|
|
name: 'memoryKey',
|
|
type: 'string',
|
|
default: 'chat_history'
|
|
},
|
|
{
|
|
label: 'Input Key',
|
|
name: 'inputKey',
|
|
type: 'string',
|
|
default: 'input'
|
|
}
|
|
]
|
|
}
|
|
|
|
async init(nodeData: INodeData): Promise<any> {
|
|
const model = nodeData.inputs?.model as BaseLanguageModel
|
|
const memoryKey = nodeData.inputs?.memoryKey as string
|
|
const inputKey = nodeData.inputs?.inputKey as string
|
|
|
|
const obj: ConversationSummaryMemoryInput = {
|
|
llm: model,
|
|
returnMessages: true,
|
|
memoryKey,
|
|
inputKey
|
|
}
|
|
|
|
return new ConversationSummaryMemoryExtended(obj)
|
|
}
|
|
}
|
|
|
|
class ConversationSummaryMemoryExtended extends FlowiseSummaryMemory implements MemoryMethods {
|
|
constructor(fields: ConversationSummaryMemoryInput) {
|
|
super(fields)
|
|
}
|
|
|
|
async getChatMessages(_?: string, returnBaseMessages = false): Promise<IMessage[] | BaseMessage[]> {
|
|
const memoryResult = await this.loadMemoryVariables({})
|
|
const baseMessages = memoryResult[this.memoryKey ?? 'chat_history']
|
|
return returnBaseMessages ? baseMessages : convertBaseMessagetoIMessage(baseMessages)
|
|
}
|
|
|
|
async addChatMessages(msgArray: { text: string; type: MessageType }[]): Promise<void> {
|
|
const input = msgArray.find((msg) => msg.type === 'userMessage')
|
|
const output = msgArray.find((msg) => msg.type === 'apiMessage')
|
|
|
|
const inputValues = { [this.inputKey ?? 'input']: input?.text }
|
|
const outputValues = { output: output?.text }
|
|
|
|
await this.saveContext(inputValues, outputValues)
|
|
}
|
|
|
|
async clearChatMessages(): Promise<void> {
|
|
await this.clear()
|
|
}
|
|
|
|
async resumeMessages(messages: IMessage[]): Promise<void> {
|
|
// Clear existing chatHistory to avoid duplication
|
|
if (messages.length) await this.clear()
|
|
|
|
// Insert into chatHistory
|
|
for (const msg of messages) {
|
|
if (msg.type === 'userMessage') await this.chatHistory.addUserMessage(msg.message)
|
|
else if (msg.type === 'apiMessage') await this.chatHistory.addAIChatMessage(msg.message)
|
|
}
|
|
|
|
// Replace buffer
|
|
const chatMessages = await this.chatHistory.getMessages()
|
|
this.buffer = await this.predictNewSummary(chatMessages.slice(-2), this.buffer)
|
|
}
|
|
}
|
|
|
|
module.exports = { nodeClass: ConversationSummaryMemory_Memory }
|