add fixes to multiple documents
This commit is contained in:
parent
9f53111aac
commit
324bd3a430
|
|
@ -1,7 +1,6 @@
|
||||||
import { INode, INodeData, INodeParams } from '../../../src/Interface'
|
import { INode, INodeData, INodeParams } from '../../../src/Interface'
|
||||||
import { TextSplitter } from 'langchain/text_splitter'
|
import { TextSplitter } from 'langchain/text_splitter'
|
||||||
import { CSVLoader } from 'langchain/document_loaders/fs/csv'
|
import { CSVLoader } from 'langchain/document_loaders/fs/csv'
|
||||||
import { getBlob } from '../../../src/utils'
|
|
||||||
|
|
||||||
class Csv_DocumentLoaders implements INode {
|
class Csv_DocumentLoaders implements INode {
|
||||||
label: string
|
label: string
|
||||||
|
|
@ -58,20 +57,35 @@ class Csv_DocumentLoaders implements INode {
|
||||||
const columnName = nodeData.inputs?.columnName as string
|
const columnName = nodeData.inputs?.columnName as string
|
||||||
const metadata = nodeData.inputs?.metadata
|
const metadata = nodeData.inputs?.metadata
|
||||||
|
|
||||||
const blob = new Blob(getBlob(csvFileBase64))
|
let alldocs = []
|
||||||
const loader = new CSVLoader(blob, columnName.trim().length === 0 ? undefined : columnName.trim())
|
let files: string[] = []
|
||||||
let docs = []
|
|
||||||
|
|
||||||
if (textSplitter) {
|
if (csvFileBase64.startsWith('[') && csvFileBase64.endsWith(']')) {
|
||||||
docs = await loader.loadAndSplit(textSplitter)
|
files = JSON.parse(csvFileBase64)
|
||||||
} else {
|
} else {
|
||||||
docs = await loader.load()
|
files = [csvFileBase64]
|
||||||
|
}
|
||||||
|
|
||||||
|
for (const file of files) {
|
||||||
|
const splitDataURI = file.split(',')
|
||||||
|
splitDataURI.pop()
|
||||||
|
const bf = Buffer.from(splitDataURI.pop() || '', 'base64')
|
||||||
|
const blob = new Blob([bf])
|
||||||
|
const loader = new CSVLoader(blob, columnName.trim().length === 0 ? undefined : columnName.trim())
|
||||||
|
|
||||||
|
if (textSplitter) {
|
||||||
|
const docs = await loader.loadAndSplit(textSplitter)
|
||||||
|
alldocs.push(...docs)
|
||||||
|
} else {
|
||||||
|
const docs = await loader.load()
|
||||||
|
alldocs.push(...docs)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (metadata) {
|
if (metadata) {
|
||||||
const parsedMetadata = typeof metadata === 'object' ? metadata : JSON.parse(metadata)
|
const parsedMetadata = typeof metadata === 'object' ? metadata : JSON.parse(metadata)
|
||||||
let finaldocs = []
|
let finaldocs = []
|
||||||
for (const doc of docs) {
|
for (const doc of alldocs) {
|
||||||
const newdoc = {
|
const newdoc = {
|
||||||
...doc,
|
...doc,
|
||||||
metadata: {
|
metadata: {
|
||||||
|
|
@ -84,7 +98,7 @@ class Csv_DocumentLoaders implements INode {
|
||||||
return finaldocs
|
return finaldocs
|
||||||
}
|
}
|
||||||
|
|
||||||
return docs
|
return alldocs
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -1,7 +1,6 @@
|
||||||
import { INode, INodeData, INodeParams } from '../../../src/Interface'
|
import { INode, INodeData, INodeParams } from '../../../src/Interface'
|
||||||
import { TextSplitter } from 'langchain/text_splitter'
|
import { TextSplitter } from 'langchain/text_splitter'
|
||||||
import { DocxLoader } from 'langchain/document_loaders/fs/docx'
|
import { DocxLoader } from 'langchain/document_loaders/fs/docx'
|
||||||
import { getBlob } from '../../../src/utils'
|
|
||||||
|
|
||||||
class Docx_DocumentLoaders implements INode {
|
class Docx_DocumentLoaders implements INode {
|
||||||
label: string
|
label: string
|
||||||
|
|
@ -49,20 +48,35 @@ class Docx_DocumentLoaders implements INode {
|
||||||
const docxFileBase64 = nodeData.inputs?.docxFile as string
|
const docxFileBase64 = nodeData.inputs?.docxFile as string
|
||||||
const metadata = nodeData.inputs?.metadata
|
const metadata = nodeData.inputs?.metadata
|
||||||
|
|
||||||
const blob = new Blob(getBlob(docxFileBase64))
|
let alldocs = []
|
||||||
const loader = new DocxLoader(blob)
|
let files: string[] = []
|
||||||
let docs = []
|
|
||||||
|
|
||||||
if (textSplitter) {
|
if (docxFileBase64.startsWith('[') && docxFileBase64.endsWith(']')) {
|
||||||
docs = await loader.loadAndSplit(textSplitter)
|
files = JSON.parse(docxFileBase64)
|
||||||
} else {
|
} else {
|
||||||
docs = await loader.load()
|
files = [docxFileBase64]
|
||||||
|
}
|
||||||
|
|
||||||
|
for (const file of files) {
|
||||||
|
const splitDataURI = file.split(',')
|
||||||
|
splitDataURI.pop()
|
||||||
|
const bf = Buffer.from(splitDataURI.pop() || '', 'base64')
|
||||||
|
const blob = new Blob([bf])
|
||||||
|
const loader = new DocxLoader(blob)
|
||||||
|
|
||||||
|
if (textSplitter) {
|
||||||
|
const docs = await loader.loadAndSplit(textSplitter)
|
||||||
|
alldocs.push(...docs)
|
||||||
|
} else {
|
||||||
|
const docs = await loader.load()
|
||||||
|
alldocs.push(...docs)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (metadata) {
|
if (metadata) {
|
||||||
const parsedMetadata = typeof metadata === 'object' ? metadata : JSON.parse(metadata)
|
const parsedMetadata = typeof metadata === 'object' ? metadata : JSON.parse(metadata)
|
||||||
let finaldocs = []
|
let finaldocs = []
|
||||||
for (const doc of docs) {
|
for (const doc of alldocs) {
|
||||||
const newdoc = {
|
const newdoc = {
|
||||||
...doc,
|
...doc,
|
||||||
metadata: {
|
metadata: {
|
||||||
|
|
@ -75,7 +89,7 @@ class Docx_DocumentLoaders implements INode {
|
||||||
return finaldocs
|
return finaldocs
|
||||||
}
|
}
|
||||||
|
|
||||||
return docs
|
return alldocs
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -1,7 +1,6 @@
|
||||||
import { INode, INodeData, INodeParams } from '../../../src/Interface'
|
import { INode, INodeData, INodeParams } from '../../../src/Interface'
|
||||||
import { TextSplitter } from 'langchain/text_splitter'
|
import { TextSplitter } from 'langchain/text_splitter'
|
||||||
import { JSONLoader } from 'langchain/document_loaders/fs/json'
|
import { JSONLoader } from 'langchain/document_loaders/fs/json'
|
||||||
import { getBlob } from '../../../src/utils'
|
|
||||||
|
|
||||||
class Json_DocumentLoaders implements INode {
|
class Json_DocumentLoaders implements INode {
|
||||||
label: string
|
label: string
|
||||||
|
|
@ -64,20 +63,35 @@ class Json_DocumentLoaders implements INode {
|
||||||
pointers = outputString.split(',').map((pointer) => '/' + pointer.trim())
|
pointers = outputString.split(',').map((pointer) => '/' + pointer.trim())
|
||||||
}
|
}
|
||||||
|
|
||||||
const blob = new Blob(getBlob(jsonFileBase64))
|
let alldocs = []
|
||||||
const loader = new JSONLoader(blob, pointers.length != 0 ? pointers : undefined)
|
let files: string[] = []
|
||||||
let docs = []
|
|
||||||
|
|
||||||
if (textSplitter) {
|
if (jsonFileBase64.startsWith('[') && jsonFileBase64.endsWith(']')) {
|
||||||
docs = await loader.loadAndSplit(textSplitter)
|
files = JSON.parse(jsonFileBase64)
|
||||||
} else {
|
} else {
|
||||||
docs = await loader.load()
|
files = [jsonFileBase64]
|
||||||
|
}
|
||||||
|
|
||||||
|
for (const file of files) {
|
||||||
|
const splitDataURI = file.split(',')
|
||||||
|
splitDataURI.pop()
|
||||||
|
const bf = Buffer.from(splitDataURI.pop() || '', 'base64')
|
||||||
|
const blob = new Blob([bf])
|
||||||
|
const loader = new JSONLoader(blob, pointers.length != 0 ? pointers : undefined)
|
||||||
|
|
||||||
|
if (textSplitter) {
|
||||||
|
const docs = await loader.loadAndSplit(textSplitter)
|
||||||
|
alldocs.push(...docs)
|
||||||
|
} else {
|
||||||
|
const docs = await loader.load()
|
||||||
|
alldocs.push(...docs)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (metadata) {
|
if (metadata) {
|
||||||
const parsedMetadata = typeof metadata === 'object' ? metadata : JSON.parse(metadata)
|
const parsedMetadata = typeof metadata === 'object' ? metadata : JSON.parse(metadata)
|
||||||
let finaldocs = []
|
let finaldocs = []
|
||||||
for (const doc of docs) {
|
for (const doc of alldocs) {
|
||||||
const newdoc = {
|
const newdoc = {
|
||||||
...doc,
|
...doc,
|
||||||
metadata: {
|
metadata: {
|
||||||
|
|
@ -90,7 +104,7 @@ class Json_DocumentLoaders implements INode {
|
||||||
return finaldocs
|
return finaldocs
|
||||||
}
|
}
|
||||||
|
|
||||||
return docs
|
return alldocs
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -1,7 +1,6 @@
|
||||||
import { INode, INodeData, INodeParams } from '../../../src/Interface'
|
import { INode, INodeData, INodeParams } from '../../../src/Interface'
|
||||||
import { TextSplitter } from 'langchain/text_splitter'
|
import { TextSplitter } from 'langchain/text_splitter'
|
||||||
import { PDFLoader } from 'langchain/document_loaders/fs/pdf'
|
import { PDFLoader } from 'langchain/document_loaders/fs/pdf'
|
||||||
import { getBlob } from '../../../src/utils'
|
|
||||||
|
|
||||||
class Pdf_DocumentLoaders implements INode {
|
class Pdf_DocumentLoaders implements INode {
|
||||||
label: string
|
label: string
|
||||||
|
|
@ -66,30 +65,49 @@ class Pdf_DocumentLoaders implements INode {
|
||||||
const usage = nodeData.inputs?.usage as string
|
const usage = nodeData.inputs?.usage as string
|
||||||
const metadata = nodeData.inputs?.metadata
|
const metadata = nodeData.inputs?.metadata
|
||||||
|
|
||||||
const blob = new Blob(getBlob(pdfFileBase64))
|
let alldocs = []
|
||||||
let docs = []
|
let files: string[] = []
|
||||||
if (usage === 'perFile') {
|
|
||||||
// @ts-ignore
|
if (pdfFileBase64.startsWith('[') && pdfFileBase64.endsWith(']')) {
|
||||||
const loader = new PDFLoader(blob, { splitPages: false, pdfjs: () => import('pdf-parse/lib/pdf.js/v1.10.100/build/pdf.js') })
|
files = JSON.parse(pdfFileBase64)
|
||||||
if (textSplitter) {
|
|
||||||
docs = await loader.loadAndSplit(textSplitter)
|
|
||||||
} else {
|
|
||||||
docs = await loader.load()
|
|
||||||
}
|
|
||||||
} else {
|
} else {
|
||||||
// @ts-ignore
|
files = [pdfFileBase64]
|
||||||
const loader = new PDFLoader(blob, { pdfjs: () => import('pdf-parse/lib/pdf.js/v1.10.100/build/pdf.js') })
|
}
|
||||||
if (textSplitter) {
|
|
||||||
docs = await loader.loadAndSplit(textSplitter)
|
for (const file of files) {
|
||||||
|
const splitDataURI = file.split(',')
|
||||||
|
splitDataURI.pop()
|
||||||
|
const bf = Buffer.from(splitDataURI.pop() || '', 'base64')
|
||||||
|
if (usage === 'perFile') {
|
||||||
|
const loader = new PDFLoader(new Blob([bf]), {
|
||||||
|
splitPages: false,
|
||||||
|
// @ts-ignore
|
||||||
|
pdfjs: () => import('pdf-parse/lib/pdf.js/v1.10.100/build/pdf.js')
|
||||||
|
})
|
||||||
|
if (textSplitter) {
|
||||||
|
const docs = await loader.loadAndSplit(textSplitter)
|
||||||
|
alldocs.push(...docs)
|
||||||
|
} else {
|
||||||
|
const docs = await loader.load()
|
||||||
|
alldocs.push(...docs)
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
docs = await loader.load()
|
// @ts-ignore
|
||||||
|
const loader = new PDFLoader(new Blob([bf]), { pdfjs: () => import('pdf-parse/lib/pdf.js/v1.10.100/build/pdf.js') })
|
||||||
|
if (textSplitter) {
|
||||||
|
const docs = await loader.loadAndSplit(textSplitter)
|
||||||
|
alldocs.push(...docs)
|
||||||
|
} else {
|
||||||
|
const docs = await loader.load()
|
||||||
|
alldocs.push(...docs)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (metadata) {
|
if (metadata) {
|
||||||
const parsedMetadata = typeof metadata === 'object' ? metadata : JSON.parse(metadata)
|
const parsedMetadata = typeof metadata === 'object' ? metadata : JSON.parse(metadata)
|
||||||
let finaldocs = []
|
let finaldocs = []
|
||||||
for (const doc of docs) {
|
for (const doc of alldocs) {
|
||||||
const newdoc = {
|
const newdoc = {
|
||||||
...doc,
|
...doc,
|
||||||
metadata: {
|
metadata: {
|
||||||
|
|
@ -102,7 +120,7 @@ class Pdf_DocumentLoaders implements INode {
|
||||||
return finaldocs
|
return finaldocs
|
||||||
}
|
}
|
||||||
|
|
||||||
return docs
|
return alldocs
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -1,7 +1,6 @@
|
||||||
import { INode, INodeData, INodeParams } from '../../../src/Interface'
|
import { INode, INodeData, INodeParams } from '../../../src/Interface'
|
||||||
import { TextSplitter } from 'langchain/text_splitter'
|
import { TextSplitter } from 'langchain/text_splitter'
|
||||||
import { TextLoader } from 'langchain/document_loaders/fs/text'
|
import { TextLoader } from 'langchain/document_loaders/fs/text'
|
||||||
import { getBlob } from '../../../src/utils'
|
|
||||||
|
|
||||||
class Text_DocumentLoaders implements INode {
|
class Text_DocumentLoaders implements INode {
|
||||||
label: string
|
label: string
|
||||||
|
|
@ -49,20 +48,35 @@ class Text_DocumentLoaders implements INode {
|
||||||
const txtFileBase64 = nodeData.inputs?.txtFile as string
|
const txtFileBase64 = nodeData.inputs?.txtFile as string
|
||||||
const metadata = nodeData.inputs?.metadata
|
const metadata = nodeData.inputs?.metadata
|
||||||
|
|
||||||
const blob = new Blob(getBlob(txtFileBase64))
|
let alldocs = []
|
||||||
const loader = new TextLoader(blob)
|
let files: string[] = []
|
||||||
let docs = []
|
|
||||||
|
|
||||||
if (textSplitter) {
|
if (txtFileBase64.startsWith('[') && txtFileBase64.endsWith(']')) {
|
||||||
docs = await loader.loadAndSplit(textSplitter)
|
files = JSON.parse(txtFileBase64)
|
||||||
} else {
|
} else {
|
||||||
docs = await loader.load()
|
files = [txtFileBase64]
|
||||||
|
}
|
||||||
|
|
||||||
|
for (const file of files) {
|
||||||
|
const splitDataURI = file.split(',')
|
||||||
|
splitDataURI.pop()
|
||||||
|
const bf = Buffer.from(splitDataURI.pop() || '', 'base64')
|
||||||
|
const blob = new Blob([bf])
|
||||||
|
const loader = new TextLoader(blob)
|
||||||
|
|
||||||
|
if (textSplitter) {
|
||||||
|
const docs = await loader.loadAndSplit(textSplitter)
|
||||||
|
alldocs.push(...docs)
|
||||||
|
} else {
|
||||||
|
const docs = await loader.load()
|
||||||
|
alldocs.push(...docs)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (metadata) {
|
if (metadata) {
|
||||||
const parsedMetadata = typeof metadata === 'object' ? metadata : JSON.parse(metadata)
|
const parsedMetadata = typeof metadata === 'object' ? metadata : JSON.parse(metadata)
|
||||||
let finaldocs = []
|
let finaldocs = []
|
||||||
for (const doc of docs) {
|
for (const doc of alldocs) {
|
||||||
const newdoc = {
|
const newdoc = {
|
||||||
...doc,
|
...doc,
|
||||||
metadata: {
|
metadata: {
|
||||||
|
|
@ -74,7 +88,7 @@ class Text_DocumentLoaders implements INode {
|
||||||
}
|
}
|
||||||
return finaldocs
|
return finaldocs
|
||||||
}
|
}
|
||||||
return docs
|
return alldocs
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -149,27 +149,3 @@ export const getInputVariables = (paramValue: string): string[] => {
|
||||||
}
|
}
|
||||||
return inputVariables
|
return inputVariables
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Get blob
|
|
||||||
* @param {string} fileBase64Str
|
|
||||||
* @returns {Buffer[]}
|
|
||||||
*/
|
|
||||||
export const getBlob = (fileBase64Str: string) => {
|
|
||||||
let bufferArray: Buffer[] = []
|
|
||||||
let files: string[] = []
|
|
||||||
|
|
||||||
if (fileBase64Str.startsWith('[') && fileBase64Str.endsWith(']')) {
|
|
||||||
files = JSON.parse(fileBase64Str)
|
|
||||||
} else {
|
|
||||||
files = [fileBase64Str]
|
|
||||||
}
|
|
||||||
|
|
||||||
for (const file of files) {
|
|
||||||
const splitDataURI = file.split(',')
|
|
||||||
splitDataURI.pop()
|
|
||||||
const bf = Buffer.from(splitDataURI.pop() || '', 'base64')
|
|
||||||
bufferArray.push(bf)
|
|
||||||
}
|
|
||||||
return bufferArray
|
|
||||||
}
|
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue