"use strict"; var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) { if (k2 === undefined) k2 = k; var desc = Object.getOwnPropertyDescriptor(m, k); if (!desc || ("get" in desc ? !m.__esModule : desc.writable || desc.configurable)) { desc = { enumerable: true, get: function() { return m[k]; } }; } Object.defineProperty(o, k2, desc); }) : (function(o, m, k, k2) { if (k2 === undefined) k2 = k; o[k2] = m[k]; })); var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) { Object.defineProperty(o, "default", { enumerable: true, value: v }); }) : function(o, v) { o["default"] = v; }); var __importStar = (this && this.__importStar) || function (mod) { if (mod && mod.__esModule) return mod; var result = {}; if (mod != null) for (var k in mod) if (k !== "default" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k); __setModuleDefault(result, mod); return result; }; var __importDefault = (this && this.__importDefault) || function (mod) { return (mod && mod.__esModule) ? mod : { "default": mod }; }; Object.defineProperty(exports, "__esModule", { value: true }); exports.AnalyticDBVectorStore = void 0; const uuid = __importStar(require("uuid")); const pg_1 = __importDefault(require("pg")); const pg_copy_streams_1 = require("pg-copy-streams"); const promises_1 = require("node:stream/promises"); const node_stream_1 = require("node:stream"); const vectorstores_1 = require("@langchain/core/vectorstores"); const documents_1 = require("@langchain/core/documents"); const _LANGCHAIN_DEFAULT_COLLECTION_NAME = "langchain_document"; /** * Class that provides methods for creating and managing a collection of * documents in an AnalyticDB, adding documents or vectors to the * collection, performing similarity search on vectors, and creating an * instance of `AnalyticDBVectorStore` from texts or documents. */ class AnalyticDBVectorStore extends vectorstores_1.VectorStore { _vectorstoreType() { return "analyticdb"; } constructor(embeddings, args) { super(embeddings, args); Object.defineProperty(this, "pool", { enumerable: true, configurable: true, writable: true, value: void 0 }); Object.defineProperty(this, "embeddingDimension", { enumerable: true, configurable: true, writable: true, value: void 0 }); Object.defineProperty(this, "collectionName", { enumerable: true, configurable: true, writable: true, value: void 0 }); Object.defineProperty(this, "preDeleteCollection", { enumerable: true, configurable: true, writable: true, value: void 0 }); Object.defineProperty(this, "isCreateCollection", { enumerable: true, configurable: true, writable: true, value: false }); this.pool = new pg_1.default.Pool({ host: args.connectionOptions.host, port: args.connectionOptions.port, database: args.connectionOptions.database, user: args.connectionOptions.user, password: args.connectionOptions.password, }); this.embeddingDimension = args.embeddingDimension; this.collectionName = args.collectionName || _LANGCHAIN_DEFAULT_COLLECTION_NAME; this.preDeleteCollection = args.preDeleteCollection || false; } /** * Closes all the clients in the pool and terminates the pool. * @returns Promise that resolves when all clients are closed and the pool is terminated. */ async end() { return this.pool.end(); } /** * Creates a new table in the database if it does not already exist. The * table is created with columns for id, embedding, document, and * metadata. An index is also created on the embedding column if it does * not already exist. * @returns Promise that resolves when the table and index are created. */ async createTableIfNotExists() { if (!this.embeddingDimension) { this.embeddingDimension = (await this.embeddings.embedQuery("test")).length; } const client = await this.pool.connect(); try { await client.query("BEGIN"); // Create the table if it doesn't exist await client.query(` CREATE TABLE IF NOT EXISTS ${this.collectionName} ( id TEXT PRIMARY KEY DEFAULT NULL, embedding REAL[], document TEXT, metadata JSON ); `); // Check if the index exists const indexName = `${this.collectionName}_embedding_idx`; const indexQuery = ` SELECT 1 FROM pg_indexes WHERE indexname = '${indexName}'; `; const result = await client.query(indexQuery); // Create the index if it doesn't exist if (result.rowCount === 0) { const indexStatement = ` CREATE INDEX ${indexName} ON ${this.collectionName} USING ann(embedding) WITH ( "dim" = ${this.embeddingDimension}, "hnsw_m" = 100 ); `; await client.query(indexStatement); } await client.query("COMMIT"); } catch (err) { await client.query("ROLLBACK"); throw err; } finally { client.release(); } } /** * Deletes the collection from the database if it exists. * @returns Promise that resolves when the collection is deleted. */ async deleteCollection() { const dropStatement = `DROP TABLE IF EXISTS ${this.collectionName};`; await this.pool.query(dropStatement); } /** * Creates a new collection in the database. If `preDeleteCollection` is * true, any existing collection with the same name is deleted before the * new collection is created. * @returns Promise that resolves when the collection is created. */ async createCollection() { if (this.preDeleteCollection) { await this.deleteCollection(); } await this.createTableIfNotExists(); this.isCreateCollection = true; } /** * Adds an array of documents to the collection. The documents are first * converted to vectors using the `embedDocuments` method of the * `embeddings` instance. * @param documents Array of Document instances to be added to the collection. * @returns Promise that resolves when the documents are added. */ async addDocuments(documents) { // When the pageContent is empty in certain scenarios (such as when using unstructuredIo), an error occurs during embedding. const filteredDocs = documents.filter((doc) => doc.pageContent); if (filteredDocs.length !== documents.length) { console.warn(`[AnalyticDB]: Filtered out ${documents.length - filteredDocs.length} empty documents.`); } const texts = filteredDocs.map(({ pageContent }) => pageContent); return this.addVectors(await this.embeddings.embedDocuments(texts), filteredDocs); } /** * Adds an array of vectors and corresponding documents to the collection. * The vectors and documents are batch inserted into the database. * @param vectors Array of vectors to be added to the collection. * @param documents Array of Document instances corresponding to the vectors. * @returns Promise that resolves when the vectors and documents are added. */ async addVectors(vectors, documents) { if (vectors.length === 0) { return; } if (vectors.length !== documents.length) { throw new Error(`Vectors and documents must have the same length`); } if (!this.embeddingDimension) { this.embeddingDimension = (await this.embeddings.embedQuery("test")).length; } if (vectors[0].length !== this.embeddingDimension) { throw new Error(`Vectors must have the same length as the number of dimensions (${this.embeddingDimension})`); } if (!this.isCreateCollection) { await this.createCollection(); } const client = await this.pool.connect(); try { const chunkSize = 500; const chunksTableData = []; for (let i = 0; i < documents.length; i += 1) { chunksTableData.push({ id: uuid.v4(), embedding: vectors[i], document: documents[i].pageContent, metadata: documents[i].metadata, }); // Execute the batch insert when the batch size is reached if (chunksTableData.length === chunkSize) { const rs = new node_stream_1.Readable(); let currentIndex = 0; rs._read = function () { if (currentIndex === chunkSize) { rs.push(null); } else { const data = chunksTableData[currentIndex]; rs.push(`${data.id}\t{${data.embedding.join(",")}}\t${data.document}\t${JSON.stringify(data.metadata)}\n`); currentIndex += 1; } }; const ws = client.query((0, pg_copy_streams_1.from)(`COPY ${this.collectionName}(id, embedding, document, metadata) FROM STDIN`)); await (0, promises_1.pipeline)(rs, ws); // Clear the chunksTableData list for the next batch chunksTableData.length = 0; } } // Insert any remaining records that didn't make up a full batch if (chunksTableData.length > 0) { const rs = new node_stream_1.Readable(); let currentIndex = 0; rs._read = function () { if (currentIndex === chunksTableData.length) { rs.push(null); } else { const data = chunksTableData[currentIndex]; rs.push(`${data.id}\t{${data.embedding.join(",")}}\t${data.document}\t${JSON.stringify(data.metadata)}\n`); currentIndex += 1; } }; const ws = client.query((0, pg_copy_streams_1.from)(`COPY ${this.collectionName}(id, embedding, document, metadata) FROM STDIN`)); await (0, promises_1.pipeline)(rs, ws); } } finally { client.release(); } } /** * Performs a similarity search on the vectors in the collection. The * search is performed using the given query vector and returns the top k * most similar vectors along with their corresponding documents and * similarity scores. * @param query Query vector for the similarity search. * @param k Number of top similar vectors to return. * @param filter Optional. Filter to apply on the metadata of the documents. * @returns Promise that resolves to an array of tuples, each containing a Document instance and its similarity score. */ async similaritySearchVectorWithScore(query, k, filter) { if (!this.isCreateCollection) { await this.createCollection(); } let filterCondition = ""; const filterEntries = filter ? Object.entries(filter) : []; if (filterEntries.length > 0) { const conditions = filterEntries.map((_, index) => `metadata->>$${2 * index + 3} = $${2 * index + 4}`); filterCondition = `WHERE ${conditions.join(" AND ")}`; } const sqlQuery = ` SELECT *, l2_distance(embedding, $1::real[]) AS distance FROM ${this.collectionName} ${filterCondition} ORDER BY embedding <-> $1 LIMIT $2; `; // Execute the query and fetch the results const { rows } = await this.pool.query(sqlQuery, [ query, k, ...filterEntries.flatMap(([key, value]) => [key, value]), ]); const result = rows.map((row) => [ new documents_1.Document({ pageContent: row.document, metadata: row.metadata }), row.distance, ]); return result; } /** * Creates an instance of `AnalyticDBVectorStore` from an array of texts * and corresponding metadata. The texts are first converted to Document * instances before being added to the collection. * @param texts Array of texts to be added to the collection. * @param metadatas Array or object of metadata corresponding to the texts. * @param embeddings Embeddings instance used to convert the texts to vectors. * @param dbConfig Configuration for the AnalyticDB. * @returns Promise that resolves to an instance of `AnalyticDBVectorStore`. */ static async fromTexts(texts, metadatas, embeddings, dbConfig) { const docs = []; for (let i = 0; i < texts.length; i += 1) { const metadata = Array.isArray(metadatas) ? metadatas[i] : metadatas; const newDoc = new documents_1.Document({ pageContent: texts[i], metadata, }); docs.push(newDoc); } return AnalyticDBVectorStore.fromDocuments(docs, embeddings, dbConfig); } /** * Creates an instance of `AnalyticDBVectorStore` from an array of * Document instances. The documents are added to the collection. * @param docs Array of Document instances to be added to the collection. * @param embeddings Embeddings instance used to convert the documents to vectors. * @param dbConfig Configuration for the AnalyticDB. * @returns Promise that resolves to an instance of `AnalyticDBVectorStore`. */ static async fromDocuments(docs, embeddings, dbConfig) { const instance = new this(embeddings, dbConfig); await instance.addDocuments(docs); return instance; } /** * Creates an instance of `AnalyticDBVectorStore` from an existing index * in the database. A new collection is created in the database. * @param embeddings Embeddings instance used to convert the documents to vectors. * @param dbConfig Configuration for the AnalyticDB. * @returns Promise that resolves to an instance of `AnalyticDBVectorStore`. */ static async fromExistingIndex(embeddings, dbConfig) { const instance = new this(embeddings, dbConfig); await instance.createCollection(); return instance; } } exports.AnalyticDBVectorStore = AnalyticDBVectorStore;