import { LLMChain } from "../llm_chain.js"; import { StuffDocumentsChain, MapReduceDocumentsChain, RefineDocumentsChain, } from "../combine_docs_chain.js"; import { QA_PROMPT_SELECTOR } from "./stuff_prompts.js"; import { COMBINE_PROMPT_SELECTOR, COMBINE_QA_PROMPT_SELECTOR, } from "./map_reduce_prompts.js"; import { QUESTION_PROMPT_SELECTOR, REFINE_PROMPT_SELECTOR, } from "./refine_prompts.js"; export const loadQAChain = (llm, params = { type: "stuff" }) => { const { type } = params; if (type === "stuff") { return loadQAStuffChain(llm, params); } if (type === "map_reduce") { return loadQAMapReduceChain(llm, params); } if (type === "refine") { return loadQARefineChain(llm, params); } throw new Error(`Invalid _type: ${type}`); }; /** * Loads a StuffQAChain based on the provided parameters. It takes an LLM * instance and StuffQAChainParams as parameters. * @param llm An instance of BaseLanguageModel. * @param params Parameters for creating a StuffQAChain. * @returns A StuffQAChain instance. */ export function loadQAStuffChain(llm, params = {}) { const { prompt = QA_PROMPT_SELECTOR.getPrompt(llm), verbose } = params; const llmChain = new LLMChain({ prompt, llm, verbose }); const chain = new StuffDocumentsChain({ llmChain, verbose }); return chain; } /** * Loads a MapReduceQAChain based on the provided parameters. It takes an * LLM instance and MapReduceQAChainParams as parameters. * @param llm An instance of BaseLanguageModel. * @param params Parameters for creating a MapReduceQAChain. * @returns A MapReduceQAChain instance. */ export function loadQAMapReduceChain(llm, params = {}) { const { combineMapPrompt = COMBINE_QA_PROMPT_SELECTOR.getPrompt(llm), combinePrompt = COMBINE_PROMPT_SELECTOR.getPrompt(llm), verbose, combineLLM, returnIntermediateSteps, } = params; const llmChain = new LLMChain({ prompt: combineMapPrompt, llm, verbose }); const combineLLMChain = new LLMChain({ prompt: combinePrompt, llm: combineLLM ?? llm, verbose, }); const combineDocumentChain = new StuffDocumentsChain({ llmChain: combineLLMChain, documentVariableName: "summaries", verbose, }); const chain = new MapReduceDocumentsChain({ llmChain, combineDocumentChain, returnIntermediateSteps, verbose, }); return chain; } /** * Loads a RefineQAChain based on the provided parameters. It takes an LLM * instance and RefineQAChainParams as parameters. * @param llm An instance of BaseLanguageModel. * @param params Parameters for creating a RefineQAChain. * @returns A RefineQAChain instance. */ export function loadQARefineChain(llm, params = {}) { const { questionPrompt = QUESTION_PROMPT_SELECTOR.getPrompt(llm), refinePrompt = REFINE_PROMPT_SELECTOR.getPrompt(llm), refineLLM, verbose, } = params; const llmChain = new LLMChain({ prompt: questionPrompt, llm, verbose }); const refineLLMChain = new LLMChain({ prompt: refinePrompt, llm: refineLLM ?? llm, verbose, }); const chain = new RefineDocumentsChain({ llmChain, refineLLMChain, verbose, }); return chain; }