import { EventStreamCodec } from "@smithy/eventstream-codec"; import { CallbackManagerForLLMRun } from "@langchain/core/callbacks/manager"; import { GenerationChunk } from "@langchain/core/outputs"; import { LLM, type BaseLLMParams } from "@langchain/core/language_models/llms"; import { BaseBedrockInput, type CredentialType } from "../../utils/bedrock/index.js"; import type { SerializedFields } from "../../load/map_keys.js"; /** * A type of Large Language Model (LLM) that interacts with the Bedrock * service. It extends the base `LLM` class and implements the * `BaseBedrockInput` interface. The class is designed to authenticate and * interact with the Bedrock service, which is a part of Amazon Web * Services (AWS). It uses AWS credentials for authentication and can be * configured with various parameters such as the model to use, the AWS * region, and the maximum number of tokens to generate. */ export declare class Bedrock extends LLM implements BaseBedrockInput { model: string; modelProvider: string; region: string; credentials: CredentialType; temperature?: number | undefined; maxTokens?: number | undefined; fetchFn: typeof fetch; endpointHost?: string; /** @deprecated */ stopSequences?: string[]; modelKwargs?: Record; codec: EventStreamCodec; streaming: boolean; lc_serializable: boolean; get lc_aliases(): Record; get lc_secrets(): { [key: string]: string; } | undefined; get lc_attributes(): SerializedFields | undefined; _llmType(): string; static lc_name(): string; constructor(fields?: Partial & BaseLLMParams); /** Call out to Bedrock service model. Arguments: prompt: The prompt to pass into the model. Returns: The string generated by the model. Example: response = model.invoke("Tell me a joke.") */ _call(prompt: string, options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise; _signedFetch(prompt: string, options: this["ParsedCallOptions"], fields: { bedrockMethod: "invoke" | "invoke-with-response-stream"; endpointHost: string; provider: string; }): Promise; invocationParams(options?: this["ParsedCallOptions"]): { model: string; region: string; temperature: number | undefined; maxTokens: number | undefined; stop: string[] | undefined; modelKwargs: Record | undefined; }; _streamResponseChunks(prompt: string, options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): AsyncGenerator; _readChunks(reader: any): { [Symbol.asyncIterator](): AsyncGenerator; }; }