67 lines
2.8 KiB
TypeScript
67 lines
2.8 KiB
TypeScript
import { EventStreamCodec } from "@smithy/eventstream-codec";
|
|
import { CallbackManagerForLLMRun } from "@langchain/core/callbacks/manager";
|
|
import { GenerationChunk } from "@langchain/core/outputs";
|
|
import { LLM, type BaseLLMParams } from "@langchain/core/language_models/llms";
|
|
import { BaseBedrockInput, type CredentialType } from "../../utils/bedrock/index.js";
|
|
import type { SerializedFields } from "../../load/map_keys.js";
|
|
/**
|
|
* A type of Large Language Model (LLM) that interacts with the Bedrock
|
|
* service. It extends the base `LLM` class and implements the
|
|
* `BaseBedrockInput` interface. The class is designed to authenticate and
|
|
* interact with the Bedrock service, which is a part of Amazon Web
|
|
* Services (AWS). It uses AWS credentials for authentication and can be
|
|
* configured with various parameters such as the model to use, the AWS
|
|
* region, and the maximum number of tokens to generate.
|
|
*/
|
|
export declare class Bedrock extends LLM implements BaseBedrockInput {
|
|
model: string;
|
|
modelProvider: string;
|
|
region: string;
|
|
credentials: CredentialType;
|
|
temperature?: number | undefined;
|
|
maxTokens?: number | undefined;
|
|
fetchFn: typeof fetch;
|
|
endpointHost?: string;
|
|
/** @deprecated */
|
|
stopSequences?: string[];
|
|
modelKwargs?: Record<string, unknown>;
|
|
codec: EventStreamCodec;
|
|
streaming: boolean;
|
|
lc_serializable: boolean;
|
|
get lc_aliases(): Record<string, string>;
|
|
get lc_secrets(): {
|
|
[key: string]: string;
|
|
} | undefined;
|
|
get lc_attributes(): SerializedFields | undefined;
|
|
_llmType(): string;
|
|
static lc_name(): string;
|
|
constructor(fields?: Partial<BaseBedrockInput> & BaseLLMParams);
|
|
/** Call out to Bedrock service model.
|
|
Arguments:
|
|
prompt: The prompt to pass into the model.
|
|
|
|
Returns:
|
|
The string generated by the model.
|
|
|
|
Example:
|
|
response = model.invoke("Tell me a joke.")
|
|
*/
|
|
_call(prompt: string, options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise<string>;
|
|
_signedFetch(prompt: string, options: this["ParsedCallOptions"], fields: {
|
|
bedrockMethod: "invoke" | "invoke-with-response-stream";
|
|
endpointHost: string;
|
|
provider: string;
|
|
}): Promise<Response>;
|
|
invocationParams(options?: this["ParsedCallOptions"]): {
|
|
model: string;
|
|
region: string;
|
|
temperature: number | undefined;
|
|
maxTokens: number | undefined;
|
|
stop: string[] | undefined;
|
|
modelKwargs: Record<string, unknown> | undefined;
|
|
};
|
|
_streamResponseChunks(prompt: string, options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): AsyncGenerator<GenerationChunk>;
|
|
_readChunks(reader: any): {
|
|
[Symbol.asyncIterator](): AsyncGenerator<Uint8Array, void, unknown>;
|
|
};
|
|
}
|