agsamantha/node_modules/@langchain/community/dist/experimental/hubs/makersuite/llms.d.ts

107 lines
3.6 KiB
TypeScript
Raw Normal View History

2024-10-02 20:15:21 +00:00
import { protos } from "@google-ai/generativelanguage";
import { type BaseLLMParams, LLM } from "@langchain/core/language_models/llms";
/**
* @deprecated - Deprecated by Google. Will be removed in 0.3.0
*
* Input for Text generation for Google Palm
*/
export interface GooglePaLMTextInput extends BaseLLMParams {
/**
* Model Name to use
*
* Alias for `model`
*
* Note: The format must follow the pattern - `models/{model}`
*/
modelName?: string;
/**
* Model Name to use
*
* Note: The format must follow the pattern - `models/{model}`
*/
model?: string;
/**
* Controls the randomness of the output.
*
* Values can range from [0.0,1.0], inclusive. A value closer to 1.0
* will produce responses that are more varied and creative, while
* a value closer to 0.0 will typically result in more straightforward
* responses from the model.
*
* Note: The default value varies by model
*/
temperature?: number;
/**
* Maximum number of tokens to generate in the completion.
*/
maxOutputTokens?: number;
/**
* Top-p changes how the model selects tokens for output.
*
* Tokens are selected from most probable to least until the sum
* of their probabilities equals the top-p value.
*
* For example, if tokens A, B, and C have a probability of
* .3, .2, and .1 and the top-p value is .5, then the model will
* select either A or B as the next token (using temperature).
*
* Note: The default value varies by model
*/
topP?: number;
/**
* Top-k changes how the model selects tokens for output.
*
* A top-k of 1 means the selected token is the most probable among
* all tokens in the models vocabulary (also called greedy decoding),
* while a top-k of 3 means that the next token is selected from
* among the 3 most probable tokens (using temperature).
*
* Note: The default value varies by model
*/
topK?: number;
/**
* The set of character sequences (up to 5) that will stop output generation.
* If specified, the API will stop at the first appearance of a stop
* sequence.
*
* Note: The stop sequence will not be included as part of the response.
*/
stopSequences?: string[];
/**
* A list of unique `SafetySetting` instances for blocking unsafe content. The API will block
* any prompts and responses that fail to meet the thresholds set by these settings. If there
* is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use
* the default safety setting for that category.
*/
safetySettings?: protos.google.ai.generativelanguage.v1beta2.ISafetySetting[];
/**
* Google Palm API key to use
*/
apiKey?: string;
}
/**
* @deprecated - Deprecated by Google. Will be removed in 0.3.0
*
* Google Palm 2 Language Model Wrapper to generate texts
*/
export declare class GooglePaLM extends LLM implements GooglePaLMTextInput {
lc_serializable: boolean;
get lc_secrets(): {
[key: string]: string;
} | undefined;
modelName: string;
model: string;
temperature?: number;
maxOutputTokens?: number;
topP?: number;
topK?: number;
stopSequences: string[];
safetySettings?: protos.google.ai.generativelanguage.v1beta2.ISafetySetting[];
apiKey?: string;
private client;
constructor(fields?: GooglePaLMTextInput);
_llmType(): string;
_call(prompt: string, options: this["ParsedCallOptions"]): Promise<string>;
protected _generateText(prompt: string): Promise<string | null | undefined>;
}