///
import type { Client } from '@modelcontextprotocol/sdk/client/index.js';
import { GoogleAuthOptions } from 'google-auth-library';
import { ReadableStream as ReadableStream_2 } from 'stream/web';
/** Marks the end of user activity.
This can only be sent if automatic (i.e. server-side) activity detection is
disabled. */
export declare interface ActivityEnd {
}
/** The different ways of handling user activity. */
export declare enum ActivityHandling {
/**
* If unspecified, the default behavior is `START_OF_ACTIVITY_INTERRUPTS`.
*/
ACTIVITY_HANDLING_UNSPECIFIED = "ACTIVITY_HANDLING_UNSPECIFIED",
/**
* If true, start of activity will interrupt the model's response (also called "barge in"). The model's current response will be cut-off in the moment of the interruption. This is the default behavior.
*/
START_OF_ACTIVITY_INTERRUPTS = "START_OF_ACTIVITY_INTERRUPTS",
/**
* The model's response will not be interrupted.
*/
NO_INTERRUPTION = "NO_INTERRUPTION"
}
/** Marks the start of user activity.
This can only be sent if automatic (i.e. server-side) activity detection is
disabled. */
export declare interface ActivityStart {
}
/** Adapter size for tuning. This enum is not supported in Gemini API. */
export declare enum AdapterSize {
/**
* Adapter size is unspecified.
*/
ADAPTER_SIZE_UNSPECIFIED = "ADAPTER_SIZE_UNSPECIFIED",
/**
* Adapter size 1.
*/
ADAPTER_SIZE_ONE = "ADAPTER_SIZE_ONE",
/**
* Adapter size 2.
*/
ADAPTER_SIZE_TWO = "ADAPTER_SIZE_TWO",
/**
* Adapter size 4.
*/
ADAPTER_SIZE_FOUR = "ADAPTER_SIZE_FOUR",
/**
* Adapter size 8.
*/
ADAPTER_SIZE_EIGHT = "ADAPTER_SIZE_EIGHT",
/**
* Adapter size 16.
*/
ADAPTER_SIZE_SIXTEEN = "ADAPTER_SIZE_SIXTEEN",
/**
* Adapter size 32.
*/
ADAPTER_SIZE_THIRTY_TWO = "ADAPTER_SIZE_THIRTY_TWO"
}
/** Aggregation metric. This enum is not supported in Gemini API. */
export declare enum AggregationMetric {
/**
* Unspecified aggregation metric.
*/
AGGREGATION_METRIC_UNSPECIFIED = "AGGREGATION_METRIC_UNSPECIFIED",
/**
* Average aggregation metric. Not supported for Pairwise metric.
*/
AVERAGE = "AVERAGE",
/**
* Mode aggregation metric.
*/
MODE = "MODE",
/**
* Standard deviation aggregation metric. Not supported for pairwise metric.
*/
STANDARD_DEVIATION = "STANDARD_DEVIATION",
/**
* Variance aggregation metric. Not supported for pairwise metric.
*/
VARIANCE = "VARIANCE",
/**
* Minimum aggregation metric. Not supported for pairwise metric.
*/
MINIMUM = "MINIMUM",
/**
* Maximum aggregation metric. Not supported for pairwise metric.
*/
MAXIMUM = "MAXIMUM",
/**
* Median aggregation metric. Not supported for pairwise metric.
*/
MEDIAN = "MEDIAN",
/**
* 90th percentile aggregation metric. Not supported for pairwise metric.
*/
PERCENTILE_P90 = "PERCENTILE_P90",
/**
* 95th percentile aggregation metric. Not supported for pairwise metric.
*/
PERCENTILE_P95 = "PERCENTILE_P95",
/**
* 99th percentile aggregation metric. Not supported for pairwise metric.
*/
PERCENTILE_P99 = "PERCENTILE_P99"
}
/** The aggregation result for the entire dataset and all metrics. This data type is not supported in Gemini API. */
export declare interface AggregationOutput {
/** One AggregationResult per metric. */
aggregationResults?: AggregationResult[];
/** The dataset used for evaluation & aggregation. */
dataset?: EvaluationDataset;
}
/** The aggregation result for a single metric. This data type is not supported in Gemini API. */
export declare interface AggregationResult {
/** Aggregation metric. */
aggregationMetric?: AggregationMetric;
/** Results for bleu metric. */
bleuMetricValue?: BleuMetricValue;
/** Result for code execution metric. */
customCodeExecutionResult?: CustomCodeExecutionResult;
/** Results for exact match metric. */
exactMatchMetricValue?: ExactMatchMetricValue;
/** Result for pairwise metric. */
pairwiseMetricResult?: PairwiseMetricResult;
/** Result for pointwise metric. */
pointwiseMetricResult?: PointwiseMetricResult;
/** Results for rouge metric. */
rougeMetricValue?: RougeMetricValue;
}
/**
* The configuration for allowed tools.
*/
declare interface AllowedTools {
/**
* The mode of the tool choice.
*/
mode?: ToolChoiceType;
/**
* The names of the allowed tools.
*/
tools?: Array;
}
/**
* Citation information for model-generated content.
*/
declare interface Annotation {
/**
* End of the attributed segment, exclusive.
*/
end_index?: number;
/**
* Source attributed for a portion of the text. Could be a URL, title, or
* other identifier.
*/
source?: string;
/**
* Start of segment of the response that is attributed to this source.
*
* Index indicates the start of the segment, measured in bytes.
*/
start_index?: number;
}
/** The generic reusable api auth config. Deprecated. Please use AuthConfig (google/cloud/aiplatform/master/auth.proto) instead. This data type is not supported in Gemini API. */
export declare interface ApiAuth {
/** The API secret. */
apiKeyConfig?: ApiAuthApiKeyConfig;
}
/** The API secret. This data type is not supported in Gemini API. */
export declare interface ApiAuthApiKeyConfig {
/** Required. The SecretManager secret version resource name storing API key. e.g. projects/{project}/secrets/{secret}/versions/{version} */
apiKeySecretVersion?: string;
/** The API key string. Either this or `api_key_secret_version` must be set. */
apiKeyString?: string;
}
/**
* The ApiClient class is used to send requests to the Gemini API or Vertex AI
* endpoints.
*
* WARNING: This is an internal API and may change without notice. Direct usage
* is not supported and may break your application.
*/
declare class ApiClient implements GeminiNextGenAPIClientAdapter {
readonly clientOptions: ApiClientInitOptions;
private readonly customBaseUrl?;
constructor(opts: ApiClientInitOptions);
isVertexAI(): boolean;
getProject(): string | undefined;
getLocation(): string | undefined;
getCustomBaseUrl(): string | undefined;
getAuthHeaders(): Promise;
getApiVersion(): string;
getBaseUrl(): string;
getRequestUrl(): string;
getHeaders(): Record;
private getRequestUrlInternal;
getBaseResourcePath(): string;
getApiKey(): string | undefined;
getWebsocketBaseUrl(): string;
setBaseUrl(url: string): void;
private constructUrl;
private shouldPrependVertexProjectPath;
request(request: HttpRequest): Promise;
private patchHttpOptions;
requestStream(request: HttpRequest): Promise>;
private includeExtraHttpOptionsToRequestInit;
private unaryApiCall;
private streamApiCall;
processStreamResponse(response: Response): AsyncGenerator;
private apiCall;
getDefaultHeaders(): Record;
private getHeadersInternal;
private getFileName;
/**
* Uploads a file asynchronously using Gemini API only, this is not supported
* in Vertex AI.
*
* @param file The string path to the file to be uploaded or a Blob object.
* @param config Optional parameters specified in the `UploadFileConfig`
* interface. @see {@link types.UploadFileConfig}
* @return A promise that resolves to a `File` object.
* @throws An error if called on a Vertex AI client.
* @throws An error if the `mimeType` is not provided and can not be inferred,
*/
uploadFile(file: string | Blob, config?: types.UploadFileConfig): Promise;
/**
* Uploads a file to a given file search store asynchronously using Gemini API only, this is not supported
* in Vertex AI.
*
* @param fileSearchStoreName The name of the file search store to upload the file to.
* @param file The string path to the file to be uploaded or a Blob object.
* @param config Optional parameters specified in the `UploadFileConfig`
* interface. @see {@link UploadFileConfig}
* @return A promise that resolves to a `File` object.
* @throws An error if called on a Vertex AI client.
* @throws An error if the `mimeType` is not provided and can not be inferred,
*/
uploadFileToFileSearchStore(fileSearchStoreName: string, file: string | Blob, config?: types.UploadToFileSearchStoreConfig): Promise;
/**
* Downloads a file asynchronously to the specified path.
*
* @params params - The parameters for the download request, see {@link
* types.DownloadFileParameters}
*/
downloadFile(params: types.DownloadFileParameters): Promise;
private fetchUploadUrl;
}
/**
* Options for initializing the ApiClient. The ApiClient uses the parameters
* for authentication purposes as well as to infer if SDK should send the
* request to Vertex AI or Gemini API.
*/
declare interface ApiClientInitOptions {
/**
* The object used for adding authentication headers to API requests.
*/
auth: Auth;
/**
* The uploader to use for uploading files. This field is required for
* creating a client, will be set through the Node_client or Web_client.
*/
uploader: Uploader;
/**
* Optional. The downloader to use for downloading files. This field is
* required for creating a client, will be set through the Node_client or
* Web_client.
*/
downloader: Downloader;
/**
* Optional. The Google Cloud project ID for Vertex AI users.
* It is not the numeric project name.
* If not provided, SDK will try to resolve it from runtime environment.
*/
project?: string;
/**
* Optional. The Google Cloud project location for Vertex AI users.
* If not provided, SDK will try to resolve it from runtime environment.
*/
location?: string;
/**
* The API Key. This is required for Gemini API users.
*/
apiKey?: string;
/**
* Optional. Set to true if you intend to call Vertex AI endpoints.
* If unset, default SDK behavior is to call Gemini API.
*/
vertexai?: boolean;
/**
* Optional. The API version for the endpoint.
* If unset, SDK will choose a default api version.
*/
apiVersion?: string;
/**
* Optional. A set of customizable configuration for HTTP requests.
*/
httpOptions?: types.HttpOptions;
/**
* Optional. An extra string to append at the end of the User-Agent header.
*
* This can be used to e.g specify the runtime and its version.
*/
userAgentExtra?: string;
}
declare class APIConnectionError extends APIError {
constructor({ message, cause }: {
message?: string | undefined;
cause?: Error | undefined;
});
}
declare class APIConnectionTimeoutError extends APIConnectionError {
constructor({ message }?: {
message?: string;
});
}
declare class APIError extends GeminiNextGenAPIClientError {
/** HTTP status for the response that caused the error */
readonly status: TStatus;
/** HTTP headers for the response that caused the error */
readonly headers: THeaders;
/** JSON body of the response that caused the error */
readonly error: TError;
constructor(status: TStatus, error: TError, message: string | undefined, headers: THeaders);
private static makeMessage;
static generate(status: number | undefined, errorResponse: Object | undefined, message: string | undefined, headers: Headers | undefined): APIError;
}
/**
* API errors raised by the GenAI API.
*/
export declare class ApiError extends Error {
/** HTTP status code */
status: number;
constructor(options: ApiErrorInfo);
}
/**
* @license
* Copyright 2025 Google LLC
* SPDX-License-Identifier: Apache-2.0
*/
/**
* Details for errors from calling the API.
*/
export declare interface ApiErrorInfo {
/** The error message. */
message: string;
/** The HTTP status code. */
status: number;
}
/** Config for authentication with API key. This data type is not supported in Gemini API. */
export declare interface ApiKeyConfig {
/** Optional. The name of the SecretManager secret version resource storing the API key. Format: `projects/{project}/secrets/{secrete}/versions/{version}` - If both `api_key_secret` and `api_key_string` are specified, this field takes precedence over `api_key_string`. - If specified, the `secretmanager.versions.access` permission should be granted to Vertex AI Extension Service Agent (https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents) on the specified resource. */
apiKeySecret?: string;
/** Optional. The API key to be used in the request directly. */
apiKeyString?: string;
/** Optional. The location of the API key. */
httpElementLocation?: HttpElementLocation;
/** Optional. The parameter name of the API key. E.g. If the API request is "https://example.com/act?api_key=", "api_key" would be the parameter name. */
name?: string;
}
/**
* A subclass of `Promise` providing additional helper methods
* for interacting with the SDK.
*/
declare class APIPromise extends Promise {
private responsePromise;
private parseResponse;
private parsedPromise;
private client;
constructor(client: BaseGeminiNextGenAPIClient, responsePromise: Promise, parseResponse?: (client: BaseGeminiNextGenAPIClient, props: APIResponseProps) => PromiseOrValue);
_thenUnwrap(transform: (data: T, props: APIResponseProps) => U): APIPromise;
/**
* Gets the raw `Response` instance instead of parsing the response
* data.
*
* If you want to parse the response body but still get the `Response`
* instance, you can use {@link withResponse()}.
*
* 👋 Getting the wrong TypeScript type for `Response`?
* Try setting `"moduleResolution": "NodeNext"` or add `"lib": ["DOM"]`
* to your `tsconfig.json`.
*/
asResponse(): Promise;
/**
* Gets the parsed response data and the raw `Response` instance.
*
* If you just want to get the raw `Response` instance without parsing it,
* you can use {@link asResponse()}.
*
* 👋 Getting the wrong TypeScript type for `Response`?
* Try setting `"moduleResolution": "NodeNext"` or add `"lib": ["DOM"]`
* to your `tsconfig.json`.
*/
withResponse(): Promise<{
data: T;
response: Response;
}>;
private parse;
then(onfulfilled?: ((value: T) => TResult1 | PromiseLike) | undefined | null, onrejected?: ((reason: any) => TResult2 | PromiseLike) | undefined | null): Promise;
catch(onrejected?: ((reason: any) => TResult | PromiseLike) | undefined | null): Promise;
finally(onfinally?: (() => void) | undefined | null): Promise;
}
declare abstract class APIResource {
/**
* The key path from the client. For example, a resource accessible as `client.resource.subresource` would
* have a property `static override readonly _key = Object.freeze(['resource', 'subresource'] as const);`.
*/
static readonly _key: readonly string[];
protected _client: BaseGeminiNextGenAPIClient;
constructor(client: BaseGeminiNextGenAPIClient);
}
declare type APIResponseProps = {
response: Response;
options: FinalRequestOptions;
controller: AbortController;
requestLogID: string;
retryOfRequestLogID: string | undefined;
startTime: number;
};
/** The API spec that the external API implements. This enum is not supported in Gemini API. */
export declare enum ApiSpec {
/**
* Unspecified API spec. This value should not be used.
*/
API_SPEC_UNSPECIFIED = "API_SPEC_UNSPECIFIED",
/**
* Simple search API spec.
*/
SIMPLE_SEARCH = "SIMPLE_SEARCH",
/**
* Elastic search API spec.
*/
ELASTIC_SEARCH = "ELASTIC_SEARCH"
}
declare class APIUserAbortError extends APIError {
constructor({ message }?: {
message?: string;
});
}
/** Representation of an audio chunk. */
export declare interface AudioChunk {
/** Raw bytes of audio data.
* @remarks Encoded as base64 string. */
data?: string;
/** MIME type of the audio chunk. */
mimeType?: string;
/** Prompts and config used for generating this audio chunk. */
sourceMetadata?: LiveMusicSourceMetadata;
}
/**
* An audio content block.
*/
declare interface AudioContent {
type: 'audio';
/**
* The audio content.
*/
data?: string;
/**
* The mime type of the audio.
*/
mime_type?: 'audio/wav' | 'audio/mp3' | 'audio/aiff' | 'audio/aac' | 'audio/ogg' | 'audio/flac';
/**
* The URI of the audio.
*/
uri?: string;
}
/** The audio transcription configuration in Setup. */
export declare interface AudioTranscriptionConfig {
/** The language codes of the audio. BCP-47 language code. If not set, the transcription will be in the language detected by the model. If set, the server will use the language code specified in the model config as a hint for the language of the audio
*/
languageCodes?: string[];
}
/**
* @license
* Copyright 2025 Google LLC
* SPDX-License-Identifier: Apache-2.0
*/
/**
* The Auth interface is used to authenticate with the API service.
*/
declare interface Auth {
/**
* Sets the headers needed to authenticate with the API service.
*
* @param headers - The Headers object that will be updated with the authentication headers.
* @param url - The URL of the request.
*/
addAuthHeaders(headers: Headers, url?: string): Promise;
}
/** The authentication config to access the API. */
export declare interface AuthConfig {
/** The authentication config to access the API. Only API key is supported. This field is not supported in Gemini API. */
apiKey?: string;
/** Config for API key auth. */
apiKeyConfig?: ApiKeyConfig;
/** Type of auth scheme. */
authType?: AuthType;
/** Config for Google Service Account auth. */
googleServiceAccountConfig?: AuthConfigGoogleServiceAccountConfig;
/** Config for HTTP Basic auth. */
httpBasicAuthConfig?: AuthConfigHttpBasicAuthConfig;
/** Config for user oauth. */
oauthConfig?: AuthConfigOauthConfig;
/** Config for user OIDC auth. */
oidcConfig?: AuthConfigOidcConfig;
}
/** Config for Google Service Account Authentication. This data type is not supported in Gemini API. */
export declare interface AuthConfigGoogleServiceAccountConfig {
/** Optional. The service account that the extension execution service runs as. - If the service account is specified, the `iam.serviceAccounts.getAccessToken` permission should be granted to Vertex AI Extension Service Agent (https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents) on the specified service account. - If not specified, the Vertex AI Extension Service Agent will be used to execute the Extension. */
serviceAccount?: string;
}
/** Config for HTTP Basic Authentication. This data type is not supported in Gemini API. */
export declare interface AuthConfigHttpBasicAuthConfig {
/** Required. The name of the SecretManager secret version resource storing the base64 encoded credentials. Format: `projects/{project}/secrets/{secrete}/versions/{version}` - If specified, the `secretmanager.versions.access` permission should be granted to Vertex AI Extension Service Agent (https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents) on the specified resource. */
credentialSecret?: string;
}
/** Config for user oauth. This data type is not supported in Gemini API. */
export declare interface AuthConfigOauthConfig {
/** Access token for extension endpoint. Only used to propagate token from [[ExecuteExtensionRequest.runtime_auth_config]] at request time. */
accessToken?: string;
/** The service account used to generate access tokens for executing the Extension. - If the service account is specified, the `iam.serviceAccounts.getAccessToken` permission should be granted to Vertex AI Extension Service Agent (https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents) on the provided service account. */
serviceAccount?: string;
}
/** Config for user OIDC auth. This data type is not supported in Gemini API. */
export declare interface AuthConfigOidcConfig {
/** OpenID Connect formatted ID token for extension endpoint. Only used to propagate token from [[ExecuteExtensionRequest.runtime_auth_config]] at request time. */
idToken?: string;
/** The service account used to generate an OpenID Connect (OIDC)-compatible JWT token signed by the Google OIDC Provider (accounts.google.com) for extension endpoint (https://cloud.google.com/iam/docs/create-short-lived-credentials-direct#sa-credentials-oidc). - The audience for the token will be set to the URL in the server url defined in the OpenApi spec. - If the service account is provided, the service account should grant `iam.serviceAccounts.getOpenIdToken` permission to Vertex AI Extension Service Agent (https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents). */
serviceAccount?: string;
}
declare class AuthenticationError extends APIError<401, Headers> {
}
/** Config for auth_tokens.create parameters. */
export declare interface AuthToken {
/** The name of the auth token. */
name?: string;
}
/** Type of auth scheme. This enum is not supported in Gemini API. */
export declare enum AuthType {
AUTH_TYPE_UNSPECIFIED = "AUTH_TYPE_UNSPECIFIED",
/**
* No Auth.
*/
NO_AUTH = "NO_AUTH",
/**
* API Key Auth.
*/
API_KEY_AUTH = "API_KEY_AUTH",
/**
* HTTP Basic Auth.
*/
HTTP_BASIC_AUTH = "HTTP_BASIC_AUTH",
/**
* Google Service Account Auth.
*/
GOOGLE_SERVICE_ACCOUNT_AUTH = "GOOGLE_SERVICE_ACCOUNT_AUTH",
/**
* OAuth auth.
*/
OAUTH = "OAUTH",
/**
* OpenID Connect (OIDC) Auth.
*/
OIDC_AUTH = "OIDC_AUTH"
}
/** Configures automatic detection of activity. */
export declare interface AutomaticActivityDetection {
/** If enabled, detected voice and text input count as activity. If disabled, the client must send activity signals. */
disabled?: boolean;
/** Determines how likely speech is to be detected. */
startOfSpeechSensitivity?: StartSensitivity;
/** Determines how likely detected speech is ended. */
endOfSpeechSensitivity?: EndSensitivity;
/** The required duration of detected speech before start-of-speech is committed. The lower this value the more sensitive the start-of-speech detection is and the shorter speech can be recognized. However, this also increases the probability of false positives. */
prefixPaddingMs?: number;
/** The required duration of detected non-speech (e.g. silence) before end-of-speech is committed. The larger this value, the longer speech gaps can be without interrupting the user's activity but this will increase the model's latency. */
silenceDurationMs?: number;
}
/** The configuration for automatic function calling. */
export declare interface AutomaticFunctionCallingConfig {
/** Whether to disable automatic function calling.
If not set or set to False, will enable automatic function calling.
If set to True, will disable automatic function calling.
*/
disable?: boolean;
/** If automatic function calling is enabled,
maximum number of remote calls for automatic function calling.
This number should be a positive integer.
If not set, SDK will set maximum number of remote calls to 10.
*/
maximumRemoteCalls?: number;
/** If automatic function calling is enabled,
whether to ignore call history to the response.
If not set, SDK will set ignore_call_history to false,
and will append the call history to
GenerateContentResponse.automatic_function_calling_history.
*/
ignoreCallHistory?: boolean;
}
declare class BadRequestError extends APIError<400, Headers> {
}
declare interface BaseCreateAgentInteractionParams {
/**
* Path param: Which version of the API to use.
*/
api_version?: string;
/**
* Body param: The name of the `Agent` used for generating the interaction.
*/
agent: (string & {}) | 'deep-research-pro-preview-12-2025';
/**
* Body param: The input for the interaction.
*/
input: Array | string | Array | TextContent | ImageContent | AudioContent | DocumentContent | VideoContent | ThoughtContent | FunctionCallContent | FunctionResultContent | CodeExecutionCallContent | CodeExecutionResultContent | URLContextCallContent | URLContextResultContent | GoogleSearchCallContent | GoogleSearchResultContent | MCPServerToolCallContent | MCPServerToolResultContent | FileSearchCallContent | FileSearchResultContent;
/**
* Body param: Configuration parameters for the agent interaction.
*/
agent_config?: DynamicAgentConfig | DeepResearchAgentConfig;
/**
* Body param: Input only. Whether to run the model interaction in the background.
*/
background?: boolean;
/**
* Body param: The ID of the previous interaction, if any.
*/
previous_interaction_id?: string;
/**
* Body param: Enforces that the generated response is a JSON object that complies with
* the JSON schema specified in this field.
*/
response_format?: unknown;
/**
* Body param: The mime type of the response. This is required if response_format is set.
*/
response_mime_type?: string;
/**
* Body param: The requested modalities of the response (TEXT, IMAGE, AUDIO).
*/
response_modalities?: Array<'text' | 'image' | 'audio'>;
/**
* Body param: Input only. Whether to store the response and request for later retrieval.
*/
store?: boolean;
/**
* Body param: Input only. Whether the interaction will be streamed.
*/
stream?: boolean;
/**
* Body param: System instruction for the interaction.
*/
system_instruction?: string;
/**
* Body param: A list of tool declarations the model may call during interaction.
*/
tools?: Array;
}
declare interface BaseCreateModelInteractionParams {
/**
* Path param: Which version of the API to use.
*/
api_version?: string;
/**
* Body param: The input for the interaction.
*/
input: Array | string | Array | TextContent | ImageContent | AudioContent | DocumentContent | VideoContent | ThoughtContent | FunctionCallContent | FunctionResultContent | CodeExecutionCallContent | CodeExecutionResultContent | URLContextCallContent | URLContextResultContent | GoogleSearchCallContent | GoogleSearchResultContent | MCPServerToolCallContent | MCPServerToolResultContent | FileSearchCallContent | FileSearchResultContent;
/**
* Body param: The name of the `Model` used for generating the interaction.
*/
model: Model_2;
/**
* Body param: Input only. Whether to run the model interaction in the background.
*/
background?: boolean;
/**
* Body param: Input only. Configuration parameters for the model interaction.
*/
generation_config?: GenerationConfig_2;
/**
* Body param: The ID of the previous interaction, if any.
*/
previous_interaction_id?: string;
/**
* Body param: Enforces that the generated response is a JSON object that complies with
* the JSON schema specified in this field.
*/
response_format?: unknown;
/**
* Body param: The mime type of the response. This is required if response_format is set.
*/
response_mime_type?: string;
/**
* Body param: The requested modalities of the response (TEXT, IMAGE, AUDIO).
*/
response_modalities?: Array<'text' | 'image' | 'audio'>;
/**
* Body param: Input only. Whether to store the response and request for later retrieval.
*/
store?: boolean;
/**
* Body param: Input only. Whether the interaction will be streamed.
*/
stream?: boolean;
/**
* Body param: System instruction for the interaction.
*/
system_instruction?: string;
/**
* Body param: A list of tool declarations the model may call during interaction.
*/
tools?: Array;
}
/**
* Base class for Gemini Next Gen API API clients.
*/
declare class BaseGeminiNextGenAPIClient {
apiKey: string | null;
apiVersion: string;
baseURL: string;
maxRetries: number;
timeout: number;
logger: Logger;
logLevel: LogLevel | undefined;
fetchOptions: MergedRequestInit | undefined;
private fetch;
private encoder;
protected idempotencyHeader?: string;
private _options;
private clientAdapter;
/**
* API Client for interfacing with the Gemini Next Gen API API.
*
* @param {string | null | undefined} [opts.apiKey=process.env['GEMINI_API_KEY'] ?? null]
* @param {string | undefined} [opts.apiVersion=v1beta]
* @param {string} [opts.baseURL=process.env['GEMINI_NEXT_GEN_API_BASE_URL'] ?? https://generativelanguage.googleapis.com] - Override the default base URL for the API.
* @param {number} [opts.timeout=1 minute] - The maximum amount of time (in milliseconds) the client will wait for a response before timing out.
* @param {MergedRequestInit} [opts.fetchOptions] - Additional `RequestInit` options to be passed to `fetch` calls.
* @param {Fetch} [opts.fetch] - Specify a custom `fetch` function implementation.
* @param {number} [opts.maxRetries=2] - The maximum number of times the client will retry a request.
* @param {HeadersLike} opts.defaultHeaders - Default headers to include with every request to the API.
* @param {Record} opts.defaultQuery - Default query parameters to include with every request to the API.
*/
constructor({ baseURL, apiKey, apiVersion, ...opts }: ClientOptions);
/**
* Create a new client instance re-using the same options given to the current client with optional overriding.
*/
withOptions(options: Partial): this;
/**
* Check whether the base URL is set to its default.
*/
private baseURLOverridden;
protected defaultQuery(): Record | undefined;
protected validateHeaders({ values, nulls }: NullableHeaders): void;
protected authHeaders(opts: FinalRequestOptions): Promise;
/**
* Basic re-implementation of `qs.stringify` for primitive types.
*/
protected stringifyQuery(query: object | Record): string;
private getUserAgent;
protected defaultIdempotencyKey(): string;
protected makeStatusError(status: number, error: Object, message: string | undefined, headers: Headers): Errors.APIError;
buildURL(path: string, query: Record | null | undefined, defaultBaseURL?: string | undefined): string;
/**
* Used as a callback for mutating the given `FinalRequestOptions` object.
*/
protected prepareOptions(options: FinalRequestOptions): Promise;
/**
* Used as a callback for mutating the given `RequestInit` object.
*
* This is useful for cases where you want to add certain headers based off of
* the request properties, e.g. `method` or `url`.
*/
protected prepareRequest(request: _RequestInit, { url, options }: {
url: string;
options: FinalRequestOptions;
}): Promise;
get(path: string, opts?: PromiseOrValue): APIPromise;
post(path: string, opts?: PromiseOrValue): APIPromise;
patch(path: string, opts?: PromiseOrValue): APIPromise;
put(path: string, opts?: PromiseOrValue): APIPromise;
delete(path: string, opts?: PromiseOrValue): APIPromise;
private methodRequest;
request(options: PromiseOrValue, remainingRetries?: number | null): APIPromise;
private makeRequest;
fetchWithTimeout(url: _RequestInfo, init: _RequestInit | undefined, ms: number, controller: AbortController): Promise;
private shouldRetry;
private retryRequest;
private calculateDefaultRetryTimeoutMillis;
buildRequest(inputOptions: FinalRequestOptions, { retryCount }?: {
retryCount?: number;
}): Promise<{
req: FinalizedRequestInit;
url: string;
timeout: number;
}>;
private buildHeaders;
private _makeAbort;
private buildBody;
static DEFAULT_TIMEOUT: number;
}
declare class BaseInteractions extends APIResource {
static readonly _key: readonly ['interactions'];
/**
* Creates a new interaction.
*
* @example
* ```ts
* const interaction = await client.interactions.create({
* api_version: 'api_version',
* input: [{ text: 'text', type: 'text' }],
* model: 'gemini-2.5-flash',
* });
* ```
*/
create(params: CreateModelInteractionParamsNonStreaming, options?: RequestOptions): APIPromise;
create(params: CreateModelInteractionParamsStreaming, options?: RequestOptions): APIPromise>;
create(params: CreateAgentInteractionParamsNonStreaming, options?: RequestOptions): APIPromise;
create(params: CreateAgentInteractionParamsStreaming, options?: RequestOptions): APIPromise>;
create(params: BaseCreateModelInteractionParams | BaseCreateAgentInteractionParams, options?: RequestOptions): APIPromise | Interaction>;
/**
* Deletes the interaction by id.
*
* @example
* ```ts
* const interaction = await client.interactions.delete('id', {
* api_version: 'api_version',
* });
* ```
*/
delete(id: string, params?: InteractionDeleteParams | null | undefined, options?: RequestOptions): APIPromise;
/**
* Cancels an interaction by id. This only applies to background interactions that are still running.
*
* @example
* ```ts
* const interaction = await client.interactions.cancel('id', {
* api_version: 'api_version',
* });
* ```
*/
cancel(id: string, params?: InteractionCancelParams | null | undefined, options?: RequestOptions): APIPromise;
/**
* Retrieves the full details of a single interaction based on its `Interaction.id`.
*
* @example
* ```ts
* const interaction = await client.interactions.get('id', {
* api_version: 'api_version',
* });
* ```
*/
get(id: string, params?: InteractionGetParamsNonStreaming, options?: RequestOptions): APIPromise;
get(id: string, params: InteractionGetParamsStreaming, options?: RequestOptions): APIPromise>;
get(id: string, params?: InteractionGetParamsBase | undefined, options?: RequestOptions): APIPromise | Interaction>;
}
/**
* @license
* Copyright 2025 Google LLC
* SPDX-License-Identifier: Apache-2.0
*/
declare class BaseModule {
}
/**
* Parameters for setting the base URLs for the Gemini API and Vertex AI API.
*/
export declare interface BaseUrlParameters {
geminiUrl?: string;
vertexUrl?: string;
}
export declare class Batches extends BaseModule {
private readonly apiClient;
constructor(apiClient: ApiClient);
/**
* Lists batch jobs.
*
* @param params - The parameters for the list request.
* @return - A pager of batch jobs.
*
* @example
* ```ts
* const batchJobs = await ai.batches.list({config: {'pageSize': 2}});
* for await (const batchJob of batchJobs) {
* console.log(batchJob);
* }
* ```
*/
list: (params?: types.ListBatchJobsParameters) => Promise>;
/**
* Create batch job.
*
* @param params - The parameters for create batch job request.
* @return The created batch job.
*
* @example
* ```ts
* const response = await ai.batches.create({
* model: 'gemini-2.0-flash',
* src: {gcsUri: 'gs://bucket/path/to/file.jsonl', format: 'jsonl'},
* config: {
* dest: {gcsUri: 'gs://bucket/path/output/directory', format: 'jsonl'},
* }
* });
* console.log(response);
* ```
*/
create: (params: types.CreateBatchJobParameters) => Promise;
/**
* **Experimental** Creates an embedding batch job.
*
* @param params - The parameters for create embedding batch job request.
* @return The created batch job.
*
* @example
* ```ts
* const response = await ai.batches.createEmbeddings({
* model: 'text-embedding-004',
* src: {fileName: 'files/my_embedding_input'},
* });
* console.log(response);
* ```
*/
createEmbeddings: (params: types.CreateEmbeddingsBatchJobParameters) => Promise;
private createInlinedGenerateContentRequest;
private getGcsUri;
private getBigqueryUri;
private formatDestination;
/**
* Internal method to create batch job.
*
* @param params - The parameters for create batch job request.
* @return The created batch job.
*
*/
private createInternal;
/**
* Internal method to create batch job.
*
* @param params - The parameters for create batch job request.
* @return The created batch job.
*
*/
private createEmbeddingsInternal;
/**
* Gets batch job configurations.
*
* @param params - The parameters for the get request.
* @return The batch job.
*
* @example
* ```ts
* await ai.batches.get({name: '...'}); // The server-generated resource name.
* ```
*/
get(params: types.GetBatchJobParameters): Promise;
/**
* Cancels a batch job.
*
* @param params - The parameters for the cancel request.
* @return The empty response returned by the API.
*
* @example
* ```ts
* await ai.batches.cancel({name: '...'}); // The server-generated resource name.
* ```
*/
cancel(params: types.CancelBatchJobParameters): Promise;
private listInternal;
/**
* Deletes a batch job.
*
* @param params - The parameters for the delete request.
* @return The empty response returned by the API.
*
* @example
* ```ts
* await ai.batches.delete({name: '...'}); // The server-generated resource name.
* ```
*/
delete(params: types.DeleteBatchJobParameters): Promise;
}
/** Config for batches.create return value. */
export declare interface BatchJob {
/** The resource name of the BatchJob. Output only.".
*/
name?: string;
/** The display name of the BatchJob.
*/
displayName?: string;
/** The state of the BatchJob.
*/
state?: JobState;
/** Output only. Only populated when the job's state is JOB_STATE_FAILED or JOB_STATE_CANCELLED. */
error?: JobError;
/** The time when the BatchJob was created.
*/
createTime?: string;
/** Output only. Time when the Job for the first time entered the `JOB_STATE_RUNNING` state. */
startTime?: string;
/** The time when the BatchJob was completed. This field is for Vertex AI only.
*/
endTime?: string;
/** The time when the BatchJob was last updated.
*/
updateTime?: string;
/** The name of the model that produces the predictions via the BatchJob.
*/
model?: string;
/** Configuration for the input data. This field is for Vertex AI only.
*/
src?: BatchJobSource;
/** Configuration for the output data.
*/
dest?: BatchJobDestination;
/** Statistics on completed and failed prediction instances. This field is for Vertex AI only.
*/
completionStats?: CompletionStats;
}
/** Config for `des` parameter. */
export declare interface BatchJobDestination {
/** Storage format of the output files. Must be one of:
'jsonl', 'bigquery'.
*/
format?: string;
/** The Google Cloud Storage URI to the output file.
*/
gcsUri?: string;
/** The BigQuery URI to the output table.
*/
bigqueryUri?: string;
/** The Gemini Developer API's file resource name of the output data
(e.g. "files/12345"). The file will be a JSONL file with a single response
per line. The responses will be GenerateContentResponse messages formatted
as JSON. The responses will be written in the same order as the input
requests.
*/
fileName?: string;
/** The responses to the requests in the batch. Returned when the batch was
built using inlined requests. The responses will be in the same order as
the input requests.
*/
inlinedResponses?: InlinedResponse[];
/** The responses to the requests in the batch. Returned when the batch was
built using inlined requests. The responses will be in the same order as
the input requests.
*/
inlinedEmbedContentResponses?: InlinedEmbedContentResponse[];
}
export declare type BatchJobDestinationUnion = BatchJobDestination | string;
/** Config for `src` parameter. */
export declare interface BatchJobSource {
/** Storage format of the input files. Must be one of:
'jsonl', 'bigquery'.
*/
format?: string;
/** The Google Cloud Storage URIs to input files.
*/
gcsUri?: string[];
/** The BigQuery URI to input table.
*/
bigqueryUri?: string;
/** The Gemini Developer API's file resource name of the input data
(e.g. "files/12345").
*/
fileName?: string;
/** The Gemini Developer API's inlined input data to run batch job.
*/
inlinedRequests?: InlinedRequest[];
}
export declare type BatchJobSourceUnion = BatchJobSource | InlinedRequest[] | string;
/** Specifies the function Behavior. Currently only supported by the BidiGenerateContent method. This enum is not supported in Vertex AI. */
export declare enum Behavior {
/**
* This value is unused.
*/
UNSPECIFIED = "UNSPECIFIED",
/**
* If set, the system will wait to receive the function response before continuing the conversation.
*/
BLOCKING = "BLOCKING",
/**
* If set, the system will not wait to receive the function response. Instead, it will attempt to handle function responses as they become available while maintaining the conversation between the user and the model.
*/
NON_BLOCKING = "NON_BLOCKING"
}
/** The BigQuery location for the input content. This data type is not supported in Gemini API. */
export declare interface BigQuerySource {
/** Required. BigQuery URI to a table, up to 2000 characters long. Accepted forms: * BigQuery path. For example: `bq://projectId.bqDatasetId.bqTableId`. */
inputUri?: string;
}
/** Bleu metric value for an instance. This data type is not supported in Gemini API. */
export declare interface BleuMetricValue {
/** Output only. Bleu score. */
score?: number;
}
/** A content blob. A Blob contains data of a specific media type. It is used to represent images, audio, and video. */
declare interface Blob_2 {
/** Required. The raw bytes of the data.
* @remarks Encoded as base64 string. */
data?: string;
/** Optional. The display name of the blob. Used to provide a label or filename to distinguish blobs. This field is only returned in `PromptMessage` for prompt management. It is used in the Gemini calls only when server-side tools (`code_execution`, `google_search`, and `url_context`) are enabled. This field is not supported in Gemini API. */
displayName?: string;
/** Required. The IANA standard MIME type of the source data. */
mimeType?: string;
}
export { Blob_2 as Blob }
export declare type BlobImageUnion = Blob_2;
/** Output only. The reason why the prompt was blocked. */
export declare enum BlockedReason {
/**
* The blocked reason is unspecified.
*/
BLOCKED_REASON_UNSPECIFIED = "BLOCKED_REASON_UNSPECIFIED",
/**
* The prompt was blocked for safety reasons.
*/
SAFETY = "SAFETY",
/**
* The prompt was blocked for other reasons. For example, it may be due to the prompt's language, or because it contains other harmful content.
*/
OTHER = "OTHER",
/**
* The prompt was blocked because it contains a term from the terminology blocklist.
*/
BLOCKLIST = "BLOCKLIST",
/**
* The prompt was blocked because it contains prohibited content.
*/
PROHIBITED_CONTENT = "PROHIBITED_CONTENT",
/**
* The prompt was blocked because it contains content that is unsafe for image generation.
*/
IMAGE_SAFETY = "IMAGE_SAFETY",
/**
* The prompt was blocked by Model Armor. This enum value is not supported in Gemini API.
*/
MODEL_ARMOR = "MODEL_ARMOR",
/**
* The prompt was blocked as a jailbreak attempt. This enum value is not supported in Gemini API.
*/
JAILBREAK = "JAILBREAK"
}
declare const brand_privateNullableHeaders: unique symbol;
/** @ts-ignore For users with \@types/bun */
declare type BunRequestInit = globalThis.FetchRequestInit;
/** A resource used in LLM queries for users to explicitly specify what to cache. */
export declare interface CachedContent {
/** The server-generated resource name of the cached content. */
name?: string;
/** The user-generated meaningful display name of the cached content. */
displayName?: string;
/** The name of the publisher model to use for cached content. */
model?: string;
/** Creation time of the cache entry. */
createTime?: string;
/** When the cache entry was last updated in UTC time. */
updateTime?: string;
/** Expiration time of the cached content. */
expireTime?: string;
/** Metadata on the usage of the cached content. */
usageMetadata?: CachedContentUsageMetadata;
}
/** Metadata on the usage of the cached content. */
export declare interface CachedContentUsageMetadata {
/** Duration of audio in seconds. This field is not supported in Gemini API. */
audioDurationSeconds?: number;
/** Number of images. This field is not supported in Gemini API. */
imageCount?: number;
/** Number of text characters. This field is not supported in Gemini API. */
textCount?: number;
/** Total number of tokens that the cached content consumes. */
totalTokenCount?: number;
/** Duration of video in seconds. This field is not supported in Gemini API. */
videoDurationSeconds?: number;
}
export declare class Caches extends BaseModule {
private readonly apiClient;
constructor(apiClient: ApiClient);
/**
* Lists cached contents.
*
* @param params - The parameters for the list request.
* @return - A pager of cached contents.
*
* @example
* ```ts
* const cachedContents = await ai.caches.list({config: {'pageSize': 2}});
* for await (const cachedContent of cachedContents) {
* console.log(cachedContent);
* }
* ```
*/
list: (params?: types.ListCachedContentsParameters) => Promise>;
/**
* Creates a cached contents resource.
*
* @remarks
* Context caching is only supported for specific models. See [Gemini
* Developer API reference](https://ai.google.dev/gemini-api/docs/caching?lang=node/context-cac)
* and [Vertex AI reference](https://cloud.google.com/vertex-ai/generative-ai/docs/context-cache/context-cache-overview#supported_models)
* for more information.
*
* @param params - The parameters for the create request.
* @return The created cached content.
*
* @example
* ```ts
* const contents = ...; // Initialize the content to cache.
* const response = await ai.caches.create({
* model: 'gemini-2.0-flash-001',
* config: {
* 'contents': contents,
* 'displayName': 'test cache',
* 'systemInstruction': 'What is the sum of the two pdfs?',
* 'ttl': '86400s',
* }
* });
* ```
*/
create(params: types.CreateCachedContentParameters): Promise;
/**
* Gets cached content configurations.
*
* @param params - The parameters for the get request.
* @return The cached content.
*
* @example
* ```ts
* await ai.caches.get({name: '...'}); // The server-generated resource name.
* ```
*/
get(params: types.GetCachedContentParameters): Promise;
/**
* Deletes cached content.
*
* @param params - The parameters for the delete request.
* @return The empty response returned by the API.
*
* @example
* ```ts
* await ai.caches.delete({name: '...'}); // The server-generated resource name.
* ```
*/
delete(params: types.DeleteCachedContentParameters): Promise;
/**
* Updates cached content configurations.
*
* @param params - The parameters for the update request.
* @return The updated cached content.
*
* @example
* ```ts
* const response = await ai.caches.update({
* name: '...', // The server-generated resource name.
* config: {'ttl': '7600s'}
* });
* ```
*/
update(params: types.UpdateCachedContentParameters): Promise;
private listInternal;
}
/**
* CallableTool is an invokable tool that can be executed with external
* application (e.g., via Model Context Protocol) or local functions with
* function calling.
*/
export declare interface CallableTool {
/**
* Returns tool that can be called by Gemini.
*/
tool(): Promise;
/**
* Executes the callable tool with the given function call arguments and
* returns the response parts from the tool execution.
*/
callTool(functionCalls: FunctionCall[]): Promise;
}
/**
* CallableToolConfig is the configuration for a callable tool.
*/
export declare interface CallableToolConfig {
/**
* Specifies the model's behavior after invoking this tool.
*/
behavior?: Behavior;
/**
* Timeout for remote calls in milliseconds. Note this timeout applies only to
* tool remote calls, and not making HTTP requests to the API. */
timeout?: number;
}
/** Optional parameters. */
export declare interface CancelBatchJobConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Config for batches.cancel parameters. */
export declare interface CancelBatchJobParameters {
/** A fully-qualified BatchJob resource name or ID.
Example: "projects/.../locations/.../batchPredictionJobs/456"
or "456" when project and location are initialized in the client.
*/
name: string;
/** Optional parameters for the request. */
config?: CancelBatchJobConfig;
}
/** Optional parameters for tunings.cancel method. */
export declare interface CancelTuningJobConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Parameters for the cancel method. */
export declare interface CancelTuningJobParameters {
/** The resource name of the tuning job. */
name: string;
/** Optional parameters for the request. */
config?: CancelTuningJobConfig;
}
/** Empty response for tunings.cancel method. */
export declare class CancelTuningJobResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
}
/** A response candidate generated from the model. */
export declare interface Candidate {
/** Contains the multi-part content of the response.
*/
content?: Content;
/** Source attribution of the generated content.
*/
citationMetadata?: CitationMetadata;
/** Describes the reason the model stopped generating tokens.
*/
finishMessage?: string;
/** Number of tokens for this candidate.
*/
tokenCount?: number;
/** The reason why the model stopped generating tokens.
If empty, the model has not stopped generating the tokens.
*/
finishReason?: FinishReason;
/** Output only. Metadata returned when grounding is enabled. It
contains the sources used to ground the generated content.
*/
groundingMetadata?: GroundingMetadata;
/** Output only. The average log probability of the tokens in this candidate. This is a length-normalized score that can be used to compare the quality of candidates of different lengths. A higher average log probability suggests a more confident and coherent response. */
avgLogprobs?: number;
/** Output only. The 0-based index of this candidate in the list of generated responses. This is useful for distinguishing between multiple candidates when `candidate_count` > 1. */
index?: number;
/** Output only. The detailed log probability information for the tokens in this candidate. This is useful for debugging, understanding model uncertainty, and identifying potential "hallucinations". */
logprobsResult?: LogprobsResult;
/** Output only. A list of ratings for the safety of a response candidate. There is at most one rating per category. */
safetyRatings?: SafetyRating[];
/** Output only. Metadata returned when the model uses the `url_context` tool to get information from a user-provided URL. */
urlContextMetadata?: UrlContextMetadata;
}
/**
* Chat session that enables sending messages to the model with previous
* conversation context.
*
* @remarks
* The session maintains all the turns between user and model.
*/
export declare class Chat {
private readonly apiClient;
private readonly modelsModule;
private readonly model;
private readonly config;
private history;
private sendPromise;
constructor(apiClient: ApiClient, modelsModule: Models, model: string, config?: types.GenerateContentConfig, history?: types.Content[]);
/**
* Sends a message to the model and returns the response.
*
* @remarks
* This method will wait for the previous message to be processed before
* sending the next message.
*
* @see {@link Chat#sendMessageStream} for streaming method.
* @param params - parameters for sending messages within a chat session.
* @returns The model's response.
*
* @example
* ```ts
* const chat = ai.chats.create({model: 'gemini-2.0-flash'});
* const response = await chat.sendMessage({
* message: 'Why is the sky blue?'
* });
* console.log(response.text);
* ```
*/
sendMessage(params: types.SendMessageParameters): Promise;
/**
* Sends a message to the model and returns the response in chunks.
*
* @remarks
* This method will wait for the previous message to be processed before
* sending the next message.
*
* @see {@link Chat#sendMessage} for non-streaming method.
* @param params - parameters for sending the message.
* @return The model's response.
*
* @example
* ```ts
* const chat = ai.chats.create({model: 'gemini-2.0-flash'});
* const response = await chat.sendMessageStream({
* message: 'Why is the sky blue?'
* });
* for await (const chunk of response) {
* console.log(chunk.text);
* }
* ```
*/
sendMessageStream(params: types.SendMessageParameters): Promise>;
/**
* Returns the chat history.
*
* @remarks
* The history is a list of contents alternating between user and model.
*
* There are two types of history:
* - The `curated history` contains only the valid turns between user and
* model, which will be included in the subsequent requests sent to the model.
* - The `comprehensive history` contains all turns, including invalid or
* empty model outputs, providing a complete record of the history.
*
* The history is updated after receiving the response from the model,
* for streaming response, it means receiving the last chunk of the response.
*
* The `comprehensive history` is returned by default. To get the `curated
* history`, set the `curated` parameter to `true`.
*
* @param curated - whether to return the curated history or the comprehensive
* history.
* @return History contents alternating between user and model for the entire
* chat session.
*/
getHistory(curated?: boolean): types.Content[];
private processStreamResponse;
private recordHistory;
}
/**
* A utility class to create a chat session.
*/
export declare class Chats {
private readonly modelsModule;
private readonly apiClient;
constructor(modelsModule: Models, apiClient: ApiClient);
/**
* Creates a new chat session.
*
* @remarks
* The config in the params will be used for all requests within the chat
* session unless overridden by a per-request `config` in
* @see {@link types.SendMessageParameters#config}.
*
* @param params - Parameters for creating a chat session.
* @returns A new chat session.
*
* @example
* ```ts
* const chat = ai.chats.create({
* model: 'gemini-2.0-flash'
* config: {
* temperature: 0.5,
* maxOutputTokens: 1024,
* }
* });
* ```
*/
create(params: types.CreateChatParameters): Chat;
}
/** Describes the machine learning model version checkpoint. */
export declare interface Checkpoint {
/** The ID of the checkpoint.
*/
checkpointId?: string;
/** The epoch of the checkpoint.
*/
epoch?: string;
/** The step of the checkpoint.
*/
step?: string;
}
/** Config for telling the service how to chunk the file. */
export declare interface ChunkingConfig {
/** White space chunking configuration. */
whiteSpaceConfig?: WhiteSpaceConfig;
}
/** A citation for a piece of generatedcontent. This data type is not supported in Gemini API. */
export declare interface Citation {
/** Output only. The end index of the citation in the content. */
endIndex?: number;
/** Output only. The license of the source of the citation. */
license?: string;
/** Output only. The publication date of the source of the citation. */
publicationDate?: GoogleTypeDate;
/** Output only. The start index of the citation in the content. */
startIndex?: number;
/** Output only. The title of the source of the citation. */
title?: string;
/** Output only. The URI of the source of the citation. */
uri?: string;
}
/** Citation information when the model quotes another source. */
export declare interface CitationMetadata {
/** Contains citation information when the model directly quotes, at
length, from another source. Can include traditional websites and code
repositories.
*/
citations?: Citation[];
}
declare interface ClientOptions {
/**
* Defaults to process.env['GEMINI_API_KEY'].
*/
apiKey?: string | null | undefined;
apiVersion?: string | undefined;
/**
* Override the default base URL for the API, e.g., "https://api.example.com/v2/"
*
* Defaults to process.env['GEMINI_NEXT_GEN_API_BASE_URL'].
*/
baseURL?: string | null | undefined;
/**
* The maximum amount of time (in milliseconds) that the client should wait for a response
* from the server before timing out a single request.
*
* Note that request timeouts are retried by default, so in a worst-case scenario you may wait
* much longer than this timeout before the promise succeeds or fails.
*
* @unit milliseconds
*/
timeout?: number | undefined;
/**
* Additional `RequestInit` options to be passed to `fetch` calls.
* Properties will be overridden by per-request `fetchOptions`.
*/
fetchOptions?: MergedRequestInit | undefined;
/**
* Specify a custom `fetch` function implementation.
*
* If not provided, we expect that `fetch` is defined globally.
*/
fetch?: Fetch | undefined;
/**
* The maximum number of times that the client will retry a request in case of a
* temporary failure, like a network error or a 5XX error from the server.
*
* @default 2
*/
maxRetries?: number | undefined;
/**
* Default headers to include with every request to the API.
*
* These can be removed in individual requests by explicitly setting the
* header to `null` in request options.
*/
defaultHeaders?: HeadersLike | undefined;
/**
* Default query parameters to include with every request to the API.
*
* These can be removed in individual requests by explicitly setting the
* param to `undefined` in request options.
*/
defaultQuery?: Record | undefined;
/**
* Set the log level.
*
* Defaults to process.env['GEMINI_NEXT_GEN_API_LOG'] or 'warn' if it isn't set.
*/
logLevel?: LogLevel | undefined;
/**
* Set the logger.
*
* Defaults to globalThis.console.
*/
logger?: Logger | undefined;
/**
* The adapter to the parent API client instance (for accessing auth, project, location)
*/
clientAdapter: GeminiNextGenAPIClientAdapter;
}
/**
* The arguments to pass to the code execution.
*/
declare interface CodeExecutionCallArguments {
/**
* The code to be executed.
*/
code?: string;
/**
* Programming language of the `code`.
*/
language?: 'python';
}
/**
* Code execution content.
*/
declare interface CodeExecutionCallContent {
/**
* A unique ID for this specific tool call.
*/
id: string;
/**
* The arguments to pass to the code execution.
*/
arguments: CodeExecutionCallArguments;
type: 'code_execution_call';
}
/** Result of executing the [ExecutableCode]. Only generated when using the [CodeExecution] tool, and always follows a `part` containing the [ExecutableCode]. */
export declare interface CodeExecutionResult {
/** Required. Outcome of the code execution. */
outcome?: Outcome;
/** Optional. Contains stdout when code execution is successful, stderr or other description otherwise. */
output?: string;
}
/**
* Code execution result content.
*/
declare interface CodeExecutionResultContent {
/**
* ID to match the ID from the code execution call block.
*/
call_id: string;
/**
* The output of the code execution.
*/
result: string;
type: 'code_execution_result';
/**
* Whether the code execution resulted in an error.
*/
is_error?: boolean;
/**
* A signature hash for backend validation.
*/
signature?: string;
}
/** Success and error statistics of processing multiple entities (for example, DataItems or structured data rows) in batch. This data type is not supported in Gemini API. */
export declare interface CompletionStats {
/** Output only. The number of entities for which any error was encountered. */
failedCount?: string;
/** Output only. In cases when enough errors are encountered a job, pipeline, or operation may be failed as a whole. Below is the number of entities for which the processing had not been finished (either in successful or failed state). Set to -1 if the number is unknown (for example, the operation failed before the total entity number could be collected). */
incompleteCount?: string;
/** Output only. The number of entities that had been processed successfully. */
successfulCount?: string;
/** Output only. The number of the successful forecast points that are generated by the forecasting model. This is ONLY used by the forecasting batch prediction. */
successfulForecastPointCount?: string;
}
/** Tool to support computer use. */
export declare interface ComputerUse {
/** Required. The environment being operated. */
environment?: Environment;
/** By default, predefined functions are included in the final model call.
Some of them can be explicitly excluded from being automatically included.
This can serve two purposes:
1. Using a more restricted / different action space.
2. Improving the definitions / instructions of predefined functions. */
excludedPredefinedFunctions?: string[];
}
/** Optional parameters for computing tokens. */
export declare interface ComputeTokensConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Parameters for computing tokens. */
export declare interface ComputeTokensParameters {
/** ID of the model to use. For a list of models, see `Google models
`_. */
model: string;
/** Input content. */
contents: ContentListUnion;
/** Optional parameters for the request.
*/
config?: ComputeTokensConfig;
}
/** Response for computing tokens. */
export declare class ComputeTokensResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
/** Lists of tokens info from the input. A ComputeTokensRequest could have multiple instances with a prompt in each instance. We also need to return lists of tokens info for the request with multiple instances. */
tokensInfo?: TokensInfo[];
}
/** Local tokenizer compute tokens result. */
export declare interface ComputeTokensResult {
/** Lists of tokens info from the input. */
tokensInfo?: TokensInfo[];
}
declare type _ConditionalNodeReadableStream = typeof globalThis extends {
ReadableStream: any;
} ? never : _NodeReadableStream;
declare class ConflictError extends APIError<409, Headers> {
}
/** Contains the multi-part content of a message. */
export declare interface Content {
/** List of parts that constitute a single message. Each part may have
a different IANA MIME type. */
parts?: Part[];
/** Optional. The producer of the content. Must be either 'user' or 'model'. If not set, the service will default to 'user'. */
role?: string;
}
/**
* The content of the response.
*/
declare type Content_2 = TextContent | ImageContent | AudioContent | DocumentContent | VideoContent | ThoughtContent | FunctionCallContent | FunctionResultContent | CodeExecutionCallContent | CodeExecutionResultContent | URLContextCallContent | URLContextResultContent | GoogleSearchCallContent | GoogleSearchResultContent | MCPServerToolCallContent | MCPServerToolResultContent | FileSearchCallContent | FileSearchResultContent;
declare interface ContentDelta {
delta: ContentDelta.TextDelta | ContentDelta.ImageDelta | ContentDelta.AudioDelta | ContentDelta.DocumentDelta | ContentDelta.VideoDelta | ContentDelta.ThoughtSummaryDelta | ContentDelta.ThoughtSignatureDelta | ContentDelta.FunctionCallDelta | ContentDelta.FunctionResultDelta | ContentDelta.CodeExecutionCallDelta | ContentDelta.CodeExecutionResultDelta | ContentDelta.URLContextCallDelta | ContentDelta.URLContextResultDelta | ContentDelta.GoogleSearchCallDelta | ContentDelta.GoogleSearchResultDelta | ContentDelta.MCPServerToolCallDelta | ContentDelta.MCPServerToolResultDelta | ContentDelta.FileSearchCallDelta | ContentDelta.FileSearchResultDelta;
event_type: 'content.delta';
index: number;
/**
* The event_id token to be used to resume the interaction stream, from this event.
*/
event_id?: string;
}
declare namespace ContentDelta {
interface TextDelta {
text: string;
type: 'text';
/**
* Citation information for model-generated content.
*/
annotations?: Array;
}
interface ImageDelta {
type: 'image';
data?: string;
mime_type?: 'image/png' | 'image/jpeg' | 'image/webp' | 'image/heic' | 'image/heif';
/**
* The resolution of the media.
*/
resolution?: 'low' | 'medium' | 'high' | 'ultra_high';
uri?: string;
}
interface AudioDelta {
type: 'audio';
data?: string;
mime_type?: 'audio/wav' | 'audio/mp3' | 'audio/aiff' | 'audio/aac' | 'audio/ogg' | 'audio/flac';
uri?: string;
}
interface DocumentDelta {
type: 'document';
data?: string;
mime_type?: 'application/pdf';
uri?: string;
}
interface VideoDelta {
type: 'video';
data?: string;
mime_type?: 'video/mp4' | 'video/mpeg' | 'video/mpg' | 'video/mov' | 'video/avi' | 'video/x-flv' | 'video/webm' | 'video/wmv' | 'video/3gpp';
/**
* The resolution of the media.
*/
resolution?: 'low' | 'medium' | 'high' | 'ultra_high';
uri?: string;
}
interface ThoughtSummaryDelta {
type: 'thought_summary';
/**
* A new summary item to be added to the thought.
*/
content?: InteractionsAPI.TextContent | InteractionsAPI.ImageContent;
}
interface ThoughtSignatureDelta {
type: 'thought_signature';
/**
* Signature to match the backend source to be part of the generation.
*/
signature?: string;
}
interface FunctionCallDelta {
/**
* A unique ID for this specific tool call.
*/
id: string;
arguments: {
[key: string]: unknown;
};
name: string;
type: 'function_call';
}
interface FunctionResultDelta {
/**
* ID to match the ID from the function call block.
*/
call_id: string;
/**
* Tool call result delta.
*/
result: FunctionResultDelta.Items | unknown | string;
type: 'function_result';
is_error?: boolean;
name?: string;
/**
* A signature hash for backend validation.
*/
signature?: string;
}
namespace FunctionResultDelta {
interface Items {
items?: Array;
}
}
interface CodeExecutionCallDelta {
/**
* A unique ID for this specific tool call.
*/
id: string;
/**
* The arguments to pass to the code execution.
*/
arguments: InteractionsAPI.CodeExecutionCallArguments;
type: 'code_execution_call';
}
interface CodeExecutionResultDelta {
/**
* ID to match the ID from the function call block.
*/
call_id: string;
result: string;
type: 'code_execution_result';
is_error?: boolean;
/**
* A signature hash for backend validation.
*/
signature?: string;
}
interface URLContextCallDelta {
/**
* A unique ID for this specific tool call.
*/
id: string;
/**
* The arguments to pass to the URL context.
*/
arguments: InteractionsAPI.URLContextCallArguments;
type: 'url_context_call';
}
interface URLContextResultDelta {
/**
* ID to match the ID from the function call block.
*/
call_id: string;
result: Array;
type: 'url_context_result';
is_error?: boolean;
/**
* A signature hash for backend validation.
*/
signature?: string;
}
interface GoogleSearchCallDelta {
/**
* A unique ID for this specific tool call.
*/
id: string;
/**
* The arguments to pass to Google Search.
*/
arguments: InteractionsAPI.GoogleSearchCallArguments;
type: 'google_search_call';
}
interface GoogleSearchResultDelta {
/**
* ID to match the ID from the function call block.
*/
call_id: string;
result: Array;
type: 'google_search_result';
is_error?: boolean;
/**
* A signature hash for backend validation.
*/
signature?: string;
}
interface MCPServerToolCallDelta {
/**
* A unique ID for this specific tool call.
*/
id: string;
arguments: {
[key: string]: unknown;
};
name: string;
server_name: string;
type: 'mcp_server_tool_call';
}
interface MCPServerToolResultDelta {
/**
* ID to match the ID from the function call block.
*/
call_id: string;
/**
* Tool call result delta.
*/
result: MCPServerToolResultDelta.Items | unknown | string;
type: 'mcp_server_tool_result';
name?: string;
server_name?: string;
/**
* A signature hash for backend validation.
*/
signature?: string;
}
namespace MCPServerToolResultDelta {
interface Items {
items?: Array;
}
}
interface FileSearchCallDelta {
/**
* A unique ID for this specific tool call.
*/
id: string;
type: 'file_search_call';
}
interface FileSearchResultDelta {
/**
* ID to match the ID from the function call block.
*/
call_id: string;
type: 'file_search_result';
result?: Array;
/**
* A signature hash for backend validation.
*/
signature?: string;
}
namespace FileSearchResultDelta {
/**
* The result of the File Search.
*/
interface Result {
/**
* The name of the file search store.
*/
file_search_store?: string;
/**
* The text of the search result.
*/
text?: string;
/**
* The title of the search result.
*/
title?: string;
}
}
}
/** The embedding generated from an input content. */
export declare interface ContentEmbedding {
/** A list of floats representing an embedding.
*/
values?: number[];
/** Vertex API only. Statistics of the input text associated with this
embedding.
*/
statistics?: ContentEmbeddingStatistics;
}
/** Statistics of the input text associated with the result of content embedding. */
export declare interface ContentEmbeddingStatistics {
/** Vertex API only. If the input text was truncated due to having
a length longer than the allowed maximum input.
*/
truncated?: boolean;
/** Vertex API only. Number of tokens of the input text.
*/
tokenCount?: number;
}
export declare type ContentListUnion = Content | Content[] | PartUnion | PartUnion[];
/** A content reference image.
A content reference image represents a subject to reference (ex. person,
product, animal) provided by the user. It can optionally be provided in
addition to a style reference image (ex. background, style reference). */
export declare class ContentReferenceImage {
/** The reference image for the editing operation. */
referenceImage?: Image_2;
/** The id of the reference image. */
referenceId?: number;
/** The type of the reference image. Only set by the SDK. */
referenceType?: string;
/** Internal method to convert to ReferenceImageAPIInternal. */
toReferenceImageAPI(): ReferenceImageAPIInternal;
}
declare interface ContentStart {
/**
* The content of the response.
*/
content: Content_2;
event_type: 'content.start';
index: number;
/**
* The event_id token to be used to resume the interaction stream, from this event.
*/
event_id?: string;
}
declare interface ContentStop {
event_type: 'content.stop';
index: number;
/**
* The event_id token to be used to resume the interaction stream, from this event.
*/
event_id?: string;
}
export declare type ContentUnion = Content | PartUnion[] | PartUnion;
/** Enables context window compression -- mechanism managing model context window so it does not exceed given length. */
export declare interface ContextWindowCompressionConfig {
/** Number of tokens (before running turn) that triggers context window compression mechanism. */
triggerTokens?: string;
/** Sliding window compression mechanism. */
slidingWindow?: SlidingWindow;
}
/** Configuration for a Control reference image. */
export declare interface ControlReferenceConfig {
/** The type of control reference image to use. */
controlType?: ControlReferenceType;
/** Defaults to False. When set to True, the control image will be
computed by the model based on the control type. When set to False,
the control image must be provided by the user. */
enableControlImageComputation?: boolean;
}
/** A control reference image.
The image of the control reference image is either a control image provided
by the user, or a regular image which the backend will use to generate a
control image of. In the case of the latter, the
enable_control_image_computation field in the config should be set to True.
A control image is an image that represents a sketch image of areas for the
model to fill in based on the prompt. */
export declare class ControlReferenceImage {
/** The reference image for the editing operation. */
referenceImage?: Image_2;
/** The id of the reference image. */
referenceId?: number;
/** The type of the reference image. Only set by the SDK. */
referenceType?: string;
/** Configuration for the control reference image. */
config?: ControlReferenceConfig;
/** Internal method to convert to ReferenceImageAPIInternal. */
toReferenceImageAPI(): ReferenceImageAPIInternal;
}
/** Enum representing the control type of a control reference image. */
export declare enum ControlReferenceType {
CONTROL_TYPE_DEFAULT = "CONTROL_TYPE_DEFAULT",
CONTROL_TYPE_CANNY = "CONTROL_TYPE_CANNY",
CONTROL_TYPE_SCRIBBLE = "CONTROL_TYPE_SCRIBBLE",
CONTROL_TYPE_FACE_MESH = "CONTROL_TYPE_FACE_MESH"
}
/** Config for the count_tokens method. */
export declare interface CountTokensConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** Instructions for the model to steer it toward better performance.
*/
systemInstruction?: ContentUnion;
/** Code that enables the system to interact with external systems to
perform an action outside of the knowledge and scope of the model.
*/
tools?: Tool[];
/** Configuration that the model uses to generate the response. Not
supported by the Gemini Developer API.
*/
generationConfig?: GenerationConfig;
}
/** Parameters for counting tokens. */
export declare interface CountTokensParameters {
/** ID of the model to use. For a list of models, see `Google models
`_. */
model: string;
/** Input content. */
contents: ContentListUnion;
/** Configuration for counting tokens. */
config?: CountTokensConfig;
}
/** Response for counting tokens. */
export declare class CountTokensResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
/** Total number of tokens. */
totalTokens?: number;
/** Number of tokens in the cached part of the prompt (the cached content). */
cachedContentTokenCount?: number;
}
/** Local tokenizer count tokens result. */
export declare interface CountTokensResult {
/** The total number of tokens. */
totalTokens?: number;
}
declare interface CreateAgentInteractionParamsNonStreaming extends BaseCreateAgentInteractionParams {
/**
* Body param: Input only. Whether the interaction will be streamed.
*/
stream?: false;
}
declare interface CreateAgentInteractionParamsStreaming extends BaseCreateAgentInteractionParams {
/**
* Body param: Input only. Whether the interaction will be streamed.
*/
stream: true;
}
/** Optional parameters. */
export declare interface CreateAuthTokenConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** An optional time after which, when using the resulting token,
messages in Live API sessions will be rejected. (Gemini may
preemptively close the session after this time.)
If not set then this defaults to 30 minutes in the future. If set, this
value must be less than 20 hours in the future. */
expireTime?: string;
/** The time after which new Live API sessions using the token
resulting from this request will be rejected.
If not set this defaults to 60 seconds in the future. If set, this value
must be less than 20 hours in the future. */
newSessionExpireTime?: string;
/** The number of times the token can be used. If this value is zero
then no limit is applied. Default is 1. Resuming a Live API session does
not count as a use. */
uses?: number;
/** Configuration specific to Live API connections created using this token. */
liveConnectConstraints?: LiveConnectConstraints;
/** Additional fields to lock in the effective LiveConnectParameters. */
lockAdditionalFields?: string[];
}
/** Config for auth_tokens.create parameters. */
export declare interface CreateAuthTokenParameters {
/** Optional parameters for the request. */
config?: CreateAuthTokenConfig;
}
/** Config for optional parameters. */
export declare interface CreateBatchJobConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** The user-defined name of this BatchJob.
*/
displayName?: string;
/** GCS or BigQuery URI prefix for the output predictions. Example:
"gs://path/to/output/data" or "bq://projectId.bqDatasetId.bqTableId".
*/
dest?: BatchJobDestinationUnion;
}
/** Config for batches.create parameters. */
export declare interface CreateBatchJobParameters {
/** The name of the model to produces the predictions via the BatchJob.
*/
model?: string;
/** GCS URI(-s) or BigQuery URI to your input data to run batch job.
Example: "gs://path/to/input/data" or "bq://projectId.bqDatasetId.bqTableId".
*/
src: BatchJobSourceUnion;
/** Optional parameters for creating a BatchJob.
*/
config?: CreateBatchJobConfig;
}
/** Optional configuration for cached content creation. */
export declare interface CreateCachedContentConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** The TTL for this resource. The expiration time is computed: now + TTL. It is a duration string, with up to nine fractional digits, terminated by 's'. Example: "3.5s". */
ttl?: string;
/** Timestamp of when this resource is considered expired. Uses RFC 3339 format, Example: 2014-10-02T15:01:23Z. */
expireTime?: string;
/** The user-generated meaningful display name of the cached content.
*/
displayName?: string;
/** The content to cache.
*/
contents?: ContentListUnion;
/** Developer set system instruction.
*/
systemInstruction?: ContentUnion;
/** A list of `Tools` the model may use to generate the next response.
*/
tools?: Tool[];
/** Configuration for the tools to use. This config is shared for all tools.
*/
toolConfig?: ToolConfig;
/** The Cloud KMS resource identifier of the customer managed
encryption key used to protect a resource.
The key needs to be in the same region as where the compute resource is
created. See
https://cloud.google.com/vertex-ai/docs/general/cmek for more
details. If this is set, then all created CachedContent objects
will be encrypted with the provided encryption key.
Allowed formats: projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{crypto_key}
*/
kmsKeyName?: string;
}
/** Parameters for caches.create method. */
export declare interface CreateCachedContentParameters {
/** ID of the model to use. Example: gemini-2.0-flash */
model: string;
/** Configuration that contains optional parameters.
*/
config?: CreateCachedContentConfig;
}
/** Parameters for initializing a new chat session.
These parameters are used when creating a chat session with the
`chats.create()` method. */
export declare interface CreateChatParameters {
/** The name of the model to use for the chat session.
For example: 'gemini-2.0-flash', 'gemini-2.0-flash-lite', etc. See Gemini API
docs to find the available models.
*/
model: string;
/** Config for the entire chat session.
This config applies to all requests within the session
unless overridden by a per-request `config` in `SendMessageParameters`.
*/
config?: GenerateContentConfig;
/** The initial conversation history for the chat session.
This allows you to start the chat with a pre-existing history. The history
must be a list of `Content` alternating between 'user' and 'model' roles.
It should start with a 'user' message.
*/
history?: Content[];
}
/** Config for optional parameters. */
export declare interface CreateEmbeddingsBatchJobConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** The user-defined name of this BatchJob.
*/
displayName?: string;
}
/** Config for batches.create parameters. */
export declare interface CreateEmbeddingsBatchJobParameters {
/** The name of the model to produces the predictions via the BatchJob.
*/
model?: string;
/** input data to run batch job".
*/
src: EmbeddingsBatchJobSource;
/** Optional parameters for creating a BatchJob.
*/
config?: CreateEmbeddingsBatchJobConfig;
}
/** Used to override the default configuration. */
export declare interface CreateFileConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Generates the parameters for the private _create method. */
export declare interface CreateFileParameters {
/** The file to be uploaded.
mime_type: (Required) The MIME type of the file. Must be provided.
name: (Optional) The name of the file in the destination (e.g.
'files/sample-image').
display_name: (Optional) The display name of the file.
*/
file: File_2;
/** Used to override the default configuration. */
config?: CreateFileConfig;
}
/** Response for the create file method. */
export declare class CreateFileResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
}
/** Optional parameters for creating a file search store. */
export declare interface CreateFileSearchStoreConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** The human-readable display name for the file search store.
*/
displayName?: string;
}
/** Config for file_search_stores.create parameters. */
export declare interface CreateFileSearchStoreParameters {
/** Optional parameters for creating a file search store.
*/
config?: CreateFileSearchStoreConfig;
}
/**
* Creates a `FunctionResponsePart` object from a `base64` encoded `string`.
*/
export declare function createFunctionResponsePartFromBase64(data: string, mimeType: string): FunctionResponsePart;
/**
* Creates a `FunctionResponsePart` object from a `URI` string.
*/
export declare function createFunctionResponsePartFromUri(uri: string, mimeType: string): FunctionResponsePart;
/**
* Creates a `Content` object with a model role from a `PartListUnion` object or `string`.
*/
export declare function createModelContent(partOrString: PartListUnion | string): Content;
declare interface CreateModelInteractionParamsNonStreaming extends BaseCreateModelInteractionParams {
/**
* Body param: Input only. Whether the interaction will be streamed.
*/
stream?: false;
}
declare interface CreateModelInteractionParamsStreaming extends BaseCreateModelInteractionParams {
/**
* Body param: Input only. Whether the interaction will be streamed.
*/
stream: true;
}
/**
* Creates a `Part` object from a `base64` encoded `string`.
*/
export declare function createPartFromBase64(data: string, mimeType: string, mediaResolution?: PartMediaResolutionLevel): Part;
/**
* Creates a `Part` object from the `outcome` and `output` of a `CodeExecutionResult` object.
*/
export declare function createPartFromCodeExecutionResult(outcome: Outcome, output: string): Part;
/**
* Creates a `Part` object from the `code` and `language` of an `ExecutableCode` object.
*/
export declare function createPartFromExecutableCode(code: string, language: Language): Part;
/**
* Creates a `Part` object from a `FunctionCall` object.
*/
export declare function createPartFromFunctionCall(name: string, args: Record): Part;
/**
* Creates a `Part` object from a `FunctionResponse` object.
*/
export declare function createPartFromFunctionResponse(id: string, name: string, response: Record, parts?: FunctionResponsePart[]): Part;
/**
* Creates a `Part` object from a `text` string.
*/
export declare function createPartFromText(text: string): Part;
/**
* Creates a `Part` object from a `URI` string.
*/
export declare function createPartFromUri(uri: string, mimeType: string, mediaResolution?: PartMediaResolutionLevel): Part;
/** Fine-tuning job creation request - optional fields. */
export declare interface CreateTuningJobConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** The method to use for tuning (SUPERVISED_FINE_TUNING or PREFERENCE_TUNING or DISTILLATION). If not set, the default method (SFT) will be used. */
method?: TuningMethod;
/** Validation dataset for tuning. The dataset must be formatted as a JSONL file. */
validationDataset?: TuningValidationDataset;
/** The display name of the tuned Model. The name can be up to 128 characters long and can consist of any UTF-8 characters. */
tunedModelDisplayName?: string;
/** The description of the TuningJob */
description?: string;
/** Number of complete passes the model makes over the entire training dataset during training. */
epochCount?: number;
/** Multiplier for adjusting the default learning rate. 1P models only. Mutually exclusive with learning_rate. */
learningRateMultiplier?: number;
/** If set to true, disable intermediate checkpoints and only the last checkpoint will be exported. Otherwise, enable intermediate checkpoints. */
exportLastCheckpointOnly?: boolean;
/** The optional checkpoint id of the pre-tuned model to use for tuning, if applicable. */
preTunedModelCheckpointId?: string;
/** Adapter size for tuning. */
adapterSize?: AdapterSize;
/** Tuning mode for SFT tuning. */
tuningMode?: TuningMode;
/** Custom base model for tuning. This is only supported for OSS models in Vertex. */
customBaseModel?: string;
/** The batch size hyperparameter for tuning. This is only supported for OSS models in Vertex. */
batchSize?: number;
/** The learning rate for tuning. OSS models only. Mutually exclusive with learning_rate_multiplier. */
learningRate?: number;
/** Optional. The labels with user-defined metadata to organize TuningJob and generated resources such as Model and Endpoint. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. */
labels?: Record;
/** Weight for KL Divergence regularization, Preference Optimization tuning only. */
beta?: number;
/** The base teacher model that is being distilled. Distillation only. */
baseTeacherModel?: string;
/** The resource name of the Tuned teacher model. Distillation only. */
tunedTeacherModelSource?: string;
/** Multiplier for adjusting the weight of the SFT loss. Distillation only. */
sftLossWeightMultiplier?: number;
/** The Google Cloud Storage location where the tuning job outputs are written. */
outputUri?: string;
/** The encryption spec of the tuning job. Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with provided encryption key. */
encryptionSpec?: EncryptionSpec;
}
/** Fine-tuning job creation parameters - optional fields. */
export declare interface CreateTuningJobParameters {
/** The base model that is being tuned, e.g., "gemini-2.5-flash". */
baseModel: string;
/** Cloud Storage path to file containing training dataset for tuning. The dataset must be formatted as a JSONL file. */
trainingDataset: TuningDataset;
/** Configuration for the tuning job. */
config?: CreateTuningJobConfig;
}
/** Fine-tuning job creation parameters - optional fields. */
export declare interface CreateTuningJobParametersPrivate {
/** The base model that is being tuned, e.g., "gemini-2.5-flash". */
baseModel?: string;
/** The PreTunedModel that is being tuned. */
preTunedModel?: PreTunedModel;
/** Cloud Storage path to file containing training dataset for tuning. The dataset must be formatted as a JSONL file. */
trainingDataset: TuningDataset;
/** Configuration for the tuning job. */
config?: CreateTuningJobConfig;
}
/**
* Creates a `Content` object with a user role from a `PartListUnion` object or `string`.
*/
export declare function createUserContent(partOrString: PartListUnion | string): Content;
/** Result for custom code execution metric. This data type is not supported in Gemini API. */
export declare interface CustomCodeExecutionResult {
/** Output only. Custom code execution score. */
score?: number;
}
/** User provided metadata stored as key-value pairs. This data type is not supported in Vertex AI. */
export declare interface CustomMetadata {
/** Required. The key of the metadata to store. */
key?: string;
/** The numeric value of the metadata to store. */
numericValue?: number;
/** The StringList value of the metadata to store. */
stringListValue?: StringList;
/** The string value of the metadata to store. */
stringValue?: string;
}
/** Spec for custom output. This data type is not supported in Gemini API. */
export declare interface CustomOutput {
/** Output only. List of raw output strings. */
rawOutputs?: RawOutput;
}
/** Distribution computed over a tuning dataset. This data type is not supported in Gemini API. */
export declare interface DatasetDistribution {
/** Output only. Defines the histogram bucket. */
buckets?: DatasetDistributionDistributionBucket[];
/** Output only. The maximum of the population values. */
max?: number;
/** Output only. The arithmetic mean of the values in the population. */
mean?: number;
/** Output only. The median of the values in the population. */
median?: number;
/** Output only. The minimum of the population values. */
min?: number;
/** Output only. The 5th percentile of the values in the population. */
p5?: number;
/** Output only. The 95th percentile of the values in the population. */
p95?: number;
/** Output only. Sum of a given population of values. */
sum?: number;
}
/** Dataset bucket used to create a histogram for the distribution given a population of values. This data type is not supported in Gemini API. */
export declare interface DatasetDistributionDistributionBucket {
/** Output only. Number of values in the bucket. */
count?: string;
/** Output only. Left bound of the bucket. */
left?: number;
/** Output only. Right bound of the bucket. */
right?: number;
}
/** Statistics computed over a tuning dataset. This data type is not supported in Gemini API. */
export declare interface DatasetStats {
/** Output only. A partial sample of the indices (starting from 1) of the dropped examples. */
droppedExampleIndices?: string[];
/** Output only. For each index in `dropped_example_indices`, the user-facing reason why the example was dropped. */
droppedExampleReasons?: string[];
/** Output only. Number of billable characters in the tuning dataset. */
totalBillableCharacterCount?: string;
/** Output only. Number of tuning characters in the tuning dataset. */
totalTuningCharacterCount?: string;
/** Output only. Number of examples in the tuning dataset. */
tuningDatasetExampleCount?: string;
/** Output only. Number of tuning steps for this Tuning Job. */
tuningStepCount?: string;
/** Output only. Sample user messages in the training dataset uri. */
userDatasetExamples?: Content[];
/** Output only. Dataset distributions for the user input tokens. */
userInputTokenDistribution?: DatasetDistribution;
/** Output only. Dataset distributions for the messages per example. */
userMessagePerExampleDistribution?: DatasetDistribution;
/** Output only. Dataset distributions for the user output tokens. */
userOutputTokenDistribution?: DatasetDistribution;
}
/**
* Configuration for the Deep Research agent.
*/
declare interface DeepResearchAgentConfig {
type: 'deep-research';
/**
* Whether to include thought summaries in the response.
*/
thinking_summaries?: 'auto' | 'none';
}
/** Optional parameters for models.get method. */
export declare interface DeleteBatchJobConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Config for batches.delete parameters. */
export declare interface DeleteBatchJobParameters {
/** A fully-qualified BatchJob resource name or ID.
Example: "projects/.../locations/.../batchPredictionJobs/456"
or "456" when project and location are initialized in the client.
*/
name: string;
/** Optional parameters for the request. */
config?: DeleteBatchJobConfig;
}
/** Optional parameters for caches.delete method. */
export declare interface DeleteCachedContentConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Parameters for caches.delete method. */
export declare interface DeleteCachedContentParameters {
/** The server-generated resource name of the cached content.
*/
name: string;
/** Optional parameters for the request.
*/
config?: DeleteCachedContentConfig;
}
/** Empty response for caches.delete method. */
export declare class DeleteCachedContentResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
}
/** Config for optional parameters. */
export declare interface DeleteDocumentConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** If set to true, any `Chunk`s and objects related to this `Document` will
also be deleted.
*/
force?: boolean;
}
/** Config for documents.delete parameters. */
export declare interface DeleteDocumentParameters {
/** The resource name of the Document.
Example: fileSearchStores/file-search-store-foo/documents/documents-bar */
name: string;
/** Optional parameters for the request. */
config?: DeleteDocumentConfig;
}
/** Used to override the default configuration. */
export declare interface DeleteFileConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Generates the parameters for the get method. */
export declare interface DeleteFileParameters {
/** The name identifier for the file to be deleted. */
name: string;
/** Used to override the default configuration. */
config?: DeleteFileConfig;
}
/** Response for the delete file method. */
export declare class DeleteFileResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
}
/** Optional parameters for deleting a FileSearchStore. */
export declare interface DeleteFileSearchStoreConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** If set to true, any Documents and objects related to this FileSearchStore will also be deleted.
If false (the default), a FAILED_PRECONDITION error will be returned if
the FileSearchStore contains any Documents.
*/
force?: boolean;
}
/** Config for file_search_stores.delete parameters. */
export declare interface DeleteFileSearchStoreParameters {
/** The resource name of the FileSearchStore. Example: `fileSearchStores/my-file-search-store-123` */
name: string;
/** Optional parameters for the request. */
config?: DeleteFileSearchStoreConfig;
}
/** Configuration for deleting a tuned model. */
export declare interface DeleteModelConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Parameters for deleting a tuned model. */
export declare interface DeleteModelParameters {
model: string;
/** Optional parameters for the request. */
config?: DeleteModelConfig;
}
export declare class DeleteModelResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
}
/** The return value of delete operation. */
export declare interface DeleteResourceJob {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
name?: string;
done?: boolean;
error?: JobError;
}
/** Statistics for distillation prompt dataset. These statistics do not include the responses sampled from the teacher model. This data type is not supported in Gemini API. */
export declare interface DistillationDataStats {
/** Output only. Statistics computed for the training dataset. */
trainingDatasetStats?: DatasetStats;
}
/** Hyperparameters for Distillation. This data type is not supported in Gemini API. */
export declare interface DistillationHyperParameters {
/** Optional. Adapter size for distillation. */
adapterSize?: AdapterSize;
/** Optional. Number of complete passes the model makes over the entire training dataset during training. */
epochCount?: string;
/** Optional. Multiplier for adjusting the default learning rate. */
learningRateMultiplier?: number;
}
/** Distillation tuning spec for tuning. */
export declare interface DistillationSpec {
/** The GCS URI of the prompt dataset to use during distillation. */
promptDatasetUri?: string;
/** The base teacher model that is being distilled. See [Supported models](https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/tuning#supported_models). */
baseTeacherModel?: string;
/** Optional. Hyperparameters for Distillation. */
hyperParameters?: DistillationHyperParameters;
/** Deprecated. A path in a Cloud Storage bucket, which will be treated as the root output directory of the distillation pipeline. It is used by the system to generate the paths of output artifacts. */
pipelineRootDirectory?: string;
/** The student model that is being tuned, e.g., "google/gemma-2b-1.1-it". Deprecated. Use base_model instead. */
studentModel?: string;
/** Deprecated. Cloud Storage path to file containing training dataset for tuning. The dataset must be formatted as a JSONL file. */
trainingDatasetUri?: string;
/** The resource name of the Tuned teacher model. Format: `projects/{project}/locations/{location}/models/{model}`. */
tunedTeacherModelSource?: string;
/** Optional. Cloud Storage path to file containing validation dataset for tuning. The dataset must be formatted as a JSONL file. */
validationDatasetUri?: string;
}
/** A Document is a collection of Chunks. */
declare interface Document_2 {
/** The resource name of the Document.
Example: fileSearchStores/file-search-store-foo/documents/documents-bar */
name?: string;
/** The human-readable display name for the Document. */
displayName?: string;
/** The current state of the Document. */
state?: DocumentState;
/** The size of the Document in bytes. */
sizeBytes?: string;
/** The MIME type of the Document. */
mimeType?: string;
/** Output only. The Timestamp of when the `Document` was created. */
createTime?: string;
/** Optional. User provided custom metadata stored as key-value pairs used for querying. A `Document` can have a maximum of 20 `CustomMetadata`. */
customMetadata?: CustomMetadata[];
/** Output only. The Timestamp of when the `Document` was last updated. */
updateTime?: string;
}
export { Document_2 as Document }
/**
* A document content block.
*/
declare interface DocumentContent {
type: 'document';
/**
* The document content.
*/
data?: string;
/**
* The mime type of the document.
*/
mime_type?: 'application/pdf';
/**
* The URI of the document.
*/
uri?: string;
}
declare class Documents extends BaseModule {
private readonly apiClient;
constructor(apiClient: ApiClient);
/**
* Lists documents.
*
* @param params - The parameters for the list request.
* @return - A pager of documents.
*
* @example
* ```ts
* const documents = await ai.documents.list({parent:'rag_store_name', config: {'pageSize': 2}});
* for await (const document of documents) {
* console.log(document);
* }
* ```
*/
list: (params: types.ListDocumentsParameters) => Promise>;
/**
* Gets a Document.
*
* @param params - The parameters for getting a document.
* @return Document.
*/
get(params: types.GetDocumentParameters): Promise;
/**
* Deletes a Document.
*
* @param params - The parameters for deleting a document.
*/
delete(params: types.DeleteDocumentParameters): Promise;
private listInternal;
}
/** State for the lifecycle of a Document. */
export declare enum DocumentState {
STATE_UNSPECIFIED = "STATE_UNSPECIFIED",
STATE_PENDING = "STATE_PENDING",
STATE_ACTIVE = "STATE_ACTIVE",
STATE_FAILED = "STATE_FAILED"
}
/** @ts-ignore */
declare type _DOMReadableStream = globalThis.ReadableStream;
export declare type DownloadableFileUnion = string | File_2 | GeneratedVideo | Video;
declare interface Downloader {
/**
* Downloads a file to the given location.
*
* @param params The parameters for downloading the file.
* @param apiClient The ApiClient to use for uploading.
* @return A Promises that resolves when the download is complete.
*/
download(params: DownloadFileParameters, apiClient: ApiClient): Promise;
}
/** Used to override the default configuration. */
export declare interface DownloadFileConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Parameters used to download a file. */
export declare interface DownloadFileParameters {
/** The file to download. It can be a file name, a file object or a generated video. */
file: DownloadableFileUnion;
/** Location where the file should be downloaded to. */
downloadPath: string;
/** Configuration to for the download operation. */
config?: DownloadFileConfig;
}
/**
* Configuration for dynamic agents.
*/
declare interface DynamicAgentConfig {
type: 'dynamic';
[k: string]: unknown;
}
/** Describes the options to customize dynamic retrieval. */
export declare interface DynamicRetrievalConfig {
/** Optional. The threshold to be used in dynamic retrieval. If not set, a system default value is used. */
dynamicThreshold?: number;
/** The mode of the predictor to be used in dynamic retrieval. */
mode?: DynamicRetrievalConfigMode;
}
/** The mode of the predictor to be used in dynamic retrieval. */
export declare enum DynamicRetrievalConfigMode {
/**
* Always trigger retrieval.
*/
MODE_UNSPECIFIED = "MODE_UNSPECIFIED",
/**
* Run retrieval only when system decides it is necessary.
*/
MODE_DYNAMIC = "MODE_DYNAMIC"
}
/** Configuration for editing an image. */
export declare interface EditImageConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** Cloud Storage URI used to store the generated images. */
outputGcsUri?: string;
/** Description of what to discourage in the generated images. */
negativePrompt?: string;
/** Number of images to generate. */
numberOfImages?: number;
/** Aspect ratio of the generated images. Supported values are
"1:1", "3:4", "4:3", "9:16", and "16:9". */
aspectRatio?: string;
/** Controls how much the model adheres to the text prompt. Large
values increase output and prompt alignment, but may compromise image
quality. */
guidanceScale?: number;
/** Random seed for image generation. This is not available when
``add_watermark`` is set to true. */
seed?: number;
/** Filter level for safety filtering. */
safetyFilterLevel?: SafetyFilterLevel;
/** Allows generation of people by the model. */
personGeneration?: PersonGeneration;
/** Whether to report the safety scores of each generated image and
the positive prompt in the response. */
includeSafetyAttributes?: boolean;
/** Whether to include the Responsible AI filter reason if the image
is filtered out of the response. */
includeRaiReason?: boolean;
/** Language of the text in the prompt. */
language?: ImagePromptLanguage;
/** MIME type of the generated image. */
outputMimeType?: string;
/** Compression quality of the generated image (for ``image/jpeg``
only). */
outputCompressionQuality?: number;
/** Whether to add a watermark to the generated images. */
addWatermark?: boolean;
/** User specified labels to track billing usage. */
labels?: Record;
/** Describes the editing mode for the request. */
editMode?: EditMode;
/** The number of sampling steps. A higher value has better image
quality, while a lower value has better latency. */
baseSteps?: number;
}
/** Parameters for the request to edit an image. */
export declare interface EditImageParameters {
/** The model to use. */
model: string;
/** A text description of the edit to apply to the image. */
prompt: string;
/** The reference images for Imagen 3 editing. */
referenceImages: ReferenceImage[];
/** Configuration for editing. */
config?: EditImageConfig;
}
/** Response for the request to edit an image. */
export declare class EditImageResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
/** Generated images. */
generatedImages?: GeneratedImage[];
}
/** Enum representing the editing mode. */
export declare enum EditMode {
EDIT_MODE_DEFAULT = "EDIT_MODE_DEFAULT",
EDIT_MODE_INPAINT_REMOVAL = "EDIT_MODE_INPAINT_REMOVAL",
EDIT_MODE_INPAINT_INSERTION = "EDIT_MODE_INPAINT_INSERTION",
EDIT_MODE_OUTPAINT = "EDIT_MODE_OUTPAINT",
EDIT_MODE_CONTROLLED_EDITING = "EDIT_MODE_CONTROLLED_EDITING",
EDIT_MODE_STYLE = "EDIT_MODE_STYLE",
EDIT_MODE_BGSWAP = "EDIT_MODE_BGSWAP",
EDIT_MODE_PRODUCT_IMAGE = "EDIT_MODE_PRODUCT_IMAGE"
}
/** Parameters for the embed_content method. */
export declare interface EmbedContentBatch {
/** The content to embed. Only the `parts.text` fields will be counted.
*/
contents?: ContentListUnion;
/** Configuration that contains optional parameters.
*/
config?: EmbedContentConfig;
}
/** Optional parameters for the embed_content method. */
export declare interface EmbedContentConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** Type of task for which the embedding will be used.
*/
taskType?: string;
/** Title for the text. Only applicable when TaskType is
`RETRIEVAL_DOCUMENT`.
*/
title?: string;
/** Reduced dimension for the output embedding. If set,
excessive values in the output embedding are truncated from the end.
Supported by newer models since 2024 only. You cannot set this value if
using the earlier model (`models/embedding-001`).
*/
outputDimensionality?: number;
/** Vertex API only. The MIME type of the input.
*/
mimeType?: string;
/** Vertex API only. Whether to silently truncate inputs longer than
the max sequence length. If this option is set to false, oversized inputs
will lead to an INVALID_ARGUMENT error, similar to other text APIs.
*/
autoTruncate?: boolean;
}
/** Request-level metadata for the Vertex Embed Content API. */
export declare interface EmbedContentMetadata {
/** Vertex API only. The total number of billable characters included
in the request.
*/
billableCharacterCount?: number;
}
/** Parameters for the embed_content method. */
export declare interface EmbedContentParameters {
/** ID of the model to use. For a list of models, see `Google models
`_. */
model: string;
/** The content to embed. Only the `parts.text` fields will be counted.
*/
contents: ContentListUnion;
/** Configuration that contains optional parameters.
*/
config?: EmbedContentConfig;
}
/** Parameters for the _embed_content method. */
export declare interface EmbedContentParametersPrivate {
/** ID of the model to use. For a list of models, see `Google models
`_. */
model: string;
/** The content to embed. Only the `parts.text` fields will be counted.
*/
contents?: ContentListUnion;
/** The single content to embed. Only the `parts.text` fields will be counted.
*/
content?: ContentUnion;
/** The Vertex embedding API to use.
*/
embeddingApiType?: EmbeddingApiType;
/** Configuration that contains optional parameters.
*/
config?: EmbedContentConfig;
}
/** Response for the embed_content method. */
export declare class EmbedContentResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
/** The embeddings for each request, in the same order as provided in
the batch request.
*/
embeddings?: ContentEmbedding[];
/** Vertex API only. Metadata about the request.
*/
metadata?: EmbedContentMetadata;
}
/** Enum representing the Vertex embedding API to use. */
export declare enum EmbeddingApiType {
/**
* predict API endpoint (default)
*/
PREDICT = "PREDICT",
/**
* embedContent API Endpoint
*/
EMBED_CONTENT = "EMBED_CONTENT"
}
export declare interface EmbeddingsBatchJobSource {
/** The Gemini Developer API's file resource name of the input data
(e.g. "files/12345").
*/
fileName?: string;
/** The Gemini Developer API's inlined input data to run batch job.
*/
inlinedRequests?: EmbedContentBatch;
}
/** Represents a customer-managed encryption key specification that can be applied to a Vertex AI resource. This data type is not supported in Gemini API. */
export declare interface EncryptionSpec {
/** Required. Resource name of the Cloud KMS key used to protect the resource. The Cloud KMS key must be in the same region as the resource. It must have the format `projects/{project}/locations/{location}/keyRings/{key_ring}/cryptoKeys/{crypto_key}`. */
kmsKeyName?: string;
}
/** An endpoint where you deploy models. */
export declare interface Endpoint {
/** Resource name of the endpoint. */
name?: string;
/** ID of the model that's deployed to the endpoint. */
deployedModelId?: string;
}
/** End of speech sensitivity. */
export declare enum EndSensitivity {
/**
* The default is END_SENSITIVITY_LOW.
*/
END_SENSITIVITY_UNSPECIFIED = "END_SENSITIVITY_UNSPECIFIED",
/**
* Automatic detection ends speech more often.
*/
END_SENSITIVITY_HIGH = "END_SENSITIVITY_HIGH",
/**
* Automatic detection ends speech less often.
*/
END_SENSITIVITY_LOW = "END_SENSITIVITY_LOW"
}
/** Tool to search public web data, powered by Vertex AI Search and Sec4 compliance. This data type is not supported in Gemini API. */
export declare interface EnterpriseWebSearch {
/** Optional. Sites with confidence level chosen & above this value will be blocked from the search results. */
blockingConfidence?: PhishBlockThreshold;
/** Optional. List of domains to be excluded from the search results. The default limit is 2000 domains. */
excludeDomains?: string[];
}
/** An entity representing the segmented area. */
export declare interface EntityLabel {
/** The label of the segmented entity. */
label?: string;
/** The confidence score of the detected label. */
score?: number;
}
/** The environment being operated. */
export declare enum Environment {
/**
* Defaults to browser.
*/
ENVIRONMENT_UNSPECIFIED = "ENVIRONMENT_UNSPECIFIED",
/**
* Operates in a web browser.
*/
ENVIRONMENT_BROWSER = "ENVIRONMENT_BROWSER"
}
declare interface ErrorEvent_2 {
event_type: 'error';
/**
* Error message from an interaction.
*/
error?: ErrorEvent_2.Error;
/**
* The event_id token to be used to resume the interaction stream, from this event.
*/
event_id?: string;
}
declare namespace ErrorEvent_2 {
/**
* Error message from an interaction.
*/
interface Error {
/**
* A URI that identifies the error type.
*/
code?: string;
/**
* A human-readable error message.
*/
message?: string;
}
}
declare namespace Errors {
export {
GeminiNextGenAPIClientError,
APIError,
APIUserAbortError,
APIConnectionError,
APIConnectionTimeoutError,
BadRequestError,
AuthenticationError,
PermissionDeniedError,
NotFoundError,
ConflictError,
UnprocessableEntityError,
RateLimitError,
InternalServerError
}
}
/** The results from an evaluation run performed by the EvaluationService. This data type is not supported in Gemini API. */
export declare class EvaluateDatasetResponse {
/** Output only. Aggregation statistics derived from results of EvaluationService. */
aggregationOutput?: AggregationOutput;
/** Output only. Output info for EvaluationService. */
outputInfo?: OutputInfo;
}
/** Evaluate Dataset Run Result for Tuning Job. This data type is not supported in Gemini API. */
export declare interface EvaluateDatasetRun {
/** Output only. The checkpoint id used in the evaluation run. Only populated when evaluating checkpoints. */
checkpointId?: string;
/** Output only. The error of the evaluation run if any. */
error?: GoogleRpcStatus;
/** Output only. Results for EvaluationService. */
evaluateDatasetResponse?: EvaluateDatasetResponse;
/** Output only. The resource name of the evaluation run. Format: `projects/{project}/locations/{location}/evaluationRuns/{evaluation_run_id}`. */
evaluationRun?: string;
/** Output only. The operation ID of the evaluation run. Format: `projects/{project}/locations/{location}/operations/{operation_id}`. */
operationName?: string;
}
/** The dataset used for evaluation. This data type is not supported in Gemini API. */
export declare interface EvaluationDataset {
/** BigQuery source holds the dataset. */
bigquerySource?: BigQuerySource;
/** Cloud storage source holds the dataset. Currently only one Cloud Storage file path is supported. */
gcsSource?: GcsSource;
}
/** Exact match metric value for an instance. This data type is not supported in Gemini API. */
export declare interface ExactMatchMetricValue {
/** Output only. Exact match score. */
score?: number;
}
/** Code generated by the model that is meant to be executed, and the result returned to the model. Generated when using the [CodeExecution] tool, in which the code will be automatically executed, and a corresponding [CodeExecutionResult] will also be generated. */
export declare interface ExecutableCode {
/** Required. The code to be executed. */
code?: string;
/** Required. Programming language of the `code`. */
language?: Language;
}
/** Retrieve from data source powered by external API for grounding. The external API is not owned by Google, but need to follow the pre-defined API spec. This data type is not supported in Gemini API. */
export declare interface ExternalApi {
/** The authentication config to access the API. Deprecated. Please use auth_config instead. */
apiAuth?: ApiAuth;
/** The API spec that the external API implements. */
apiSpec?: ApiSpec;
/** The authentication config to access the API. */
authConfig?: AuthConfig;
/** Parameters for the elastic search API. */
elasticSearchParams?: ExternalApiElasticSearchParams;
/** The endpoint of the external API. The system will call the API at this endpoint to retrieve the data for grounding. Example: https://acme.com:443/search */
endpoint?: string;
/** Parameters for the simple search API. */
simpleSearchParams?: ExternalApiSimpleSearchParams;
}
/** The search parameters to use for the ELASTIC_SEARCH spec. This data type is not supported in Gemini API. */
export declare interface ExternalApiElasticSearchParams {
/** The ElasticSearch index to use. */
index?: string;
/** Optional. Number of hits (chunks) to request. When specified, it is passed to Elasticsearch as the `num_hits` param. */
numHits?: number;
/** The ElasticSearch search template to use. */
searchTemplate?: string;
}
/** The search parameters to use for SIMPLE_SEARCH spec. This data type is not supported in Gemini API. */
export declare interface ExternalApiSimpleSearchParams {
}
/** Options for feature selection preference. */
export declare enum FeatureSelectionPreference {
FEATURE_SELECTION_PREFERENCE_UNSPECIFIED = "FEATURE_SELECTION_PREFERENCE_UNSPECIFIED",
PRIORITIZE_QUALITY = "PRIORITIZE_QUALITY",
BALANCED = "BALANCED",
PRIORITIZE_COST = "PRIORITIZE_COST"
}
/**
* @license
* Copyright 2025 Google LLC
* SPDX-License-Identifier: Apache-2.0
*/
declare type Fetch = (input: string | URL | Request, init?: RequestInit) => Promise;
export declare interface FetchPredictOperationConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Parameters for the fetchPredictOperation method. */
export declare interface FetchPredictOperationParameters {
/** The server-assigned name for the operation. */
operationName: string;
resourceName: string;
/** Used to override the default configuration. */
config?: FetchPredictOperationConfig;
}
/** @ts-ignore For users who use Deno */
declare type FetchRequestInit = NonNullable[1]>;
/** A file uploaded to the API. */
declare interface File_2 {
/** The `File` resource name. The ID (name excluding the "files/" prefix) can contain up to 40 characters that are lowercase alphanumeric or dashes (-). The ID cannot start or end with a dash. If the name is empty on create, a unique name will be generated. Example: `files/123-456` */
name?: string;
/** Optional. The human-readable display name for the `File`. The display name must be no more than 512 characters in length, including spaces. Example: 'Welcome Image' */
displayName?: string;
/** Output only. MIME type of the file. */
mimeType?: string;
/** Output only. Size of the file in bytes. */
sizeBytes?: string;
/** Output only. The timestamp of when the `File` was created. */
createTime?: string;
/** Output only. The timestamp of when the `File` will be deleted. Only set if the `File` is scheduled to expire. */
expirationTime?: string;
/** Output only. The timestamp of when the `File` was last updated. */
updateTime?: string;
/** Output only. SHA-256 hash of the uploaded bytes. The hash value is encoded in base64 format. */
sha256Hash?: string;
/** Output only. The URI of the `File`. */
uri?: string;
/** Output only. The URI of the `File`, only set for downloadable (generated) files. */
downloadUri?: string;
/** Output only. Processing state of the File. */
state?: FileState;
/** Output only. The source of the `File`. */
source?: FileSource;
/** Output only. Metadata for a video. */
videoMetadata?: Record;
/** Output only. Error status if File processing failed. */
error?: FileStatus;
}
export { File_2 as File }
/** URI-based data. A FileData message contains a URI pointing to data of a specific media type. It is used to represent images, audio, and video stored in Google Cloud Storage. */
export declare interface FileData {
/** Optional. The display name of the file. Used to provide a label or filename to distinguish files. This field is only returned in `PromptMessage` for prompt management. It is used in the Gemini calls only when server side tools (`code_execution`, `google_search`, and `url_context`) are enabled. This field is not supported in Gemini API. */
displayName?: string;
/** Required. The URI of the file in Google Cloud Storage. */
fileUri?: string;
/** Required. The IANA standard MIME type of the source data. */
mimeType?: string;
}
export declare class Files extends BaseModule {
private readonly apiClient;
constructor(apiClient: ApiClient);
/**
* Lists files.
*
* @param params - The parameters for the list request.
* @return - A pager of files.
*
* @example
* ```ts
* const files = await ai.files.list({config: {'pageSize': 2}});
* for await (const file of files) {
* console.log(file);
* }
* ```
*/
list: (params?: types.ListFilesParameters) => Promise>;
/**
* Uploads a file asynchronously to the Gemini API.
* This method is not available in Vertex AI.
* Supported upload sources:
* - Node.js: File path (string) or Blob object.
* - Browser: Blob object (e.g., File).
*
* @remarks
* The `mimeType` can be specified in the `config` parameter. If omitted:
* - For file path (string) inputs, the `mimeType` will be inferred from the
* file extension.
* - For Blob object inputs, the `mimeType` will be set to the Blob's `type`
* property.
* Somex eamples for file extension to mimeType mapping:
* .txt -> text/plain
* .json -> application/json
* .jpg -> image/jpeg
* .png -> image/png
* .mp3 -> audio/mpeg
* .mp4 -> video/mp4
*
* This section can contain multiple paragraphs and code examples.
*
* @param params - Optional parameters specified in the
* `types.UploadFileParameters` interface.
* @see {@link types.UploadFileParameters#config} for the optional
* config in the parameters.
* @return A promise that resolves to a `types.File` object.
* @throws An error if called on a Vertex AI client.
* @throws An error if the `mimeType` is not provided and can not be inferred,
* the `mimeType` can be provided in the `params.config` parameter.
* @throws An error occurs if a suitable upload location cannot be established.
*
* @example
* The following code uploads a file to Gemini API.
*
* ```ts
* const file = await ai.files.upload({file: 'file.txt', config: {
* mimeType: 'text/plain',
* }});
* console.log(file.name);
* ```
*/
upload(params: types.UploadFileParameters): Promise;
/**
* Downloads a remotely stored file asynchronously to a location specified in
* the `params` object. This method only works on Node environment, to
* download files in the browser, use a browser compliant method like an
* tag.
*
* @param params - The parameters for the download request.
*
* @example
* The following code downloads an example file named "files/mehozpxf877d" as
* "file.txt".
*
* ```ts
* await ai.files.download({file: file.name, downloadPath: 'file.txt'});
* ```
*/
download(params: types.DownloadFileParameters): Promise;
/**
* Registers Google Cloud Storage files for use with the API.
* This method is only available in Node.js environments.
*/
registerFiles(params: types.RegisterFilesParameters): Promise;
protected _registerFiles(params: types.InternalRegisterFilesParameters): Promise;
private listInternal;
private createInternal;
/**
* Retrieves the file information from the service.
*
* @param params - The parameters for the get request
* @return The Promise that resolves to the types.File object requested.
*
* @example
* ```ts
* const config: GetFileParameters = {
* name: fileName,
* };
* file = await ai.files.get(config);
* console.log(file.name);
* ```
*/
get(params: types.GetFileParameters): Promise;
/**
* Deletes a remotely stored file.
*
* @param params - The parameters for the delete request.
* @return The DeleteFileResponse, the response for the delete method.
*
* @example
* The following code deletes an example file named "files/mehozpxf877d".
*
* ```ts
* await ai.files.delete({name: file.name});
* ```
*/
delete(params: types.DeleteFileParameters): Promise;
private registerFilesInternal;
}
/** Tool to retrieve knowledge from the File Search Stores. */
export declare interface FileSearch {
/** The names of the file_search_stores to retrieve from.
Example: `fileSearchStores/my-file-search-store-123` */
fileSearchStoreNames?: string[];
/** The number of file search retrieval chunks to retrieve. */
topK?: number;
/** Metadata filter to apply to the file search retrieval documents. See https://google.aip.dev/160 for the syntax of the filter expression. */
metadataFilter?: string;
}
/**
* File Search content.
*/
declare interface FileSearchCallContent {
/**
* A unique ID for this specific tool call.
*/
id: string;
type: 'file_search_call';
}
/**
* File Search result content.
*/
declare interface FileSearchResultContent {
/**
* ID to match the ID from the file search call block.
*/
call_id: string;
type: 'file_search_result';
/**
* The results of the File Search.
*/
result?: Array;
/**
* A signature hash for backend validation.
*/
signature?: string;
}
declare namespace FileSearchResultContent {
/**
* The result of the File Search.
*/
interface Result {
/**
* The name of the file search store.
*/
file_search_store?: string;
/**
* The text of the search result.
*/
text?: string;
/**
* The title of the search result.
*/
title?: string;
}
}
/** A collection of Documents. */
export declare interface FileSearchStore {
/** The resource name of the FileSearchStore. Example: `fileSearchStores/my-file-search-store-123` */
name?: string;
/** The human-readable display name for the FileSearchStore. */
displayName?: string;
/** The Timestamp of when the FileSearchStore was created. */
createTime?: string;
/** The Timestamp of when the FileSearchStore was last updated. */
updateTime?: string;
/** The number of documents in the FileSearchStore that are active and ready for retrieval. */
activeDocumentsCount?: string;
/** The number of documents in the FileSearchStore that are being processed. */
pendingDocumentsCount?: string;
/** The number of documents in the FileSearchStore that have failed processing. */
failedDocumentsCount?: string;
/** The size of raw bytes ingested into the FileSearchStore. This is the
total size of all the documents in the FileSearchStore. */
sizeBytes?: string;
}
declare class FileSearchStores extends BaseModule {
private readonly apiClient;
readonly documents: Documents;
constructor(apiClient: ApiClient, documents?: Documents);
/**
* Lists file search stores.
*
* @param params - The parameters for the list request.
* @return - A pager of file search stores.
*
* @example
* ```ts
* const fileSearchStores = await ai.fileSearchStores.list({config: {'pageSize': 2}});
* for await (const fileSearchStore of fileSearchStores) {
* console.log(fileSearchStore);
* }
* ```
*/
list: (params?: types.ListFileSearchStoresParameters) => Promise>;
/**
* Uploads a file asynchronously to a given File Search Store.
* This method is not available in Vertex AI.
* Supported upload sources:
* - Node.js: File path (string) or Blob object.
* - Browser: Blob object (e.g., File).
*
* @remarks
* The `mimeType` can be specified in the `config` parameter. If omitted:
* - For file path (string) inputs, the `mimeType` will be inferred from the
* file extension.
* - For Blob object inputs, the `mimeType` will be set to the Blob's `type`
* property.
*
* This section can contain multiple paragraphs and code examples.
*
* @param params - Optional parameters specified in the
* `types.UploadToFileSearchStoreParameters` interface.
* @see {@link types.UploadToFileSearchStoreParameters#config} for the optional
* config in the parameters.
* @return A promise that resolves to a long running operation.
* @throws An error if called on a Vertex AI client.
* @throws An error if the `mimeType` is not provided and can not be inferred,
* the `mimeType` can be provided in the `params.config` parameter.
* @throws An error occurs if a suitable upload location cannot be established.
*
* @example
* The following code uploads a file to a given file search store.
*
* ```ts
* const operation = await ai.fileSearchStores.upload({fileSearchStoreName: 'fileSearchStores/foo-bar', file: 'file.txt', config: {
* mimeType: 'text/plain',
* }});
* console.log(operation.name);
* ```
*/
uploadToFileSearchStore(params: types.UploadToFileSearchStoreParameters): Promise;
/**
* Creates a File Search Store.
*
* @param params - The parameters for creating a File Search Store.
* @return FileSearchStore.
*/
create(params: types.CreateFileSearchStoreParameters): Promise;
/**
* Gets a File Search Store.
*
* @param params - The parameters for getting a File Search Store.
* @return FileSearchStore.
*/
get(params: types.GetFileSearchStoreParameters): Promise;
/**
* Deletes a File Search Store.
*
* @param params - The parameters for deleting a File Search Store.
*/
delete(params: types.DeleteFileSearchStoreParameters): Promise;
private listInternal;
private uploadToFileSearchStoreInternal;
/**
* Imports a File from File Service to a FileSearchStore.
*
* This is a long-running operation, see aip.dev/151
*
* @param params - The parameters for importing a file to a file search store.
* @return ImportFileOperation.
*/
importFile(params: types.ImportFileParameters): Promise;
}
/** Source of the File. */
export declare enum FileSource {
SOURCE_UNSPECIFIED = "SOURCE_UNSPECIFIED",
UPLOADED = "UPLOADED",
GENERATED = "GENERATED",
REGISTERED = "REGISTERED"
}
/**
* Represents the size and mimeType of a file. The information is used to
* request the upload URL from the https://generativelanguage.googleapis.com/upload/v1beta/files endpoint.
* This interface defines the structure for constructing and executing HTTP
* requests.
*/
declare interface FileStat {
/**
* The size of the file in bytes.
*/
size: number;
/**
* The MIME type of the file.
*/
type: string | undefined;
}
/** State for the lifecycle of a File. */
export declare enum FileState {
STATE_UNSPECIFIED = "STATE_UNSPECIFIED",
PROCESSING = "PROCESSING",
ACTIVE = "ACTIVE",
FAILED = "FAILED"
}
/** Status of a File that uses a common error model. */
export declare interface FileStatus {
/** A list of messages that carry the error details. There is a common set of message types for APIs to use. */
details?: Record[];
/** A list of messages that carry the error details. There is a common set of message types for APIs to use. */
message?: string;
/** The status code. 0 for OK, 1 for CANCELLED */
code?: number;
}
declare type FinalizedRequestInit = RequestInit & {
headers: Headers;
};
declare type FinalRequestOptions = RequestOptions & {
method: HTTPMethod;
path: string;
};
/** Output only. The reason why the model stopped generating tokens.
If empty, the model has not stopped generating the tokens. */
export declare enum FinishReason {
/**
* The finish reason is unspecified.
*/
FINISH_REASON_UNSPECIFIED = "FINISH_REASON_UNSPECIFIED",
/**
* Token generation reached a natural stopping point or a configured stop sequence.
*/
STOP = "STOP",
/**
* Token generation reached the configured maximum output tokens.
*/
MAX_TOKENS = "MAX_TOKENS",
/**
* Token generation stopped because the content potentially contains safety violations. NOTE: When streaming, [content][] is empty if content filters blocks the output.
*/
SAFETY = "SAFETY",
/**
* The token generation stopped because of potential recitation.
*/
RECITATION = "RECITATION",
/**
* The token generation stopped because of using an unsupported language.
*/
LANGUAGE = "LANGUAGE",
/**
* All other reasons that stopped the token generation.
*/
OTHER = "OTHER",
/**
* Token generation stopped because the content contains forbidden terms.
*/
BLOCKLIST = "BLOCKLIST",
/**
* Token generation stopped for potentially containing prohibited content.
*/
PROHIBITED_CONTENT = "PROHIBITED_CONTENT",
/**
* Token generation stopped because the content potentially contains Sensitive Personally Identifiable Information (SPII).
*/
SPII = "SPII",
/**
* The function call generated by the model is invalid.
*/
MALFORMED_FUNCTION_CALL = "MALFORMED_FUNCTION_CALL",
/**
* Token generation stopped because generated images have safety violations.
*/
IMAGE_SAFETY = "IMAGE_SAFETY",
/**
* The tool call generated by the model is invalid.
*/
UNEXPECTED_TOOL_CALL = "UNEXPECTED_TOOL_CALL",
/**
* Image generation stopped because the generated images have prohibited content.
*/
IMAGE_PROHIBITED_CONTENT = "IMAGE_PROHIBITED_CONTENT",
/**
* The model was expected to generate an image, but none was generated.
*/
NO_IMAGE = "NO_IMAGE",
/**
* Image generation stopped because the generated image may be a recitation from a source.
*/
IMAGE_RECITATION = "IMAGE_RECITATION",
/**
* Image generation stopped for a reason not otherwise specified.
*/
IMAGE_OTHER = "IMAGE_OTHER"
}
/** Tuning Spec for Full Fine Tuning. This data type is not supported in Gemini API. */
export declare interface FullFineTuningSpec {
/** Optional. Hyperparameters for Full Fine Tuning. */
hyperParameters?: SupervisedHyperParameters;
/** Required. Training dataset used for tuning. The dataset can be specified as either a Cloud Storage path to a JSONL file or as the resource name of a Vertex Multimodal Dataset. */
trainingDatasetUri?: string;
/** Optional. Validation dataset used for tuning. The dataset can be specified as either a Cloud Storage path to a JSONL file or as the resource name of a Vertex Multimodal Dataset. */
validationDatasetUri?: string;
}
/**
* A tool that can be used by the model.
*/
declare interface Function_2 {
type: 'function';
/**
* A description of the function.
*/
description?: string;
/**
* The name of the function.
*/
name?: string;
/**
* The JSON Schema for the function's parameters.
*/
parameters?: unknown;
}
/** A function call. */
export declare interface FunctionCall {
/** The unique id of the function call. If populated, the client to execute the
`function_call` and return the response with the matching `id`. */
id?: string;
/** Optional. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. */
args?: Record;
/** Optional. The name of the function to call. Matches [FunctionDeclaration.name]. */
name?: string;
/** Optional. The partial argument value of the function call. If provided, represents the arguments/fields that are streamed incrementally. This field is not supported in Gemini API. */
partialArgs?: PartialArg[];
/** Optional. Whether this is the last part of the FunctionCall. If true, another partial message for the current FunctionCall is expected to follow. This field is not supported in Gemini API. */
willContinue?: boolean;
}
/**
* A function tool call content block.
*/
declare interface FunctionCallContent {
/**
* A unique ID for this specific tool call.
*/
id: string;
/**
* The arguments to pass to the function.
*/
arguments: {
[key: string]: unknown;
};
/**
* The name of the tool to call.
*/
name: string;
type: 'function_call';
}
/** Function calling config. */
export declare interface FunctionCallingConfig {
/** Optional. Function names to call. Only set when the Mode is ANY. Function names should match [FunctionDeclaration.name]. With mode set to ANY, model will predict a function call from the set of function names provided. */
allowedFunctionNames?: string[];
/** Optional. Function calling mode. */
mode?: FunctionCallingConfigMode;
/** Optional. When set to true, arguments of a single function call will be streamed out in multiple parts/contents/responses. Partial parameter results will be returned in the [FunctionCall.partial_args] field. This field is not supported in Gemini API. */
streamFunctionCallArguments?: boolean;
}
/** Function calling mode. */
export declare enum FunctionCallingConfigMode {
/**
* Unspecified function calling mode. This value should not be used.
*/
MODE_UNSPECIFIED = "MODE_UNSPECIFIED",
/**
* Default model behavior, model decides to predict either function calls or natural language response.
*/
AUTO = "AUTO",
/**
* Model is constrained to always predicting function calls only. If "allowed_function_names" are set, the predicted function calls will be limited to any one of "allowed_function_names", else the predicted function calls will be any one of the provided "function_declarations".
*/
ANY = "ANY",
/**
* Model will not predict any function calls. Model behavior is same as when not passing any function declarations.
*/
NONE = "NONE",
/**
* Model is constrained to predict either function calls or natural language response. If "allowed_function_names" are set, the predicted function calls will be limited to any one of "allowed_function_names", else the predicted function calls will be any one of the provided "function_declarations".
*/
VALIDATED = "VALIDATED"
}
/** Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. */
export declare interface FunctionDeclaration {
/** Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. */
description?: string;
/** Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots, colons and dashes, with a maximum length of 64. */
name?: string;
/** Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 */
parameters?: Schema;
/** Optional. Describes the parameters to the function in JSON Schema format. The schema must describe an object where the properties are the parameters to the function. For example: ``` { "type": "object", "properties": { "name": { "type": "string" }, "age": { "type": "integer" } }, "additionalProperties": false, "required": ["name", "age"], "propertyOrdering": ["name", "age"] } ``` This field is mutually exclusive with `parameters`. */
parametersJsonSchema?: unknown;
/** Optional. Describes the output from this function in JSON Schema format. Reflects the Open API 3.03 Response Object. The Schema defines the type used for the response value of the function. */
response?: Schema;
/** Optional. Describes the output from this function in JSON Schema format. The value specified by the schema is the response value of the function. This field is mutually exclusive with `response`. */
responseJsonSchema?: unknown;
/** Optional. Specifies the function Behavior. Currently only supported by the BidiGenerateContent method. This field is not supported in Vertex AI. */
behavior?: Behavior;
}
/** A function response. */
export declare class FunctionResponse {
/** Signals that function call continues, and more responses will be returned, turning the function call into a generator. Is only applicable to NON_BLOCKING function calls (see FunctionDeclaration.behavior for details), ignored otherwise. If false, the default, future responses will not be considered. Is only applicable to NON_BLOCKING function calls, is ignored otherwise. If set to false, future responses will not be considered. It is allowed to return empty `response` with `will_continue=False` to signal that the function call is finished. */
willContinue?: boolean;
/** Specifies how the response should be scheduled in the conversation. Only applicable to NON_BLOCKING function calls, is ignored otherwise. Defaults to WHEN_IDLE. */
scheduling?: FunctionResponseScheduling;
/** List of parts that constitute a function response. Each part may
have a different IANA MIME type. */
parts?: FunctionResponsePart[];
/** Optional. The id of the function call this response is for. Populated by the client to match the corresponding function call `id`. */
id?: string;
/** Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. */
name?: string;
/** Required. The function response in JSON object format. Use "output" key to specify function output and "error" key to specify error details (if any). If "output" and "error" keys are not specified, then whole "response" is treated as function output. */
response?: Record;
}
/** Raw media bytes for function response.
Text should not be sent as raw bytes, use the FunctionResponse.response
field. */
export declare class FunctionResponseBlob {
/** Required. The IANA standard MIME type of the source data. */
mimeType?: string;
/** Required. Inline media bytes.
* @remarks Encoded as base64 string. */
data?: string;
/** Optional. Display name of the blob.
Used to provide a label or filename to distinguish blobs. */
displayName?: string;
}
/** URI based data for function response. */
export declare class FunctionResponseFileData {
/** Required. URI. */
fileUri?: string;
/** Required. The IANA standard MIME type of the source data. */
mimeType?: string;
/** Optional. Display name of the file.
Used to provide a label or filename to distinguish files. */
displayName?: string;
}
/** A datatype containing media that is part of a `FunctionResponse` message.
A `FunctionResponsePart` consists of data which has an associated datatype. A
`FunctionResponsePart` can only contain one of the accepted types in
`FunctionResponsePart.data`.
A `FunctionResponsePart` must have a fixed IANA MIME type identifying the
type and subtype of the media if the `inline_data` field is filled with raw
bytes. */
export declare class FunctionResponsePart {
/** Optional. Inline media bytes. */
inlineData?: FunctionResponseBlob;
/** Optional. URI based data. */
fileData?: FunctionResponseFileData;
}
/** Specifies how the response should be scheduled in the conversation. */
export declare enum FunctionResponseScheduling {
/**
* This value is unused.
*/
SCHEDULING_UNSPECIFIED = "SCHEDULING_UNSPECIFIED",
/**
* Only add the result to the conversation context, do not interrupt or trigger generation.
*/
SILENT = "SILENT",
/**
* Add the result to the conversation context, and prompt to generate output without interrupting ongoing generation.
*/
WHEN_IDLE = "WHEN_IDLE",
/**
* Add the result to the conversation context, interrupt ongoing generation and prompt to generate output.
*/
INTERRUPT = "INTERRUPT"
}
/**
* A function tool result content block.
*/
declare interface FunctionResultContent {
/**
* ID to match the ID from the function call block.
*/
call_id: string;
/**
* The result of the tool call.
*/
result: unknown | Array | string;
type: 'function_result';
/**
* Whether the tool call resulted in an error.
*/
is_error?: boolean;
/**
* The name of the tool that was called.
*/
name?: string;
/**
* A signature hash for backend validation.
*/
signature?: string;
}
/** The Google Cloud Storage location for the input content. This data type is not supported in Gemini API. */
export declare interface GcsSource {
/** Required. Google Cloud Storage URI(-s) to the input file(s). May contain wildcards. For more information on wildcards, see https://cloud.google.com/storage/docs/wildcards. */
uris?: string[];
}
/**
* @license
* Copyright 2025 Google LLC
* SPDX-License-Identifier: Apache-2.0
*/
declare interface GeminiNextGenAPIClientAdapter {
isVertexAI: () => boolean;
getProject: () => string | undefined;
getLocation: () => string | undefined;
getAuthHeaders: () => Headers | Promise;
}
/**
* @license
* Copyright 2025 Google LLC
* SPDX-License-Identifier: Apache-2.0
*/
declare class GeminiNextGenAPIClientError extends Error {
}
/** Input example for preference optimization. This data type is not supported in Gemini API. */
export declare interface GeminiPreferenceExample {
/** List of completions for a given prompt. */
completions?: GeminiPreferenceExampleCompletion[];
/** Multi-turn contents that represents the Prompt. */
contents?: Content[];
}
/** Completion and its preference score. This data type is not supported in Gemini API. */
export declare interface GeminiPreferenceExampleCompletion {
/** Single turn completion for the given prompt. */
completion?: Content;
/** The score for the given completion. */
score?: number;
}
/** Optional model configuration parameters.
For more information, see `Content generation parameters
`_. */
export declare interface GenerateContentConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** Instructions for the model to steer it toward better performance.
For example, "Answer as concisely as possible" or "Don't use technical
terms in your response".
*/
systemInstruction?: ContentUnion;
/** Value that controls the degree of randomness in token selection.
Lower temperatures are good for prompts that require a less open-ended or
creative response, while higher temperatures can lead to more diverse or
creative results.
*/
temperature?: number;
/** Tokens are selected from the most to least probable until the sum
of their probabilities equals this value. Use a lower value for less
random responses and a higher value for more random responses.
*/
topP?: number;
/** For each token selection step, the ``top_k`` tokens with the
highest probabilities are sampled. Then tokens are further filtered based
on ``top_p`` with the final token selected using temperature sampling. Use
a lower number for less random responses and a higher number for more
random responses.
*/
topK?: number;
/** Number of response variations to return.
*/
candidateCount?: number;
/** Maximum number of tokens that can be generated in the response.
*/
maxOutputTokens?: number;
/** List of strings that tells the model to stop generating text if one
of the strings is encountered in the response.
*/
stopSequences?: string[];
/** Whether to return the log probabilities of the tokens that were
chosen by the model at each step.
*/
responseLogprobs?: boolean;
/** Number of top candidate tokens to return the log probabilities for
at each generation step.
*/
logprobs?: number;
/** Positive values penalize tokens that already appear in the
generated text, increasing the probability of generating more diverse
content.
*/
presencePenalty?: number;
/** Positive values penalize tokens that repeatedly appear in the
generated text, increasing the probability of generating more diverse
content.
*/
frequencyPenalty?: number;
/** When ``seed`` is fixed to a specific number, the model makes a best
effort to provide the same response for repeated requests. By default, a
random number is used.
*/
seed?: number;
/** Output response mimetype of the generated candidate text.
Supported mimetype:
- `text/plain`: (default) Text output.
- `application/json`: JSON response in the candidates.
The model needs to be prompted to output the appropriate response type,
otherwise the behavior is undefined.
*/
responseMimeType?: string;
/** The `Schema` object allows the definition of input and output data types.
These types can be objects, but also primitives and arrays.
Represents a select subset of an [OpenAPI 3.0 schema
object](https://spec.openapis.org/oas/v3.0.3#schema).
If set, a compatible response_mime_type must also be set.
Compatible mimetypes: `application/json`: Schema for JSON response.
If `response_schema` doesn't process your schema correctly, try using
`response_json_schema` instead.
*/
responseSchema?: SchemaUnion;
/** Optional. Output schema of the generated response.
This is an alternative to `response_schema` that accepts [JSON
Schema](https://json-schema.org/). If set, `response_schema` must be
omitted, but `response_mime_type` is required. While the full JSON Schema
may be sent, not all features are supported. Specifically, only the
following properties are supported: - `$id` - `$defs` - `$ref` - `$anchor`
- `type` - `format` - `title` - `description` - `enum` (for strings and
numbers) - `items` - `prefixItems` - `minItems` - `maxItems` - `minimum` -
`maximum` - `anyOf` - `oneOf` (interpreted the same as `anyOf`) -
`properties` - `additionalProperties` - `required` The non-standard
`propertyOrdering` property may also be set. Cyclic references are
unrolled to a limited degree and, as such, may only be used within
non-required properties. (Nullable properties are not sufficient.) If
`$ref` is set on a sub-schema, no other properties, except for than those
starting as a `$`, may be set. */
responseJsonSchema?: unknown;
/** Configuration for model router requests.
*/
routingConfig?: GenerationConfigRoutingConfig;
/** Configuration for model selection.
*/
modelSelectionConfig?: ModelSelectionConfig;
/** Safety settings in the request to block unsafe content in the
response.
*/
safetySettings?: SafetySetting[];
/** Code that enables the system to interact with external systems to
perform an action outside of the knowledge and scope of the model.
*/
tools?: ToolListUnion;
/** Associates model output to a specific function call.
*/
toolConfig?: ToolConfig;
/** Labels with user-defined metadata to break down billed charges. */
labels?: Record;
/** Resource name of a context cache that can be used in subsequent
requests.
*/
cachedContent?: string;
/** The requested modalities of the response. Represents the set of
modalities that the model can return.
*/
responseModalities?: string[];
/** If specified, the media resolution specified will be used.
*/
mediaResolution?: MediaResolution;
/** The speech generation configuration.
*/
speechConfig?: SpeechConfigUnion;
/** If enabled, audio timestamp will be included in the request to the
model.
*/
audioTimestamp?: boolean;
/** The configuration for automatic function calling.
*/
automaticFunctionCalling?: AutomaticFunctionCallingConfig;
/** The thinking features configuration.
*/
thinkingConfig?: ThinkingConfig;
/** The image generation configuration.
*/
imageConfig?: ImageConfig;
/** Enables enhanced civic answers. It may not be available for all
models. This field is not supported in Vertex AI.
*/
enableEnhancedCivicAnswers?: boolean;
/** Settings for prompt and response sanitization using the Model Armor
service. If supplied, safety_settings must not be supplied.
*/
modelArmorConfig?: ModelArmorConfig;
}
/** Config for models.generate_content parameters. */
export declare interface GenerateContentParameters {
/** ID of the model to use. For a list of models, see `Google models
`_. */
model: string;
/** Content of the request.
*/
contents: ContentListUnion;
/** Configuration that contains optional model parameters.
*/
config?: GenerateContentConfig;
}
/** Response message for PredictionService.GenerateContent. */
export declare class GenerateContentResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
/** Response variations returned by the model.
*/
candidates?: Candidate[];
/** Timestamp when the request is made to the server.
*/
createTime?: string;
/** The history of automatic function calling.
*/
automaticFunctionCallingHistory?: Content[];
/** Output only. The model version used to generate the response. */
modelVersion?: string;
/** Output only. Content filter results for a prompt sent in the request. Note: Sent only in the first stream chunk. Only happens when no candidates were generated due to content violations. */
promptFeedback?: GenerateContentResponsePromptFeedback;
/** Output only. response_id is used to identify each response. It is the encoding of the event_id. */
responseId?: string;
/** Usage metadata about the response(s). */
usageMetadata?: GenerateContentResponseUsageMetadata;
/**
* Returns the concatenation of all text parts from the first candidate in the response.
*
* @remarks
* If there are multiple candidates in the response, the text from the first
* one will be returned.
* If there are non-text parts in the response, the concatenation of all text
* parts will be returned, and a warning will be logged.
* If there are thought parts in the response, the concatenation of all text
* parts excluding the thought parts will be returned.
*
* @example
* ```ts
* const response = await ai.models.generateContent({
* model: 'gemini-2.0-flash',
* contents:
* 'Why is the sky blue?',
* });
*
* console.debug(response.text);
* ```
*/
get text(): string | undefined;
/**
* Returns the concatenation of all inline data parts from the first candidate
* in the response.
*
* @remarks
* If there are multiple candidates in the response, the inline data from the
* first one will be returned. If there are non-inline data parts in the
* response, the concatenation of all inline data parts will be returned, and
* a warning will be logged.
*/
get data(): string | undefined;
/**
* Returns the function calls from the first candidate in the response.
*
* @remarks
* If there are multiple candidates in the response, the function calls from
* the first one will be returned.
* If there are no function calls in the response, undefined will be returned.
*
* @example
* ```ts
* const controlLightFunctionDeclaration: FunctionDeclaration = {
* name: 'controlLight',
* parameters: {
* type: Type.OBJECT,
* description: 'Set the brightness and color temperature of a room light.',
* properties: {
* brightness: {
* type: Type.NUMBER,
* description:
* 'Light level from 0 to 100. Zero is off and 100 is full brightness.',
* },
* colorTemperature: {
* type: Type.STRING,
* description:
* 'Color temperature of the light fixture which can be `daylight`, `cool` or `warm`.',
* },
* },
* required: ['brightness', 'colorTemperature'],
* };
* const response = await ai.models.generateContent({
* model: 'gemini-2.0-flash',
* contents: 'Dim the lights so the room feels cozy and warm.',
* config: {
* tools: [{functionDeclarations: [controlLightFunctionDeclaration]}],
* toolConfig: {
* functionCallingConfig: {
* mode: FunctionCallingConfigMode.ANY,
* allowedFunctionNames: ['controlLight'],
* },
* },
* },
* });
* console.debug(JSON.stringify(response.functionCalls));
* ```
*/
get functionCalls(): FunctionCall[] | undefined;
/**
* Returns the first executable code from the first candidate in the response.
*
* @remarks
* If there are multiple candidates in the response, the executable code from
* the first one will be returned.
* If there are no executable code in the response, undefined will be
* returned.
*
* @example
* ```ts
* const response = await ai.models.generateContent({
* model: 'gemini-2.0-flash',
* contents:
* 'What is the sum of the first 50 prime numbers? Generate and run code for the calculation, and make sure you get all 50.'
* config: {
* tools: [{codeExecution: {}}],
* },
* });
*
* console.debug(response.executableCode);
* ```
*/
get executableCode(): string | undefined;
/**
* Returns the first code execution result from the first candidate in the response.
*
* @remarks
* If there are multiple candidates in the response, the code execution result from
* the first one will be returned.
* If there are no code execution result in the response, undefined will be returned.
*
* @example
* ```ts
* const response = await ai.models.generateContent({
* model: 'gemini-2.0-flash',
* contents:
* 'What is the sum of the first 50 prime numbers? Generate and run code for the calculation, and make sure you get all 50.'
* config: {
* tools: [{codeExecution: {}}],
* },
* });
*
* console.debug(response.codeExecutionResult);
* ```
*/
get codeExecutionResult(): string | undefined;
}
/** Content filter results for a prompt sent in the request. Note: This is sent only in the first stream chunk and only if no candidates were generated due to content violations. */
export declare class GenerateContentResponsePromptFeedback {
/** Output only. The reason why the prompt was blocked. */
blockReason?: BlockedReason;
/** Output only. A readable message that explains the reason why the prompt was blocked. This field is not supported in Gemini API. */
blockReasonMessage?: string;
/** Output only. A list of safety ratings for the prompt. There is one rating per category. */
safetyRatings?: SafetyRating[];
}
/** Usage metadata about the content generation request and response. This message provides a detailed breakdown of token usage and other relevant metrics. This data type is not supported in Gemini API. */
export declare class GenerateContentResponseUsageMetadata {
/** Output only. A detailed breakdown of the token count for each modality in the cached content. */
cacheTokensDetails?: ModalityTokenCount[];
/** Output only. The number of tokens in the cached content that was used for this request. */
cachedContentTokenCount?: number;
/** The total number of tokens in the generated candidates. */
candidatesTokenCount?: number;
/** Output only. A detailed breakdown of the token count for each modality in the generated candidates. */
candidatesTokensDetails?: ModalityTokenCount[];
/** The total number of tokens in the prompt. This includes any text, images, or other media provided in the request. When `cached_content` is set, this also includes the number of tokens in the cached content. */
promptTokenCount?: number;
/** Output only. A detailed breakdown of the token count for each modality in the prompt. */
promptTokensDetails?: ModalityTokenCount[];
/** Output only. The number of tokens that were part of the model's generated "thoughts" output, if applicable. */
thoughtsTokenCount?: number;
/** Output only. The number of tokens in the results from tool executions, which are provided back to the model as input, if applicable. */
toolUsePromptTokenCount?: number;
/** Output only. A detailed breakdown by modality of the token counts from the results of tool executions, which are provided back to the model as input. */
toolUsePromptTokensDetails?: ModalityTokenCount[];
/** The total number of tokens for the entire request. This is the sum of `prompt_token_count`, `candidates_token_count`, `tool_use_prompt_token_count`, and `thoughts_token_count`. */
totalTokenCount?: number;
/** Output only. The traffic type for this request. */
trafficType?: TrafficType;
}
/** An output image. */
export declare interface GeneratedImage {
/** The output image data. */
image?: Image_2;
/** Responsible AI filter reason if the image is filtered out of the
response. */
raiFilteredReason?: string;
/** Safety attributes of the image. Lists of RAI categories and their
scores of each content. */
safetyAttributes?: SafetyAttributes;
/** The rewritten prompt used for the image generation if the prompt
enhancer is enabled. */
enhancedPrompt?: string;
}
/** A generated image mask. */
export declare interface GeneratedImageMask {
/** The generated image mask. */
mask?: Image_2;
/** The detected entities on the segmented area. */
labels?: EntityLabel[];
}
/** A generated video. */
export declare interface GeneratedVideo {
/** The output video */
video?: Video;
}
/** The config for generating an images. */
export declare interface GenerateImagesConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** Cloud Storage URI used to store the generated images. */
outputGcsUri?: string;
/** Description of what to discourage in the generated images. */
negativePrompt?: string;
/** Number of images to generate. */
numberOfImages?: number;
/** Aspect ratio of the generated images. Supported values are
"1:1", "3:4", "4:3", "9:16", and "16:9". */
aspectRatio?: string;
/** Controls how much the model adheres to the text prompt. Large
values increase output and prompt alignment, but may compromise image
quality. */
guidanceScale?: number;
/** Random seed for image generation. This is not available when
``add_watermark`` is set to true. */
seed?: number;
/** Filter level for safety filtering. */
safetyFilterLevel?: SafetyFilterLevel;
/** Allows generation of people by the model. */
personGeneration?: PersonGeneration;
/** Whether to report the safety scores of each generated image and
the positive prompt in the response. */
includeSafetyAttributes?: boolean;
/** Whether to include the Responsible AI filter reason if the image
is filtered out of the response. */
includeRaiReason?: boolean;
/** Language of the text in the prompt. */
language?: ImagePromptLanguage;
/** MIME type of the generated image. */
outputMimeType?: string;
/** Compression quality of the generated image (for ``image/jpeg``
only). */
outputCompressionQuality?: number;
/** Whether to add a watermark to the generated images. */
addWatermark?: boolean;
/** User specified labels to track billing usage. */
labels?: Record;
/** The size of the largest dimension of the generated image.
Supported sizes are 1K and 2K (not supported for Imagen 3 models). */
imageSize?: string;
/** Whether to use the prompt rewriting logic. */
enhancePrompt?: boolean;
}
/** The parameters for generating images. */
export declare interface GenerateImagesParameters {
/** ID of the model to use. For a list of models, see `Google models
`_. */
model: string;
/** Text prompt that typically describes the images to output.
*/
prompt: string;
/** Configuration for generating images.
*/
config?: GenerateImagesConfig;
}
/** The output images response. */
export declare class GenerateImagesResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
/** List of generated images. */
generatedImages?: GeneratedImage[];
/** Safety attributes of the positive prompt. Only populated if
``include_safety_attributes`` is set to True. */
positivePromptSafetyAttributes?: SafetyAttributes;
}
/** Configuration for generating videos. */
export declare interface GenerateVideosConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** Number of output videos. */
numberOfVideos?: number;
/** The gcs bucket where to save the generated videos. */
outputGcsUri?: string;
/** Frames per second for video generation. */
fps?: number;
/** Duration of the clip for video generation in seconds. */
durationSeconds?: number;
/** The RNG seed. If RNG seed is exactly same for each request with
unchanged inputs, the prediction results will be consistent. Otherwise,
a random RNG seed will be used each time to produce a different
result. */
seed?: number;
/** The aspect ratio for the generated video. 16:9 (landscape) and
9:16 (portrait) are supported. */
aspectRatio?: string;
/** The resolution for the generated video. 720p and 1080p are
supported. */
resolution?: string;
/** Whether allow to generate person videos, and restrict to specific
ages. Supported values are: dont_allow, allow_adult. */
personGeneration?: string;
/** The pubsub topic where to publish the video generation
progress. */
pubsubTopic?: string;
/** Explicitly state what should not be included in the generated
videos. */
negativePrompt?: string;
/** Whether to use the prompt rewriting logic. */
enhancePrompt?: boolean;
/** Whether to generate audio along with the video. */
generateAudio?: boolean;
/** Image to use as the last frame of generated videos.
Only supported for image to video use cases. */
lastFrame?: Image_2;
/** The images to use as the references to generate the videos.
If this field is provided, the text prompt field must also be provided.
The image, video, or last_frame field are not supported. Each image must
be associated with a type. Veo 2 supports up to 3 asset images *or* 1
style image. */
referenceImages?: VideoGenerationReferenceImage[];
/** The mask to use for generating videos. */
mask?: VideoGenerationMask;
/** Compression quality of the generated videos. */
compressionQuality?: VideoCompressionQuality;
}
/** A video generation operation. */
export declare class GenerateVideosOperation implements Operation {
/** The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. */
name?: string;
/** Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. */
metadata?: Record;
/** If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. */
done?: boolean;
/** The error result of the operation in case of failure or cancellation. */
error?: Record;
/** The generated videos. */
response?: GenerateVideosResponse;
/**
* Instantiates an Operation of the same type as the one being called with the fields set from the API response.
* @internal
*/
_fromAPIResponse({ apiResponse, _isVertexAI, }: OperationFromAPIResponseParameters): Operation;
/** The full HTTP response. */
sdkHttpResponse?: HttpResponse;
}
/** Class that represents the parameters for generating videos. */
export declare interface GenerateVideosParameters {
/** ID of the model to use. For a list of models, see `Google models
`_. */
model: string;
/** The text prompt for generating the videos.
Optional if image or video is provided. */
prompt?: string;
/** The input image for generating the videos.
Optional if prompt is provided. Not allowed if video is provided. */
image?: Image_2;
/** The input video for video extension use cases.
Optional if prompt is provided. Not allowed if image is provided. */
video?: Video;
/** A set of source input(s) for video generation. */
source?: GenerateVideosSource;
/** Configuration for generating videos. */
config?: GenerateVideosConfig;
}
/** Response with generated videos. */
export declare class GenerateVideosResponse {
/** List of the generated videos */
generatedVideos?: GeneratedVideo[];
/** Returns if any videos were filtered due to RAI policies. */
raiMediaFilteredCount?: number;
/** Returns rai failure reasons if any. */
raiMediaFilteredReasons?: string[];
}
/** A set of source input(s) for video generation. */
export declare interface GenerateVideosSource {
/** The text prompt for generating the videos.
Optional if image or video is provided. */
prompt?: string;
/** The input image for generating the videos.
Optional if prompt is provided. Not allowed if video is provided. */
image?: Image_2;
/** The input video for video extension use cases.
Optional if prompt is provided. Not allowed if image is provided. */
video?: Video;
}
/** Generation config. */
export declare interface GenerationConfig {
/** Optional. Config for model selection. */
modelSelectionConfig?: ModelSelectionConfig;
/** Output schema of the generated response. This is an alternative to
`response_schema` that accepts [JSON Schema](https://json-schema.org/).
*/
responseJsonSchema?: unknown;
/** Optional. If enabled, audio timestamps will be included in the request to the model. This can be useful for synchronizing audio with other modalities in the response. This field is not supported in Gemini API. */
audioTimestamp?: boolean;
/** Optional. The number of candidate responses to generate. A higher `candidate_count` can provide more options to choose from, but it also consumes more resources. This can be useful for generating a variety of responses and selecting the best one. */
candidateCount?: number;
/** Optional. If enabled, the model will detect emotions and adapt its responses accordingly. For example, if the model detects that the user is frustrated, it may provide a more empathetic response. This field is not supported in Gemini API. */
enableAffectiveDialog?: boolean;
/** Optional. Penalizes tokens based on their frequency in the generated text. A positive value helps to reduce the repetition of words and phrases. Valid values can range from [-2.0, 2.0]. */
frequencyPenalty?: number;
/** Optional. The number of top log probabilities to return for each token. This can be used to see which other tokens were considered likely candidates for a given position. A higher value will return more options, but it will also increase the size of the response. */
logprobs?: number;
/** Optional. The maximum number of tokens to generate in the response. A token is approximately four characters. The default value varies by model. This parameter can be used to control the length of the generated text and prevent overly long responses. */
maxOutputTokens?: number;
/** Optional. The token resolution at which input media content is sampled. This is used to control the trade-off between the quality of the response and the number of tokens used to represent the media. A higher resolution allows the model to perceive more detail, which can lead to a more nuanced response, but it will also use more tokens. This does not affect the image dimensions sent to the model. */
mediaResolution?: MediaResolution;
/** Optional. Penalizes tokens that have already appeared in the generated text. A positive value encourages the model to generate more diverse and less repetitive text. Valid values can range from [-2.0, 2.0]. */
presencePenalty?: number;
/** Optional. If set to true, the log probabilities of the output tokens are returned. Log probabilities are the logarithm of the probability of a token appearing in the output. A higher log probability means the token is more likely to be generated. This can be useful for analyzing the model's confidence in its own output and for debugging. */
responseLogprobs?: boolean;
/** Optional. The IANA standard MIME type of the response. The model will generate output that conforms to this MIME type. Supported values include 'text/plain' (default) and 'application/json'. The model needs to be prompted to output the appropriate response type, otherwise the behavior is undefined. This is a preview feature. */
responseMimeType?: string;
/** Optional. The modalities of the response. The model will generate a response that includes all the specified modalities. For example, if this is set to `[TEXT, IMAGE]`, the response will include both text and an image. */
responseModalities?: Modality[];
/** Optional. Lets you to specify a schema for the model's response, ensuring that the output conforms to a particular structure. This is useful for generating structured data such as JSON. The schema is a subset of the [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema) object. When this field is set, you must also set the `response_mime_type` to `application/json`. */
responseSchema?: Schema;
/** Optional. Routing configuration. This field is not supported in Gemini API. */
routingConfig?: GenerationConfigRoutingConfig;
/** Optional. A seed for the random number generator. By setting a seed, you can make the model's output mostly deterministic. For a given prompt and parameters (like temperature, top_p, etc.), the model will produce the same response every time. However, it's not a guaranteed absolute deterministic behavior. This is different from parameters like `temperature`, which control the *level* of randomness. `seed` ensures that the "random" choices the model makes are the same on every run, making it essential for testing and ensuring reproducible results. */
seed?: number;
/** Optional. The speech generation config. */
speechConfig?: SpeechConfig;
/** Optional. A list of character sequences that will stop the model from generating further tokens. If a stop sequence is generated, the output will end at that point. This is useful for controlling the length and structure of the output. For example, you can use ["\n", "###"] to stop generation at a new line or a specific marker. */
stopSequences?: string[];
/** Optional. Controls the randomness of the output. A higher temperature results in more creative and diverse responses, while a lower temperature makes the output more predictable and focused. The valid range is (0.0, 2.0]. */
temperature?: number;
/** Optional. Configuration for thinking features. An error will be returned if this field is set for models that don't support thinking. */
thinkingConfig?: ThinkingConfig;
/** Optional. Specifies the top-k sampling threshold. The model considers only the top k most probable tokens for the next token. This can be useful for generating more coherent and less random text. For example, a `top_k` of 40 means the model will choose the next word from the 40 most likely words. */
topK?: number;
/** Optional. Specifies the nucleus sampling threshold. The model considers only the smallest set of tokens whose cumulative probability is at least `top_p`. This helps generate more diverse and less repetitive responses. For example, a `top_p` of 0.9 means the model considers tokens until the cumulative probability of the tokens to select from reaches 0.9. It's recommended to adjust either temperature or `top_p`, but not both. */
topP?: number;
/** Optional. Enables enhanced civic answers. It may not be available for all models. This field is not supported in Vertex AI. */
enableEnhancedCivicAnswers?: boolean;
}
/**
* Configuration parameters for model interactions.
*/
declare interface GenerationConfig_2 {
/**
* Configuration for image interaction.
*/
image_config?: ImageConfig_2;
/**
* The maximum number of tokens to include in the response.
*/
max_output_tokens?: number;
/**
* Seed used in decoding for reproducibility.
*/
seed?: number;
/**
* Configuration for speech interaction.
*/
speech_config?: Array;
/**
* A list of character sequences that will stop output interaction.
*/
stop_sequences?: Array;
/**
* Controls the randomness of the output.
*/
temperature?: number;
/**
* The level of thought tokens that the model should generate.
*/
thinking_level?: ThinkingLevel_2;
/**
* Whether to include thought summaries in the response.
*/
thinking_summaries?: 'auto' | 'none';
/**
* The tool choice for the interaction.
*/
tool_choice?: ToolChoiceType | ToolChoiceConfig;
/**
* The maximum cumulative probability of tokens to consider when sampling.
*/
top_p?: number;
}
/** The configuration for routing the request to a specific model. This can be used to control which model is used for the generation, either automatically or by specifying a model name. This data type is not supported in Gemini API. */
export declare interface GenerationConfigRoutingConfig {
/** In this mode, the model is selected automatically based on the content of the request. */
autoMode?: GenerationConfigRoutingConfigAutoRoutingMode;
/** In this mode, the model is specified manually. */
manualMode?: GenerationConfigRoutingConfigManualRoutingMode;
}
/** The configuration for automated routing. When automated routing is specified, the routing will be determined by the pretrained routing model and customer provided model routing preference. This data type is not supported in Gemini API. */
export declare interface GenerationConfigRoutingConfigAutoRoutingMode {
/** The model routing preference. */
modelRoutingPreference?: 'UNKNOWN' | 'PRIORITIZE_QUALITY' | 'BALANCED' | 'PRIORITIZE_COST';
}
/** The configuration for manual routing. When manual routing is specified, the model will be selected based on the model name provided. This data type is not supported in Gemini API. */
export declare interface GenerationConfigRoutingConfigManualRoutingMode {
/** The name of the model to use. Only public LLM models are accepted. */
modelName?: string;
}
/**
* Config for thinking feature.
*
* @deprecated This interface will be deprecated. Please use `ThinkingConfig` instead.
*/
export declare interface GenerationConfigThinkingConfig extends ThinkingConfig {
}
/** Optional parameters. */
export declare interface GetBatchJobConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Config for batches.get parameters. */
export declare interface GetBatchJobParameters {
/** A fully-qualified BatchJob resource name or ID.
Example: "projects/.../locations/.../batchPredictionJobs/456"
or "456" when project and location are initialized in the client.
*/
name: string;
/** Optional parameters for the request. */
config?: GetBatchJobConfig;
}
/** Optional parameters for caches.get method. */
export declare interface GetCachedContentConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Parameters for caches.get method. */
export declare interface GetCachedContentParameters {
/** The server-generated resource name of the cached content.
*/
name: string;
/** Optional parameters for the request.
*/
config?: GetCachedContentConfig;
}
/** Optional Config. */
export declare interface GetDocumentConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Parameters for documents.get. */
export declare interface GetDocumentParameters {
/** The resource name of the Document.
Example: fileSearchStores/file-search-store-foo/documents/documents-bar */
name: string;
/** Optional parameters for the request. */
config?: GetDocumentConfig;
}
/** Used to override the default configuration. */
export declare interface GetFileConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Generates the parameters for the get method. */
export declare interface GetFileParameters {
/** The name identifier for the file to retrieve. */
name: string;
/** Used to override the default configuration. */
config?: GetFileConfig;
}
/** Optional parameters for getting a FileSearchStore. */
export declare interface GetFileSearchStoreConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Config for file_search_stores.get parameters. */
export declare interface GetFileSearchStoreParameters {
/** The resource name of the FileSearchStore. Example: `fileSearchStores/my-file-search-store-123` */
name: string;
/** Optional parameters for the request. */
config?: GetFileSearchStoreConfig;
}
/** Optional parameters for models.get method. */
export declare interface GetModelConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
export declare interface GetModelParameters {
model: string;
/** Optional parameters for the request. */
config?: GetModelConfig;
}
export declare interface GetOperationConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Parameters for the GET method. */
export declare interface GetOperationParameters {
/** The server-assigned name for the operation. */
operationName: string;
/** Used to override the default configuration. */
config?: GetOperationConfig;
}
/** Optional parameters for tunings.get method. */
export declare interface GetTuningJobConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Parameters for the get method. */
export declare interface GetTuningJobParameters {
name: string;
/** Optional parameters for the request. */
config?: GetTuningJobConfig;
}
/**
* The Google GenAI SDK.
*
* @remarks
* Provides access to the GenAI features through either the {@link
* https://cloud.google.com/vertex-ai/docs/reference/rest | Gemini API} or
* the {@link https://cloud.google.com/vertex-ai/docs/reference/rest | Vertex AI
* API}.
*
* The {@link GoogleGenAIOptions.vertexai} value determines which of the API
* services to use.
*
* When using the Gemini API, a {@link GoogleGenAIOptions.apiKey} must also be
* set. When using Vertex AI, both {@link GoogleGenAIOptions.project} and {@link
* GoogleGenAIOptions.location} must be set, or a {@link
* GoogleGenAIOptions.apiKey} must be set when using Express Mode.
*
* Explicitly passed in values in {@link GoogleGenAIOptions} will always take
* precedence over environment variables. If both project/location and api_key
* exist in the environment variables, the project/location will be used.
*
* @example
* Initializing the SDK for using the Gemini API:
* ```ts
* import {GoogleGenAI} from '@google/genai';
* const ai = new GoogleGenAI({apiKey: 'GEMINI_API_KEY'});
* ```
*
* @example
* Initializing the SDK for using the Vertex AI API:
* ```ts
* import {GoogleGenAI} from '@google/genai';
* const ai = new GoogleGenAI({
* vertexai: true,
* project: 'PROJECT_ID',
* location: 'PROJECT_LOCATION'
* });
* ```
*
*/
export declare class GoogleGenAI {
protected readonly apiClient: ApiClient;
private readonly apiKey?;
readonly vertexai: boolean;
private readonly googleAuthOptions?;
private readonly project?;
private readonly location?;
private readonly apiVersion?;
private readonly httpOptions?;
readonly models: Models;
readonly live: Live;
readonly batches: Batches;
readonly chats: Chats;
readonly caches: Caches;
readonly files: Files;
readonly operations: Operations;
readonly authTokens: Tokens;
readonly tunings: Tunings;
readonly fileSearchStores: FileSearchStores;
private _interactions;
get interactions(): Interactions;
constructor(options: GoogleGenAIOptions);
}
/**
* Google Gen AI SDK's configuration options.
*
* See {@link GoogleGenAI} for usage samples.
*/
export declare interface GoogleGenAIOptions {
/**
* Optional. Determines whether to use the Vertex AI or the Gemini API.
*
* @remarks
* When true, the {@link https://cloud.google.com/vertex-ai/docs/reference/rest | Vertex AI API} will used.
* When false, the {@link https://ai.google.dev/api | Gemini API} will be used.
*
* If unset, default SDK behavior is to use the Gemini API service.
*/
vertexai?: boolean;
/**
* Optional. The Google Cloud project ID for Vertex AI clients.
*
* Find your project ID: https://cloud.google.com/resource-manager/docs/creating-managing-projects#identifying_projects
*
* @remarks
* Only supported on Node runtimes, ignored on browser runtimes.
*/
project?: string;
/**
* Optional. The Google Cloud project {@link https://cloud.google.com/vertex-ai/generative-ai/docs/learn/locations | location} for Vertex AI clients.
*
* @remarks
* Only supported on Node runtimes, ignored on browser runtimes.
*
*/
location?: string;
/**
* The API Key, required for Gemini API clients.
*
* @remarks
* Required on browser runtimes.
*/
apiKey?: string;
/**
* Optional. The API version to use.
*
* @remarks
* If unset, the default API version will be used.
*/
apiVersion?: string;
/**
* Optional. Authentication options defined by the by google-auth-library for Vertex AI clients.
*
* @remarks
* @see {@link https://github.com/googleapis/google-auth-library-nodejs/blob/v9.15.0/src/auth/googleauth.ts | GoogleAuthOptions interface in google-auth-library-nodejs}.
*
* Only supported on Node runtimes, ignored on browser runtimes.
*
*/
googleAuthOptions?: GoogleAuthOptions;
/**
* Optional. A set of customizable configuration for HTTP requests.
*/
httpOptions?: HttpOptions;
}
/** Tool to retrieve knowledge from Google Maps. */
export declare interface GoogleMaps {
/** The authentication config to access the API. Only API key is supported. This field is not supported in Gemini API. */
authConfig?: AuthConfig;
/** Optional. Whether to return a widget context token in the GroundingMetadata of the response. Developers can use the widget context token to render a Google Maps widget with geospatial context related to the places that the model references in the response. */
enableWidget?: boolean;
}
/** The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). This data type is not supported in Gemini API. */
export declare interface GoogleRpcStatus {
/** The status code, which should be an enum value of google.rpc.Code. */
code?: number;
/** A list of messages that carry the error details. There is a common set of message types for APIs to use. */
details?: Record[];
/** A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. */
message?: string;
}
/** Tool to support web search. */
export declare interface GoogleSearch {
/** Different types of search that can be enabled on the GoogleSearch tool. */
searchTypes?: SearchTypes;
/** Optional. Sites with confidence level chosen & above this value will be blocked from the search results. This field is not supported in Gemini API. */
blockingConfidence?: PhishBlockThreshold;
/** Optional. List of domains to be excluded from the search results. The default limit is 2000 domains. Example: ["amazon.com", "facebook.com"]. This field is not supported in Gemini API. */
excludeDomains?: string[];
/** Optional. Filter search results to a specific time range. If customers set a start time, they must set an end time (and vice versa). This field is not supported in Vertex AI. */
timeRangeFilter?: Interval;
}
/**
* The arguments to pass to Google Search.
*/
declare interface GoogleSearchCallArguments {
/**
* Web search queries for the following-up web search.
*/
queries?: Array;
}
/**
* Google Search content.
*/
declare interface GoogleSearchCallContent {
/**
* A unique ID for this specific tool call.
*/
id: string;
/**
* The arguments to pass to Google Search.
*/
arguments: GoogleSearchCallArguments;
type: 'google_search_call';
/**
* The type of search grounding enabled.
*/
search_type?: 'web_search' | 'image_search';
}
/**
* The result of the Google Search.
*/
declare interface GoogleSearchResult {
/**
* Web content snippet that can be embedded in a web page or an app webview.
*/
rendered_content?: string;
/**
* Title of the search result.
*/
title?: string;
/**
* URI reference of the search result.
*/
url?: string;
}
/**
* Google Search result content.
*/
declare interface GoogleSearchResultContent {
/**
* ID to match the ID from the google search call block.
*/
call_id: string;
/**
* The results of the Google Search.
*/
result: Array;
type: 'google_search_result';
/**
* Whether the Google Search resulted in an error.
*/
is_error?: boolean;
/**
* A signature hash for backend validation.
*/
signature?: string;
}
/** Tool to retrieve public web data for grounding, powered by Google. */
export declare interface GoogleSearchRetrieval {
/** Specifies the dynamic retrieval configuration for the given source. */
dynamicRetrievalConfig?: DynamicRetrievalConfig;
}
/** Represents a whole or partial calendar date, such as a birthday. The time of day and time zone are either specified elsewhere or are insignificant. The date is relative to the Gregorian Calendar. This can represent one of the following: * A full date, with non-zero year, month, and day values. * A month and day, with a zero year (for example, an anniversary). * A year on its own, with a zero month and a zero day. * A year and month, with a zero day (for example, a credit card expiration date). Related types: * google.type.TimeOfDay * google.type.DateTime * google.protobuf.Timestamp. This data type is not supported in Gemini API. */
export declare interface GoogleTypeDate {
/** Day of a month. Must be from 1 to 31 and valid for the year and month, or 0 to specify a year by itself or a year and month where the day isn't significant. */
day?: number;
/** Month of a year. Must be from 1 to 12, or 0 to specify a year without a month and day. */
month?: number;
/** Year of the date. Must be from 1 to 9999, or 0 to specify a date without a year. */
year?: number;
}
/** A piece of evidence that supports a claim made by the model.
This is used to show a citation for a claim made by the model. When grounding
is enabled, the model returns a `GroundingChunk` that contains a reference to
the source of the information. */
export declare interface GroundingChunk {
/** A grounding chunk from an image search result. See the `Image`
message for details.
*/
image?: GroundingChunkImage;
/** A `Maps` chunk is a piece of evidence that comes from Google Maps.
It contains information about a place, such as its name, address, and
reviews. This is used to provide the user with rich, location-based
information. */
maps?: GroundingChunkMaps;
/** A grounding chunk from a data source retrieved by a retrieval tool, such as Vertex AI Search. See the `RetrievedContext` message for details. This field is not supported in Gemini API. */
retrievedContext?: GroundingChunkRetrievedContext;
/** A grounding chunk from a web page, typically from Google Search. See the `Web` message for details. */
web?: GroundingChunkWeb;
}
/** A piece of evidence that comes from an image search result.
It contains the URI of the image search result and the URI of the image.
This is used to provide the user with a link to the source of the
information. */
export declare interface GroundingChunkImage {
/** The URI of the image search result page. */
sourceUri?: string;
/** The URI of the image. */
imageUri?: string;
/** The title of the image search result page. */
title?: string;
/** The domain of the image search result page. */
domain?: string;
}
/** A `Maps` chunk is a piece of evidence that comes from Google Maps.
It contains information about a place, such as its name, address, and reviews.
This is used to provide the user with rich, location-based information. */
export declare interface GroundingChunkMaps {
/** The sources that were used to generate the place answer.
This includes review snippets and photos that were used to generate the
answer, as well as URIs to flag content. */
placeAnswerSources?: GroundingChunkMapsPlaceAnswerSources;
/** This Place's resource name, in `places/{place_id}` format.
This can be used to look up the place in the Google Maps API. */
placeId?: string;
/** The text of the place answer. */
text?: string;
/** The title of the place. */
title?: string;
/** The URI of the place. */
uri?: string;
}
/** The sources that were used to generate the place answer.
This includes review snippets and photos that were used to generate the
answer, as well as URIs to flag content. */
export declare interface GroundingChunkMapsPlaceAnswerSources {
/** Snippets of reviews that were used to generate the answer. */
reviewSnippet?: GroundingChunkMapsPlaceAnswerSourcesReviewSnippet[];
/** A link where users can flag a problem with the generated answer. */
flagContentUri?: string;
/** Snippets of reviews that were used to generate the answer. */
reviewSnippets?: GroundingChunkMapsPlaceAnswerSourcesReviewSnippet[];
}
/** Author attribution for a photo or review. */
export declare interface GroundingChunkMapsPlaceAnswerSourcesAuthorAttribution {
/** Name of the author of the Photo or Review. */
displayName?: string;
/** Profile photo URI of the author of the Photo or Review. */
photoUri?: string;
/** URI of the author of the Photo or Review. */
uri?: string;
}
/** Encapsulates a review snippet. */
export declare interface GroundingChunkMapsPlaceAnswerSourcesReviewSnippet {
/** This review's author. */
authorAttribution?: GroundingChunkMapsPlaceAnswerSourcesAuthorAttribution;
/** A link where users can flag a problem with the review. */
flagContentUri?: string;
/** A link to show the review on Google Maps. */
googleMapsUri?: string;
/** A string of formatted recent time, expressing the review time relative to the current time in a form appropriate for the language and country. */
relativePublishTimeDescription?: string;
/** A reference representing this place review which may be used to look up this place review again. */
review?: string;
/** Id of the review referencing the place. */
reviewId?: string;
/** Title of the review. */
title?: string;
}
/** Context retrieved from a data source to ground the model's response. This is used when a retrieval tool fetches information from a user-provided corpus or a public dataset. This data type is not supported in Gemini API. */
export declare interface GroundingChunkRetrievedContext {
/** Output only. The full resource name of the referenced Vertex AI Search document. This is used to identify the specific document that was retrieved. The format is `projects/{project}/locations/{location}/collections/{collection}/dataStores/{data_store}/branches/{branch}/documents/{document}`. */
documentName?: string;
/** Additional context for a Retrieval-Augmented Generation (RAG) retrieval result. This is populated only when the RAG retrieval tool is used. */
ragChunk?: RagChunk;
/** The content of the retrieved data source. */
text?: string;
/** The title of the retrieved data source. */
title?: string;
/** The URI of the retrieved data source. */
uri?: string;
}
/** A `Web` chunk is a piece of evidence that comes from a web page. It contains the URI of the web page, the title of the page, and the domain of the page. This is used to provide the user with a link to the source of the information. */
export declare interface GroundingChunkWeb {
/** The domain of the web page that contains the evidence. This can be used to filter out low-quality sources. This field is not supported in Gemini API. */
domain?: string;
/** The title of the web page that contains the evidence. */
title?: string;
/** The URI of the web page that contains the evidence. */
uri?: string;
}
/** Information for various kinds of grounding. */
export declare interface GroundingMetadata {
/** Optional. The image search queries that were used to generate the
content. This field is populated only when the grounding source is Google
Search with the Image Search search_type enabled.
*/
imageSearchQueries?: string[];
/** A list of supporting references retrieved from the grounding
source. This field is populated when the grounding source is Google
Search, Vertex AI Search, or Google Maps.
*/
groundingChunks?: GroundingChunk[];
/** List of grounding support. */
groundingSupports?: GroundingSupport[];
/** Metadata related to retrieval in the grounding flow. */
retrievalMetadata?: RetrievalMetadata;
/** Optional. Google search entry for the following-up web
searches. */
searchEntryPoint?: SearchEntryPoint;
/** Web search queries for the following-up web search. */
webSearchQueries?: string[];
/** Optional. Output only. A token that can be used to render a Google Maps widget with the contextual data. This field is populated only when the grounding source is Google Maps. This field is not supported in Gemini API. */
googleMapsWidgetContextToken?: string;
/** Optional. The queries that were executed by the retrieval tools. This field is populated only when the grounding source is a retrieval tool, such as Vertex AI Search. This field is not supported in Gemini API. */
retrievalQueries?: string[];
/** Optional. Output only. A list of URIs that can be used to flag a place or review for inappropriate content. This field is populated only when the grounding source is Google Maps. This field is not supported in Gemini API. */
sourceFlaggingUris?: GroundingMetadataSourceFlaggingUri[];
}
/** A URI that can be used to flag a place or review for inappropriate content. This is populated only when the grounding source is Google Maps. This data type is not supported in Gemini API. */
export declare interface GroundingMetadataSourceFlaggingUri {
/** The URI that can be used to flag the content. */
flagContentUri?: string;
/** The ID of the place or review. */
sourceId?: string;
}
/** Grounding support. */
export declare interface GroundingSupport {
/** Confidence score of the support references.
Ranges from 0 to 1. 1 is the most confident. This list must have the
same size as the grounding_chunk_indices. */
confidenceScores?: number[];
/** A list of indices (into 'grounding_chunk') specifying the
citations associated with the claim. For instance [1,3,4] means that
grounding_chunk[1], grounding_chunk[3], grounding_chunk[4] are the
retrieved content attributed to the claim. */
groundingChunkIndices?: number[];
/** Segment of the content this support belongs to. */
segment?: Segment;
}
/** The method for blocking content. If not specified, the default behavior is to use the probability score. This enum is not supported in Gemini API. */
export declare enum HarmBlockMethod {
/**
* The harm block method is unspecified.
*/
HARM_BLOCK_METHOD_UNSPECIFIED = "HARM_BLOCK_METHOD_UNSPECIFIED",
/**
* The harm block method uses both probability and severity scores.
*/
SEVERITY = "SEVERITY",
/**
* The harm block method uses the probability score.
*/
PROBABILITY = "PROBABILITY"
}
/** The threshold for blocking content. If the harm probability exceeds this threshold, the content will be blocked. */
export declare enum HarmBlockThreshold {
/**
* The harm block threshold is unspecified.
*/
HARM_BLOCK_THRESHOLD_UNSPECIFIED = "HARM_BLOCK_THRESHOLD_UNSPECIFIED",
/**
* Block content with a low harm probability or higher.
*/
BLOCK_LOW_AND_ABOVE = "BLOCK_LOW_AND_ABOVE",
/**
* Block content with a medium harm probability or higher.
*/
BLOCK_MEDIUM_AND_ABOVE = "BLOCK_MEDIUM_AND_ABOVE",
/**
* Block content with a high harm probability.
*/
BLOCK_ONLY_HIGH = "BLOCK_ONLY_HIGH",
/**
* Do not block any content, regardless of its harm probability.
*/
BLOCK_NONE = "BLOCK_NONE",
/**
* Turn off the safety filter entirely.
*/
OFF = "OFF"
}
/** The harm category to be blocked. */
export declare enum HarmCategory {
/**
* Default value. This value is unused.
*/
HARM_CATEGORY_UNSPECIFIED = "HARM_CATEGORY_UNSPECIFIED",
/**
* Abusive, threatening, or content intended to bully, torment, or ridicule.
*/
HARM_CATEGORY_HARASSMENT = "HARM_CATEGORY_HARASSMENT",
/**
* Content that promotes violence or incites hatred against individuals or groups based on certain attributes.
*/
HARM_CATEGORY_HATE_SPEECH = "HARM_CATEGORY_HATE_SPEECH",
/**
* Content that contains sexually explicit material.
*/
HARM_CATEGORY_SEXUALLY_EXPLICIT = "HARM_CATEGORY_SEXUALLY_EXPLICIT",
/**
* Content that promotes, facilitates, or enables dangerous activities.
*/
HARM_CATEGORY_DANGEROUS_CONTENT = "HARM_CATEGORY_DANGEROUS_CONTENT",
/**
* Deprecated: Election filter is not longer supported. The harm category is civic integrity.
*/
HARM_CATEGORY_CIVIC_INTEGRITY = "HARM_CATEGORY_CIVIC_INTEGRITY",
/**
* Images that contain hate speech. This enum value is not supported in Gemini API.
*/
HARM_CATEGORY_IMAGE_HATE = "HARM_CATEGORY_IMAGE_HATE",
/**
* Images that contain dangerous content. This enum value is not supported in Gemini API.
*/
HARM_CATEGORY_IMAGE_DANGEROUS_CONTENT = "HARM_CATEGORY_IMAGE_DANGEROUS_CONTENT",
/**
* Images that contain harassment. This enum value is not supported in Gemini API.
*/
HARM_CATEGORY_IMAGE_HARASSMENT = "HARM_CATEGORY_IMAGE_HARASSMENT",
/**
* Images that contain sexually explicit content. This enum value is not supported in Gemini API.
*/
HARM_CATEGORY_IMAGE_SEXUALLY_EXPLICIT = "HARM_CATEGORY_IMAGE_SEXUALLY_EXPLICIT",
/**
* Prompts designed to bypass safety filters. This enum value is not supported in Gemini API.
*/
HARM_CATEGORY_JAILBREAK = "HARM_CATEGORY_JAILBREAK"
}
/** Output only. The probability of harm for this category. */
export declare enum HarmProbability {
/**
* The harm probability is unspecified.
*/
HARM_PROBABILITY_UNSPECIFIED = "HARM_PROBABILITY_UNSPECIFIED",
/**
* The harm probability is negligible.
*/
NEGLIGIBLE = "NEGLIGIBLE",
/**
* The harm probability is low.
*/
LOW = "LOW",
/**
* The harm probability is medium.
*/
MEDIUM = "MEDIUM",
/**
* The harm probability is high.
*/
HIGH = "HIGH"
}
/** Output only. The severity of harm for this category. This enum is not supported in Gemini API. */
export declare enum HarmSeverity {
/**
* The harm severity is unspecified.
*/
HARM_SEVERITY_UNSPECIFIED = "HARM_SEVERITY_UNSPECIFIED",
/**
* The harm severity is negligible.
*/
HARM_SEVERITY_NEGLIGIBLE = "HARM_SEVERITY_NEGLIGIBLE",
/**
* The harm severity is low.
*/
HARM_SEVERITY_LOW = "HARM_SEVERITY_LOW",
/**
* The harm severity is medium.
*/
HARM_SEVERITY_MEDIUM = "HARM_SEVERITY_MEDIUM",
/**
* The harm severity is high.
*/
HARM_SEVERITY_HIGH = "HARM_SEVERITY_HIGH"
}
declare type HeadersLike = Headers | readonly HeaderValue[][] | Record | undefined | null | NullableHeaders;
/**
* @license
* Copyright 2025 Google LLC
* SPDX-License-Identifier: Apache-2.0
*/
declare type HeaderValue = string | undefined | null;
/** The location of the API key. This enum is not supported in Gemini API. */
export declare enum HttpElementLocation {
HTTP_IN_UNSPECIFIED = "HTTP_IN_UNSPECIFIED",
/**
* Element is in the HTTP request query.
*/
HTTP_IN_QUERY = "HTTP_IN_QUERY",
/**
* Element is in the HTTP request header.
*/
HTTP_IN_HEADER = "HTTP_IN_HEADER",
/**
* Element is in the HTTP request path.
*/
HTTP_IN_PATH = "HTTP_IN_PATH",
/**
* Element is in the HTTP request body.
*/
HTTP_IN_BODY = "HTTP_IN_BODY",
/**
* Element is in the HTTP request cookie.
*/
HTTP_IN_COOKIE = "HTTP_IN_COOKIE"
}
declare type HTTPMethod = 'get' | 'post' | 'put' | 'patch' | 'delete';
/** HTTP options to be used in each of the requests. */
export declare interface HttpOptions {
/** The base URL for the AI platform service endpoint. */
baseUrl?: string;
/** The resource scope used to constructing the resource name when base_url is set */
baseUrlResourceScope?: ResourceScope;
/** Specifies the version of the API to use. */
apiVersion?: string;
/** Additional HTTP headers to be sent with the request. */
headers?: Record;
/** Timeout for the request in milliseconds. */
timeout?: number;
/** Extra parameters to add to the request body.
The structure must match the backend API's request structure.
- VertexAI backend API docs: https://cloud.google.com/vertex-ai/docs/reference/rest
- GeminiAPI backend API docs: https://ai.google.dev/api/rest */
extraBody?: Record;
/** HTTP retry options for the request. */
retryOptions?: HttpRetryOptions;
}
/**
* Represents the necessary information to send a request to an API endpoint.
* This interface defines the structure for constructing and executing HTTP
* requests.
*/
declare interface HttpRequest {
/**
* URL path from the modules, this path is appended to the base API URL to
* form the complete request URL.
*
* If you wish to set full URL, use httpOptions.baseUrl instead. Example to
* set full URL in the request:
*
* const request: HttpRequest = {
* path: '',
* httpOptions: {
* baseUrl: 'https://',
* apiVersion: '',
* },
* httpMethod: 'GET',
* };
*
* The result URL will be: https://
*
*/
path: string;
/**
* Optional query parameters to be appended to the request URL.
*/
queryParams?: Record;
/**
* Optional request body in json string or Blob format, GET request doesn't
* need a request body.
*/
body?: string | Blob;
/**
* The HTTP method to be used for the request.
*/
httpMethod: 'GET' | 'POST' | 'PATCH' | 'DELETE';
/**
* Optional set of customizable configuration for HTTP requests.
*/
httpOptions?: types.HttpOptions;
/**
* Optional abort signal which can be used to cancel the request.
*/
abortSignal?: AbortSignal;
}
/** A wrapper class for the http response. */
export declare class HttpResponse {
/** Used to retain the processed HTTP headers in the response. */
headers?: Record;
/**
* The original http response.
*/
responseInternal: Response;
constructor(response: Response);
json(): Promise;
}
/** HTTP retry options to be used in each of the requests. */
export declare interface HttpRetryOptions {
/** Maximum number of attempts, including the original request.
If 0 or 1, it means no retries. If not specified, default to 5. */
attempts?: number;
}
/** An image. */
declare interface Image_2 {
/** The Cloud Storage URI of the image. ``Image`` can contain a value
for this field or the ``image_bytes`` field but not both. */
gcsUri?: string;
/** The image bytes data. ``Image`` can contain a value for this field
or the ``gcs_uri`` field but not both.
* @remarks Encoded as base64 string. */
imageBytes?: string;
/** The MIME type of the image. */
mimeType?: string;
}
export { Image_2 as Image }
/** The image generation configuration to be used in GenerateContentConfig. */
export declare interface ImageConfig {
/** Aspect ratio of the generated images. Supported values are
"1:1", "2:3", "3:2", "3:4", "4:3", "9:16", "16:9", and "21:9". */
aspectRatio?: string;
/** Optional. Specifies the size of generated images. Supported
values are `1K`, `2K`, `4K`. If not specified, the model will use default
value `1K`. */
imageSize?: string;
/** Controls the generation of people. Supported values are:
ALLOW_ALL, ALLOW_ADULT, ALLOW_NONE. */
personGeneration?: string;
/** Controls whether prominent people (celebrities)
generation is allowed. If used with personGeneration, personGeneration
enum would take precedence. For instance, if ALLOW_NONE is set, all person
generation would be blocked. If this field is unspecified, the default
behavior is to allow prominent people. */
prominentPeople?: ProminentPeople;
/** MIME type of the generated image. This field is not
supported in Gemini API. */
outputMimeType?: string;
/** Compression quality of the generated image (for
``image/jpeg`` only). This field is not supported in Gemini API. */
outputCompressionQuality?: number;
/** Optional. The image output format for generated images. */
imageOutputOptions?: ImageConfigImageOutputOptions;
}
/**
* The configuration for image interaction.
*/
declare interface ImageConfig_2 {
aspect_ratio?: '1:1' | '2:3' | '3:2' | '3:4' | '4:3' | '4:5' | '5:4' | '9:16' | '16:9' | '21:9' | '1:8' | '8:1' | '1:4' | '4:1';
image_size?: '1K' | '2K' | '4K' | '512';
}
/** The image output format for generated images. This data type is not supported in Gemini API. */
export declare interface ImageConfigImageOutputOptions {
/** Optional. The compression quality of the output image. */
compressionQuality?: number;
/** Optional. The image format that the output should be saved as. */
mimeType?: string;
}
/**
* An image content block.
*/
declare interface ImageContent {
type: 'image';
/**
* The image content.
*/
data?: string;
/**
* The mime type of the image.
*/
mime_type?: 'image/png' | 'image/jpeg' | 'image/webp' | 'image/heic' | 'image/heif';
/**
* The resolution of the media.
*/
resolution?: 'low' | 'medium' | 'high' | 'ultra_high';
/**
* The URI of the image.
*/
uri?: string;
}
/** Enum that specifies the language of the text in the prompt. */
export declare enum ImagePromptLanguage {
/**
* Auto-detect the language.
*/
auto = "auto",
/**
* English
*/
en = "en",
/**
* Japanese
*/
ja = "ja",
/**
* Korean
*/
ko = "ko",
/**
* Hindi
*/
hi = "hi",
/**
* Chinese
*/
zh = "zh",
/**
* Portuguese
*/
pt = "pt",
/**
* Spanish
*/
es = "es"
}
/** Image search for grounding and related configurations. */
export declare interface ImageSearch {
}
/** Optional parameters for importing a file. */
export declare interface ImportFileConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** User provided custom metadata stored as key-value pairs used for querying. */
customMetadata?: CustomMetadata[];
/** Config for telling the service how to chunk the file. */
chunkingConfig?: ChunkingConfig;
}
/** Long-running operation for importing a file to a FileSearchStore. */
export declare class ImportFileOperation implements Operation {
/** The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. */
name?: string;
/** Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. */
metadata?: Record;
/** If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. */
done?: boolean;
/** The error result of the operation in case of failure or cancellation. */
error?: Record;
/** The result of the ImportFile operation, available when the operation is done. */
response?: ImportFileResponse;
/**
* Instantiates an Operation of the same type as the one being called with the fields set from the API response.
* @internal
*/
_fromAPIResponse({ apiResponse, _isVertexAI, }: OperationFromAPIResponseParameters): Operation;
/** The full HTTP response. */
sdkHttpResponse?: HttpResponse;
}
/** Config for file_search_stores.import_file parameters. */
export declare interface ImportFileParameters {
/** The resource name of the FileSearchStore. Example: `fileSearchStores/my-file-search-store-123` */
fileSearchStoreName: string;
/** The name of the File API File to import. Example: `files/abc-123` */
fileName: string;
/** Optional parameters for the request. */
config?: ImportFileConfig;
}
/** Response for ImportFile to import a File API file with a file search store. */
export declare class ImportFileResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
/** The name of the FileSearchStore containing Documents. */
parent?: string;
/** The identifier for the Document imported. */
documentName?: string;
}
/** Config for `inlined_embedding_responses` parameter. */
export declare class InlinedEmbedContentResponse {
/** The response to the request.
*/
response?: SingleEmbedContentResponse;
/** The error encountered while processing the request.
*/
error?: JobError;
}
/** Config for inlined request. */
export declare interface InlinedRequest {
/** ID of the model to use. For a list of models, see `Google models
`_. */
model?: string;
/** Content of the request.
*/
contents?: ContentListUnion;
/** The metadata to be associated with the request. */
metadata?: Record;
/** Configuration that contains optional model parameters.
*/
config?: GenerateContentConfig;
}
/** Config for `inlined_responses` parameter. */
export declare class InlinedResponse {
/** The response to the request.
*/
response?: GenerateContentResponse;
/** The metadata to be associated with the request. */
metadata?: Record;
/** The error encountered while processing the request.
*/
error?: JobError;
}
/**
* The Interaction resource.
*/
declare interface Interaction {
/**
* Output only. A unique identifier for the interaction completion.
*/
id: string;
/**
* Output only. The time at which the response was created in ISO 8601 format
* (YYYY-MM-DDThh:mm:ssZ).
*/
created: string;
/**
* Output only. The status of the interaction.
*/
status: 'in_progress' | 'requires_action' | 'completed' | 'failed' | 'cancelled' | 'incomplete';
/**
* Output only. The time at which the response was last updated in ISO 8601 format
* (YYYY-MM-DDThh:mm:ssZ).
*/
updated: string;
/**
* The name of the `Agent` used for generating the interaction.
*/
agent?: (string & {}) | 'deep-research-pro-preview-12-2025';
/**
* Configuration parameters for the agent interaction.
*/
agent_config?: DynamicAgentConfig | DeepResearchAgentConfig;
/**
* The input for the interaction.
*/
input?: Array | string | Array | TextContent | ImageContent | AudioContent | DocumentContent | VideoContent | ThoughtContent | FunctionCallContent | FunctionResultContent | CodeExecutionCallContent | CodeExecutionResultContent | URLContextCallContent | URLContextResultContent | GoogleSearchCallContent | GoogleSearchResultContent | MCPServerToolCallContent | MCPServerToolResultContent | FileSearchCallContent | FileSearchResultContent;
/**
* The name of the `Model` used for generating the interaction.
*/
model?: Model_2;
/**
* Output only. Responses from the model.
*/
outputs?: Array;
/**
* The ID of the previous interaction, if any.
*/
previous_interaction_id?: string;
/**
* Enforces that the generated response is a JSON object that complies with
* the JSON schema specified in this field.
*/
response_format?: unknown;
/**
* The mime type of the response. This is required if response_format is set.
*/
response_mime_type?: string;
/**
* The requested modalities of the response (TEXT, IMAGE, AUDIO).
*/
response_modalities?: Array<'text' | 'image' | 'audio'>;
/**
* Output only. The role of the interaction.
*/
role?: string;
/**
* System instruction for the interaction.
*/
system_instruction?: string;
/**
* A list of tool declarations the model may call during interaction.
*/
tools?: Array;
/**
* Output only. Statistics on the interaction request's token usage.
*/
usage?: Usage;
}
declare interface InteractionCancelParams {
/**
* Which version of the API to use.
*/
api_version?: string;
}
declare interface InteractionCompleteEvent {
event_type: 'interaction.complete';
/**
* The completed interaction with empty outputs to reduce the payload size.
* Use the preceding ContentDelta events for the actual output.
*/
interaction: Interaction;
/**
* The event_id token to be used to resume the interaction stream, from this event.
*/
event_id?: string;
}
declare type InteractionCreateParams = CreateModelInteractionParamsNonStreaming | CreateModelInteractionParamsStreaming | CreateAgentInteractionParamsNonStreaming | CreateAgentInteractionParamsStreaming;
declare interface InteractionDeleteParams {
/**
* Which version of the API to use.
*/
api_version?: string;
}
declare type InteractionDeleteResponse = unknown;
declare type InteractionGetParams = InteractionGetParamsNonStreaming | InteractionGetParamsStreaming;
declare namespace InteractionGetParams {
type InteractionGetParamsNonStreaming = InteractionsAPI.InteractionGetParamsNonStreaming;
type InteractionGetParamsStreaming = InteractionsAPI.InteractionGetParamsStreaming;
}
declare interface InteractionGetParamsBase {
/**
* Path param: Which version of the API to use.
*/
api_version?: string;
/**
* Query param: If set to true, includes the input in the response.
*/
include_input?: boolean;
/**
* Query param: Optional. If set, resumes the interaction stream from the next chunk after the event marked by the event id. Can only be used if `stream` is true.
*/
last_event_id?: string;
/**
* Query param: If set to true, the generated content will be streamed incrementally.
*/
stream?: boolean;
}
declare interface InteractionGetParamsNonStreaming extends InteractionGetParamsBase {
/**
* Query param: If set to true, the generated content will be streamed incrementally.
*/
stream?: false;
}
declare interface InteractionGetParamsStreaming extends InteractionGetParamsBase {
/**
* Query param: If set to true, the generated content will be streamed incrementally.
*/
stream: true;
}
export declare class Interactions extends BaseInteractions {
}
export declare namespace Interactions {
export { type AllowedTools as AllowedTools, type Annotation as Annotation, type AudioContent as AudioContent, type CodeExecutionCallArguments as CodeExecutionCallArguments, type CodeExecutionCallContent as CodeExecutionCallContent, type CodeExecutionResultContent as CodeExecutionResultContent, type Content_2 as Content, type ContentDelta as ContentDelta, type ContentStart as ContentStart, type ContentStop as ContentStop, type DeepResearchAgentConfig as DeepResearchAgentConfig, type DocumentContent as DocumentContent, type DynamicAgentConfig as DynamicAgentConfig, type ErrorEvent_2 as ErrorEvent, type FileSearchCallContent as FileSearchCallContent, type FileSearchResultContent as FileSearchResultContent, type Function_2 as Function, type FunctionCallContent as FunctionCallContent, type FunctionResultContent as FunctionResultContent, type GenerationConfig_2 as GenerationConfig, type GoogleSearchCallArguments as GoogleSearchCallArguments, type GoogleSearchCallContent as GoogleSearchCallContent, type GoogleSearchResult as GoogleSearchResult, type GoogleSearchResultContent as GoogleSearchResultContent, type ImageConfig_2 as ImageConfig, type ImageContent as ImageContent, type Interaction as Interaction, type InteractionCompleteEvent as InteractionCompleteEvent, type InteractionSSEEvent as InteractionSSEEvent, type InteractionStartEvent as InteractionStartEvent, type InteractionStatusUpdate as InteractionStatusUpdate, type MCPServerToolCallContent as MCPServerToolCallContent, type MCPServerToolResultContent as MCPServerToolResultContent, type Model_2 as Model, type SpeechConfig_2 as SpeechConfig, type TextContent as TextContent, type ThinkingLevel_2 as ThinkingLevel, type ThoughtContent as ThoughtContent, type Tool_2 as Tool, type ToolChoiceConfig as ToolChoiceConfig, type ToolChoiceType as ToolChoiceType, type Turn as Turn, type URLContextCallArguments as URLContextCallArguments, type URLContextCallContent as URLContextCallContent, type URLContextResult as URLContextResult, type URLContextResultContent as URLContextResultContent, type Usage as Usage, type VideoContent as VideoContent, type InteractionDeleteResponse as InteractionDeleteResponse, type InteractionCreateParams as InteractionCreateParams, type CreateModelInteractionParamsNonStreaming as CreateModelInteractionParamsNonStreaming, type CreateModelInteractionParamsStreaming as CreateModelInteractionParamsStreaming, type CreateAgentInteractionParamsNonStreaming as CreateAgentInteractionParamsNonStreaming, type CreateAgentInteractionParamsStreaming as CreateAgentInteractionParamsStreaming, type InteractionDeleteParams as InteractionDeleteParams, type InteractionCancelParams as InteractionCancelParams, type InteractionGetParams as InteractionGetParams, type InteractionGetParamsNonStreaming as InteractionGetParamsNonStreaming, type InteractionGetParamsStreaming as InteractionGetParamsStreaming, };
}
declare namespace InteractionsAPI {
export {
BaseInteractions,
Interactions,
AllowedTools,
Annotation,
AudioContent,
CodeExecutionCallArguments,
CodeExecutionCallContent,
CodeExecutionResultContent,
Content_2 as Content,
ContentDelta,
ContentStart,
ContentStop,
DeepResearchAgentConfig,
DocumentContent,
DynamicAgentConfig,
ErrorEvent_2 as ErrorEvent,
FileSearchCallContent,
FileSearchResultContent,
Function_2 as Function,
FunctionCallContent,
FunctionResultContent,
GenerationConfig_2 as GenerationConfig,
GoogleSearchCallArguments,
GoogleSearchCallContent,
GoogleSearchResult,
GoogleSearchResultContent,
ImageConfig_2 as ImageConfig,
ImageContent,
Interaction,
InteractionCompleteEvent,
InteractionSSEEvent,
InteractionStartEvent,
InteractionStatusUpdate,
MCPServerToolCallContent,
MCPServerToolResultContent,
Model_2 as Model,
SpeechConfig_2 as SpeechConfig,
TextContent,
ThinkingLevel_2 as ThinkingLevel,
ThoughtContent,
Tool_2 as Tool,
ToolChoiceConfig,
ToolChoiceType,
Turn,
URLContextCallArguments,
URLContextCallContent,
URLContextResult,
URLContextResultContent,
Usage,
VideoContent,
InteractionDeleteResponse,
InteractionCreateParams,
BaseCreateModelInteractionParams,
BaseCreateAgentInteractionParams,
CreateModelInteractionParamsNonStreaming,
CreateModelInteractionParamsStreaming,
CreateAgentInteractionParamsNonStreaming,
CreateAgentInteractionParamsStreaming,
InteractionDeleteParams,
InteractionCancelParams,
InteractionGetParams,
InteractionGetParamsBase,
InteractionGetParamsNonStreaming,
InteractionGetParamsStreaming
}
}
declare type InteractionSSEEvent = InteractionStartEvent | InteractionCompleteEvent | InteractionStatusUpdate | ContentStart | ContentDelta | ContentStop | ErrorEvent_2;
declare interface InteractionStartEvent {
event_type: 'interaction.start';
/**
* The Interaction resource.
*/
interaction: Interaction;
/**
* The event_id token to be used to resume the interaction stream, from this event.
*/
event_id?: string;
}
declare interface InteractionStatusUpdate {
event_type: 'interaction.status_update';
interaction_id: string;
status: 'in_progress' | 'requires_action' | 'completed' | 'failed' | 'cancelled' | 'incomplete';
/**
* The event_id token to be used to resume the interaction stream, from this event.
*/
event_id?: string;
}
/** Parameters for the private _Register method. */
export declare interface InternalRegisterFilesParameters {
/** The Google Cloud Storage URIs to register. Example: `gs://bucket/object`. */
uris: string[];
/** Used to override the default configuration. */
config?: RegisterFilesConfig;
}
declare class InternalServerError extends APIError {
}
/** Represents a time interval, encoded as a Timestamp start (inclusive) and a Timestamp end (exclusive). The start must be less than or equal to the end. When the start equals the end, the interval is empty (matches no time). When both start and end are unspecified, the interval matches any time. */
export declare interface Interval {
/** Optional. Exclusive end of the interval. If specified, a Timestamp matching this interval will have to be before the end. */
endTime?: string;
/** Optional. Inclusive start of the interval. If specified, a Timestamp matching this interval will have to be the same or after the start. */
startTime?: string;
}
/** Job error. */
export declare interface JobError {
/** A list of messages that carry the error details. There is a common set of message types for APIs to use. */
details?: string[];
/** The status code. */
code?: number;
/** A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the `details` field. */
message?: string;
}
/** Job state. */
export declare enum JobState {
/**
* The job state is unspecified.
*/
JOB_STATE_UNSPECIFIED = "JOB_STATE_UNSPECIFIED",
/**
* The job has been just created or resumed and processing has not yet begun.
*/
JOB_STATE_QUEUED = "JOB_STATE_QUEUED",
/**
* The service is preparing to run the job.
*/
JOB_STATE_PENDING = "JOB_STATE_PENDING",
/**
* The job is in progress.
*/
JOB_STATE_RUNNING = "JOB_STATE_RUNNING",
/**
* The job completed successfully.
*/
JOB_STATE_SUCCEEDED = "JOB_STATE_SUCCEEDED",
/**
* The job failed.
*/
JOB_STATE_FAILED = "JOB_STATE_FAILED",
/**
* The job is being cancelled. From this state the job may only go to either `JOB_STATE_SUCCEEDED`, `JOB_STATE_FAILED` or `JOB_STATE_CANCELLED`.
*/
JOB_STATE_CANCELLING = "JOB_STATE_CANCELLING",
/**
* The job has been cancelled.
*/
JOB_STATE_CANCELLED = "JOB_STATE_CANCELLED",
/**
* The job has been stopped, and can be resumed.
*/
JOB_STATE_PAUSED = "JOB_STATE_PAUSED",
/**
* The job has expired.
*/
JOB_STATE_EXPIRED = "JOB_STATE_EXPIRED",
/**
* The job is being updated. Only jobs in the `JOB_STATE_RUNNING` state can be updated. After updating, the job goes back to the `JOB_STATE_RUNNING` state.
*/
JOB_STATE_UPDATING = "JOB_STATE_UPDATING",
/**
* The job is partially succeeded, some results may be missing due to errors.
*/
JOB_STATE_PARTIALLY_SUCCEEDED = "JOB_STATE_PARTIALLY_SUCCEEDED"
}
/** Programming language of the `code`. */
export declare enum Language {
/**
* Unspecified language. This value should not be used.
*/
LANGUAGE_UNSPECIFIED = "LANGUAGE_UNSPECIFIED",
/**
* Python >= 3.10, with numpy and simpy available.
*/
PYTHON = "PYTHON"
}
/** An object that represents a latitude/longitude pair.
This is expressed as a pair of doubles to represent degrees latitude and
degrees longitude. Unless specified otherwise, this object must conform to the
WGS84 standard. Values must be within normalized ranges. */
export declare interface LatLng {
/** The latitude in degrees. It must be in the range [-90.0, +90.0]. */
latitude?: number;
/** The longitude in degrees. It must be in the range [-180.0, +180.0] */
longitude?: number;
}
/** Config for optional parameters. */
export declare interface ListBatchJobsConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
pageSize?: number;
pageToken?: string;
filter?: string;
}
/** Config for batches.list parameters. */
export declare interface ListBatchJobsParameters {
config?: ListBatchJobsConfig;
}
/** Config for batches.list return value. */
export declare class ListBatchJobsResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
nextPageToken?: string;
batchJobs?: BatchJob[];
}
/** Config for caches.list method. */
export declare interface ListCachedContentsConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
pageSize?: number;
pageToken?: string;
}
/** Parameters for caches.list method. */
export declare interface ListCachedContentsParameters {
/** Configuration that contains optional parameters.
*/
config?: ListCachedContentsConfig;
}
export declare class ListCachedContentsResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
nextPageToken?: string;
/** List of cached contents.
*/
cachedContents?: CachedContent[];
}
/** Config for optional parameters. */
export declare interface ListDocumentsConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
pageSize?: number;
pageToken?: string;
}
/** Config for documents.list parameters. */
export declare interface ListDocumentsParameters {
/** The resource name of the FileSearchStores. Example: `fileSearchStore/file-search-store-foo` */
parent: string;
config?: ListDocumentsConfig;
}
/** Config for documents.list return value. */
export declare class ListDocumentsResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
/** A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no more pages. */
nextPageToken?: string;
/** The returned `Document`s. */
documents?: Document_2[];
}
export declare interface ListFilesConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
pageSize?: number;
pageToken?: string;
}
/** Optional parameters for listing FileSearchStore. */
export declare interface ListFileSearchStoresConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
pageSize?: number;
pageToken?: string;
}
/** Config for file_search_stores.list parameters. */
export declare interface ListFileSearchStoresParameters {
/** Optional parameters for the list request. */
config?: ListFileSearchStoresConfig;
}
/** Config for file_search_stores.list return value. */
export declare class ListFileSearchStoresResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
nextPageToken?: string;
/** The returned file search stores. */
fileSearchStores?: FileSearchStore[];
}
/** Generates the parameters for the list method. */
export declare interface ListFilesParameters {
/** Used to override the default configuration. */
config?: ListFilesConfig;
}
/** Response for the list files method. */
export declare class ListFilesResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
/** A token that can be sent as a `page_token` into a subsequent `ListFiles` call. */
nextPageToken?: string;
/** The list of `File`s. */
files?: File_2[];
}
export declare interface ListModelsConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
pageSize?: number;
pageToken?: string;
filter?: string;
/** Set true to list base models, false to list tuned models. */
queryBase?: boolean;
}
export declare interface ListModelsParameters {
config?: ListModelsConfig;
}
export declare class ListModelsResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
nextPageToken?: string;
models?: Model[];
}
/** Configuration for the list tuning jobs method. */
export declare interface ListTuningJobsConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
pageSize?: number;
pageToken?: string;
filter?: string;
}
/** Parameters for the list tuning jobs method. */
export declare interface ListTuningJobsParameters {
config?: ListTuningJobsConfig;
}
/** Response for the list tuning jobs method. */
export declare class ListTuningJobsResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
/** A token to retrieve the next page of results. Pass this token in a subsequent [GenAiTuningService.ListTuningJobs] call to retrieve the next page of results. */
nextPageToken?: string;
/** The tuning jobs that match the request. */
tuningJobs?: TuningJob[];
}
/**
Live class encapsulates the configuration for live interaction with the
Generative Language API. It embeds ApiClient for general API settings.
@experimental
*/
export declare class Live {
private readonly apiClient;
private readonly auth;
private readonly webSocketFactory;
readonly music: LiveMusic;
constructor(apiClient: ApiClient, auth: Auth, webSocketFactory: WebSocketFactory);
/**
Establishes a connection to the specified model with the given
configuration and returns a Session object representing that connection.
@experimental Built-in MCP support is an experimental feature, may change in
future versions.
@remarks
@param params - The parameters for establishing a connection to the model.
@return A live session.
@example
```ts
let model: string;
if (GOOGLE_GENAI_USE_VERTEXAI) {
model = 'gemini-2.0-flash-live-preview-04-09';
} else {
model = 'gemini-live-2.5-flash-preview';
}
const session = await ai.live.connect({
model: model,
config: {
responseModalities: [Modality.AUDIO],
},
callbacks: {
onopen: () => {
console.log('Connected to the socket.');
},
onmessage: (e: MessageEvent) => {
console.log('Received message from the server: %s\n', debug(e.data));
},
onerror: (e: ErrorEvent) => {
console.log('Error occurred: %s\n', debug(e.error));
},
onclose: (e: CloseEvent) => {
console.log('Connection closed.');
},
},
});
```
*/
connect(params: types.LiveConnectParameters): Promise;
private isCallableTool;
}
/** Callbacks for the live API. */
export declare interface LiveCallbacks {
/**
* Called when the websocket connection is established.
*/
onopen?: (() => void) | null;
/**
* Called when a message is received from the server.
*/
onmessage: (e: LiveServerMessage) => void;
/**
* Called when an error occurs.
*/
onerror?: ((e: ErrorEvent) => void) | null;
/**
* Called when the websocket connection is closed.
*/
onclose?: ((e: CloseEvent) => void) | null;
}
/** Incremental update of the current conversation delivered from the client.
All the content here will unconditionally be appended to the conversation
history and used as part of the prompt to the model to generate content.
A message here will interrupt any current model generation. */
export declare interface LiveClientContent {
/** The content appended to the current conversation with the model.
For single-turn queries, this is a single instance. For multi-turn
queries, this is a repeated field that contains conversation history and
latest request.
*/
turns?: Content[];
/** If true, indicates that the server content generation should start with
the currently accumulated prompt. Otherwise, the server will await
additional messages before starting generation. */
turnComplete?: boolean;
}
/** Messages sent by the client in the API call. */
export declare interface LiveClientMessage {
/** Message to be sent by the system when connecting to the API. SDK users should not send this message. */
setup?: LiveClientSetup;
/** Incremental update of the current conversation delivered from the client. */
clientContent?: LiveClientContent;
/** User input that is sent in real time. */
realtimeInput?: LiveClientRealtimeInput;
/** Response to a `ToolCallMessage` received from the server. */
toolResponse?: LiveClientToolResponse;
}
/** User input that is sent in real time.
This is different from `LiveClientContent` in a few ways:
- Can be sent continuously without interruption to model generation.
- If there is a need to mix data interleaved across the
`LiveClientContent` and the `LiveClientRealtimeInput`, server attempts to
optimize for best response, but there are no guarantees.
- End of turn is not explicitly specified, but is rather derived from user
activity (for example, end of speech).
- Even before the end of turn, the data is processed incrementally
to optimize for a fast start of the response from the model.
- Is always assumed to be the user's input (cannot be used to populate
conversation history). */
export declare interface LiveClientRealtimeInput {
/** Inlined bytes data for media input. */
mediaChunks?: Blob_2[];
/** The realtime audio input stream. */
audio?: Blob_2;
/**
Indicates that the audio stream has ended, e.g. because the microphone was
turned off.
This should only be sent when automatic activity detection is enabled
(which is the default).
The client can reopen the stream by sending an audio message.
*/
audioStreamEnd?: boolean;
/** The realtime video input stream. */
video?: Blob_2;
/** The realtime text input stream. */
text?: string;
/** Marks the start of user activity. */
activityStart?: ActivityStart;
/** Marks the end of user activity. */
activityEnd?: ActivityEnd;
}
/** Message contains configuration that will apply for the duration of the streaming session. */
export declare interface LiveClientSetup {
/**
The fully qualified name of the publisher model or tuned model endpoint to
use.
*/
model?: string;
/** The generation configuration for the session.
Note: only a subset of fields are supported.
*/
generationConfig?: GenerationConfig;
/** The user provided system instructions for the model.
Note: only text should be used in parts and content in each part will be
in a separate paragraph. */
systemInstruction?: ContentUnion;
/** A list of `Tools` the model may use to generate the next response.
A `Tool` is a piece of code that enables the system to interact with
external systems to perform an action, or set of actions, outside of
knowledge and scope of the model. */
tools?: ToolListUnion;
/** Configures the realtime input behavior in BidiGenerateContent. */
realtimeInputConfig?: RealtimeInputConfig;
/** Configures session resumption mechanism.
If included server will send SessionResumptionUpdate messages. */
sessionResumption?: SessionResumptionConfig;
/** Configures context window compression mechanism.
If included, server will compress context window to fit into given length. */
contextWindowCompression?: ContextWindowCompressionConfig;
/** The transcription of the input aligns with the input audio language.
*/
inputAudioTranscription?: AudioTranscriptionConfig;
/** The transcription of the output aligns with the language code
specified for the output audio.
*/
outputAudioTranscription?: AudioTranscriptionConfig;
/** Configures the proactivity of the model. This allows the model to respond proactively to
the input and to ignore irrelevant input. */
proactivity?: ProactivityConfig;
/** Configures the explicit VAD signal. If enabled, the client will send
vad_signal to indicate the start and end of speech. This allows the server
to process the audio more efficiently. */
explicitVadSignal?: boolean;
}
/** Client generated response to a `ToolCall` received from the server.
Individual `FunctionResponse` objects are matched to the respective
`FunctionCall` objects by the `id` field.
Note that in the unary and server-streaming GenerateContent APIs function
calling happens by exchanging the `Content` parts, while in the bidi
GenerateContent APIs function calling happens over this dedicated set of
messages. */
export declare class LiveClientToolResponse {
/** The response to the function calls. */
functionResponses?: FunctionResponse[];
}
/** Session config for the API connection. */
export declare interface LiveConnectConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** The generation configuration for the session. */
generationConfig?: GenerationConfig;
/** The requested modalities of the response. Represents the set of
modalities that the model can return. Defaults to AUDIO if not specified.
*/
responseModalities?: Modality[];
/** Value that controls the degree of randomness in token selection.
Lower temperatures are good for prompts that require a less open-ended or
creative response, while higher temperatures can lead to more diverse or
creative results.
*/
temperature?: number;
/** Tokens are selected from the most to least probable until the sum
of their probabilities equals this value. Use a lower value for less
random responses and a higher value for more random responses.
*/
topP?: number;
/** For each token selection step, the ``top_k`` tokens with the
highest probabilities are sampled. Then tokens are further filtered based
on ``top_p`` with the final token selected using temperature sampling. Use
a lower number for less random responses and a higher number for more
random responses.
*/
topK?: number;
/** Maximum number of tokens that can be generated in the response.
*/
maxOutputTokens?: number;
/** If specified, the media resolution specified will be used.
*/
mediaResolution?: MediaResolution;
/** When ``seed`` is fixed to a specific number, the model makes a best
effort to provide the same response for repeated requests. By default, a
random number is used.
*/
seed?: number;
/** The speech generation configuration.
*/
speechConfig?: SpeechConfig;
/** Config for thinking features.
An error will be returned if this field is set for models that don't
support thinking.
*/
thinkingConfig?: ThinkingConfig;
/** If enabled, the model will detect emotions and adapt its responses accordingly. */
enableAffectiveDialog?: boolean;
/** The user provided system instructions for the model.
Note: only text should be used in parts and content in each part will be
in a separate paragraph. */
systemInstruction?: ContentUnion;
/** A list of `Tools` the model may use to generate the next response.
A `Tool` is a piece of code that enables the system to interact with
external systems to perform an action, or set of actions, outside of
knowledge and scope of the model. */
tools?: ToolListUnion;
/** Configures session resumption mechanism.
If included the server will send SessionResumptionUpdate messages. */
sessionResumption?: SessionResumptionConfig;
/** The transcription of the input aligns with the input audio language.
*/
inputAudioTranscription?: AudioTranscriptionConfig;
/** The transcription of the output aligns with the language code
specified for the output audio.
*/
outputAudioTranscription?: AudioTranscriptionConfig;
/** Configures the realtime input behavior in BidiGenerateContent. */
realtimeInputConfig?: RealtimeInputConfig;
/** Configures context window compression mechanism.
If included, server will compress context window to fit into given length. */
contextWindowCompression?: ContextWindowCompressionConfig;
/** Configures the proactivity of the model. This allows the model to respond proactively to
the input and to ignore irrelevant input. */
proactivity?: ProactivityConfig;
/** Configures the explicit VAD signal. If enabled, the client will send
vad_signal to indicate the start and end of speech. This allows the server
to process the audio more efficiently. */
explicitVadSignal?: boolean;
}
/** Config for LiveConnectConstraints for Auth Token creation. */
export declare interface LiveConnectConstraints {
/** ID of the model to configure in the ephemeral token for Live API.
For a list of models, see `Gemini models
`. */
model?: string;
/** Configuration specific to Live API connections created using this token. */
config?: LiveConnectConfig;
}
/** Parameters for connecting to the live API. */
export declare interface LiveConnectParameters {
/** ID of the model to use. For a list of models, see `Google models
`_. */
model: string;
/** callbacks */
callbacks: LiveCallbacks;
/** Optional configuration parameters for the request.
*/
config?: LiveConnectConfig;
}
/**
LiveMusic class encapsulates the configuration for live music
generation via Lyria Live models.
@experimental
*/
declare class LiveMusic {
private readonly apiClient;
private readonly auth;
private readonly webSocketFactory;
constructor(apiClient: ApiClient, auth: Auth, webSocketFactory: WebSocketFactory);
/**
Establishes a connection to the specified model and returns a
LiveMusicSession object representing that connection.
@experimental
@remarks
@param params - The parameters for establishing a connection to the model.
@return A live session.
@example
```ts
let model = 'models/lyria-realtime-exp';
const session = await ai.live.music.connect({
model: model,
callbacks: {
onmessage: (e: MessageEvent) => {
console.log('Received message from the server: %s\n', debug(e.data));
},
onerror: (e: ErrorEvent) => {
console.log('Error occurred: %s\n', debug(e.error));
},
onclose: (e: CloseEvent) => {
console.log('Connection closed.');
},
},
});
```
*/
connect(params: types.LiveMusicConnectParameters): Promise;
}
/** Callbacks for the realtime music API. */
export declare interface LiveMusicCallbacks {
/**
* Called when a message is received from the server.
*/
onmessage: (e: LiveMusicServerMessage) => void;
/**
* Called when an error occurs.
*/
onerror?: ((e: ErrorEvent) => void) | null;
/**
* Called when the websocket connection is closed.
*/
onclose?: ((e: CloseEvent) => void) | null;
}
/** User input to start or steer the music. */
export declare interface LiveMusicClientContent {
/** Weighted prompts as the model input. */
weightedPrompts?: WeightedPrompt[];
}
/** Messages sent by the client in the LiveMusicClientMessage call. */
export declare interface LiveMusicClientMessage {
/** Message to be sent in the first (and only in the first) `LiveMusicClientMessage`.
Clients should wait for a `LiveMusicSetupComplete` message before
sending any additional messages. */
setup?: LiveMusicClientSetup;
/** User input to influence music generation. */
clientContent?: LiveMusicClientContent;
/** Configuration for music generation. */
musicGenerationConfig?: LiveMusicGenerationConfig;
/** Playback control signal for the music generation. */
playbackControl?: LiveMusicPlaybackControl;
}
/** Message to be sent by the system when connecting to the API. */
export declare interface LiveMusicClientSetup {
/** The model's resource name. Format: `models/{model}`. */
model?: string;
}
/** Parameters for connecting to the live API. */
export declare interface LiveMusicConnectParameters {
/** The model's resource name. */
model: string;
/** Callbacks invoked on server events. */
callbacks: LiveMusicCallbacks;
}
/** A prompt that was filtered with the reason. */
export declare interface LiveMusicFilteredPrompt {
/** The text prompt that was filtered. */
text?: string;
/** The reason the prompt was filtered. */
filteredReason?: string;
}
/** Configuration for music generation. */
export declare interface LiveMusicGenerationConfig {
/** Controls the variance in audio generation. Higher values produce
higher variance. Range is [0.0, 3.0]. */
temperature?: number;
/** Controls how the model selects tokens for output. Samples the topK
tokens with the highest probabilities. Range is [1, 1000]. */
topK?: number;
/** Seeds audio generation. If not set, the request uses a randomly
generated seed. */
seed?: number;
/** Controls how closely the model follows prompts.
Higher guidance follows more closely, but will make transitions more
abrupt. Range is [0.0, 6.0]. */
guidance?: number;
/** Beats per minute. Range is [60, 200]. */
bpm?: number;
/** Density of sounds. Range is [0.0, 1.0]. */
density?: number;
/** Brightness of the music. Range is [0.0, 1.0]. */
brightness?: number;
/** Scale of the generated music. */
scale?: Scale;
/** Whether the audio output should contain bass. */
muteBass?: boolean;
/** Whether the audio output should contain drums. */
muteDrums?: boolean;
/** Whether the audio output should contain only bass and drums. */
onlyBassAndDrums?: boolean;
/** The mode of music generation. Default mode is QUALITY. */
musicGenerationMode?: MusicGenerationMode;
}
/** The playback control signal to apply to the music generation. */
export declare enum LiveMusicPlaybackControl {
/**
* This value is unused.
*/
PLAYBACK_CONTROL_UNSPECIFIED = "PLAYBACK_CONTROL_UNSPECIFIED",
/**
* Start generating the music.
*/
PLAY = "PLAY",
/**
* Hold the music generation. Use PLAY to resume from the current position.
*/
PAUSE = "PAUSE",
/**
* Stop the music generation and reset the context (prompts retained).
Use PLAY to restart the music generation.
*/
STOP = "STOP",
/**
* Reset the context of the music generation without stopping it.
Retains the current prompts and config.
*/
RESET_CONTEXT = "RESET_CONTEXT"
}
/** Server update generated by the model in response to client messages.
Content is generated as quickly as possible, and not in real time.
Clients may choose to buffer and play it out in real time. */
export declare interface LiveMusicServerContent {
/** The audio chunks that the model has generated. */
audioChunks?: AudioChunk[];
}
/** Response message for the LiveMusicClientMessage call. */
export declare class LiveMusicServerMessage {
/** Message sent in response to a `LiveMusicClientSetup` message from the client.
Clients should wait for this message before sending any additional messages. */
setupComplete?: LiveMusicServerSetupComplete;
/** Content generated by the model in response to client messages. */
serverContent?: LiveMusicServerContent;
/** A prompt that was filtered with the reason. */
filteredPrompt?: LiveMusicFilteredPrompt;
/**
* Returns the first audio chunk from the server content, if present.
*
* @remarks
* If there are no audio chunks in the response, undefined will be returned.
*/
get audioChunk(): AudioChunk | undefined;
}
/** Sent in response to a `LiveMusicClientSetup` message from the client. */
export declare interface LiveMusicServerSetupComplete {
}
/**
Represents a connection to the API.
@experimental
*/
export declare class LiveMusicSession {
readonly conn: WebSocket_2;
private readonly apiClient;
constructor(conn: WebSocket_2, apiClient: ApiClient);
/**
Sets inputs to steer music generation. Updates the session's current
weighted prompts.
@param params - Contains one property, `weightedPrompts`.
- `weightedPrompts` to send to the model; weights are normalized to
sum to 1.0.
@experimental
*/
setWeightedPrompts(params: types.LiveMusicSetWeightedPromptsParameters): Promise;
/**
Sets a configuration to the model. Updates the session's current
music generation config.
@param params - Contains one property, `musicGenerationConfig`.
- `musicGenerationConfig` to set in the model. Passing an empty or
undefined config to the model will reset the config to defaults.
@experimental
*/
setMusicGenerationConfig(params: types.LiveMusicSetConfigParameters): Promise;
private sendPlaybackControl;
/**
* Start the music stream.
*
* @experimental
*/
play(): void;
/**
* Temporarily halt the music stream. Use `play` to resume from the current
* position.
*
* @experimental
*/
pause(): void;
/**
* Stop the music stream and reset the state. Retains the current prompts
* and config.
*
* @experimental
*/
stop(): void;
/**
* Resets the context of the music generation without stopping it.
* Retains the current prompts and config.
*
* @experimental
*/
resetContext(): void;
/**
Terminates the WebSocket connection.
@experimental
*/
close(): void;
}
/** Parameters for setting config for the live music API. */
export declare interface LiveMusicSetConfigParameters {
/** Configuration for music generation. */
musicGenerationConfig: LiveMusicGenerationConfig;
}
/** Parameters for setting weighted prompts for the live music API. */
export declare interface LiveMusicSetWeightedPromptsParameters {
/** A map of text prompts to weights to use for the generation request. */
weightedPrompts: WeightedPrompt[];
}
/** Prompts and config used for generating this audio chunk. */
export declare interface LiveMusicSourceMetadata {
/** Weighted prompts for generating this audio chunk. */
clientContent?: LiveMusicClientContent;
/** Music generation config for generating this audio chunk. */
musicGenerationConfig?: LiveMusicGenerationConfig;
}
/** Parameters for sending client content to the live API. */
export declare interface LiveSendClientContentParameters {
/** Client content to send to the session. */
turns?: ContentListUnion;
/** If true, indicates that the server content generation should start with
the currently accumulated prompt. Otherwise, the server will await
additional messages before starting generation. */
turnComplete?: boolean;
}
/** Parameters for sending realtime input to the live API. */
export declare interface LiveSendRealtimeInputParameters {
/** Realtime input to send to the session. */
media?: BlobImageUnion;
/** The realtime audio input stream. */
audio?: Blob_2;
/**
Indicates that the audio stream has ended, e.g. because the microphone was
turned off.
This should only be sent when automatic activity detection is enabled
(which is the default).
The client can reopen the stream by sending an audio message.
*/
audioStreamEnd?: boolean;
/** The realtime video input stream. */
video?: BlobImageUnion;
/** The realtime text input stream. */
text?: string;
/** Marks the start of user activity. */
activityStart?: ActivityStart;
/** Marks the end of user activity. */
activityEnd?: ActivityEnd;
}
/** Parameters for sending tool responses to the live API. */
export declare class LiveSendToolResponseParameters {
/** Tool responses to send to the session. */
functionResponses: FunctionResponse[] | FunctionResponse;
}
/** Incremental server update generated by the model in response to client messages.
Content is generated as quickly as possible, and not in real time. Clients
may choose to buffer and play it out in real time. */
export declare interface LiveServerContent {
/** The content that the model has generated as part of the current conversation with the user. */
modelTurn?: Content;
/** If true, indicates that the model is done generating. Generation will only start in response to additional client messages. Can be set alongside `content`, indicating that the `content` is the last in the turn. */
turnComplete?: boolean;
/** If true, indicates that a client message has interrupted current model generation. If the client is playing out the content in realtime, this is a good signal to stop and empty the current queue. */
interrupted?: boolean;
/** Metadata returned to client when grounding is enabled. */
groundingMetadata?: GroundingMetadata;
/** If true, indicates that the model is done generating. When model is
interrupted while generating there will be no generation_complete message
in interrupted turn, it will go through interrupted > turn_complete.
When model assumes realtime playback there will be delay between
generation_complete and turn_complete that is caused by model
waiting for playback to finish. If true, indicates that the model
has finished generating all content. This is a signal to the client
that it can stop sending messages. */
generationComplete?: boolean;
/** Input transcription. The transcription is independent to the model
turn which means it doesn’t imply any ordering between transcription and
model turn. */
inputTranscription?: Transcription;
/** Output transcription. The transcription is independent to the model
turn which means it doesn’t imply any ordering between transcription and
model turn.
*/
outputTranscription?: Transcription;
/** Metadata related to url context retrieval tool. */
urlContextMetadata?: UrlContextMetadata;
/** Reason for the turn is complete. */
turnCompleteReason?: TurnCompleteReason;
/** If true, indicates that the model is not generating content because
it is waiting for more input from the user, e.g. because it expects the
user to continue talking. */
waitingForInput?: boolean;
}
/** Server will not be able to service client soon. */
export declare interface LiveServerGoAway {
/** The remaining time before the connection will be terminated as ABORTED. The minimal time returned here is specified differently together with the rate limits for a given model. */
timeLeft?: string;
}
/** Response message for API call. */
export declare class LiveServerMessage {
/** Sent in response to a `LiveClientSetup` message from the client. */
setupComplete?: LiveServerSetupComplete;
/** Content generated by the model in response to client messages. */
serverContent?: LiveServerContent;
/** Request for the client to execute the `function_calls` and return the responses with the matching `id`s. */
toolCall?: LiveServerToolCall;
/** Notification for the client that a previously issued `ToolCallMessage` with the specified `id`s should have been not executed and should be cancelled. */
toolCallCancellation?: LiveServerToolCallCancellation;
/** Usage metadata about model response(s). */
usageMetadata?: UsageMetadata;
/** Server will disconnect soon. */
goAway?: LiveServerGoAway;
/** Update of the session resumption state. */
sessionResumptionUpdate?: LiveServerSessionResumptionUpdate;
/** Voice activity detection signal. Allowlisted only. */
voiceActivityDetectionSignal?: VoiceActivityDetectionSignal;
/** Voice activity signal. */
voiceActivity?: VoiceActivity;
/**
* Returns the concatenation of all text parts from the server content if present.
*
* @remarks
* If there are non-text parts in the response, the concatenation of all text
* parts will be returned, and a warning will be logged.
*/
get text(): string | undefined;
/**
* Returns the concatenation of all inline data parts from the server content if present.
*
* @remarks
* If there are non-inline data parts in the
* response, the concatenation of all inline data parts will be returned, and
* a warning will be logged.
*/
get data(): string | undefined;
}
/** Update of the session resumption state.
Only sent if `session_resumption` was set in the connection config. */
export declare interface LiveServerSessionResumptionUpdate {
/** New handle that represents state that can be resumed. Empty if `resumable`=false. */
newHandle?: string;
/** True if session can be resumed at this point. It might be not possible to resume session at some points. In that case we send update empty new_handle and resumable=false. Example of such case could be model executing function calls or just generating. Resuming session (using previous session token) in such state will result in some data loss. */
resumable?: boolean;
/** Index of last message sent by client that is included in state represented by this SessionResumptionToken. Only sent when `SessionResumptionConfig.transparent` is set.
Presence of this index allows users to transparently reconnect and avoid issue of losing some part of realtime audio input/video. If client wishes to temporarily disconnect (for example as result of receiving GoAway) they can do it without losing state by buffering messages sent since last `SessionResmumptionTokenUpdate`. This field will enable them to limit buffering (avoid keeping all requests in RAM).
Note: This should not be used for when resuming a session at some time later -- in those cases partial audio and video frames arelikely not needed. */
lastConsumedClientMessageIndex?: string;
}
/** Sent in response to a `LiveGenerateContentSetup` message from the client. */
export declare interface LiveServerSetupComplete {
/** The session id of the live session. */
sessionId?: string;
}
/** Request for the client to execute the `function_calls` and return the responses with the matching `id`s. */
export declare interface LiveServerToolCall {
/** The function call to be executed. */
functionCalls?: FunctionCall[];
}
/** Notification for the client that a previously issued `ToolCallMessage` with the specified `id`s should have been not executed and should be cancelled.
If there were side-effects to those tool calls, clients may attempt to undo
the tool calls. This message occurs only in cases where the clients interrupt
server turns. */
export declare interface LiveServerToolCallCancellation {
/** The ids of the tool calls to be cancelled. */
ids?: string[];
}
declare type LogFn = (message: string, ...rest: unknown[]) => void;
declare type Logger = {
error: LogFn;
warn: LogFn;
info: LogFn;
debug: LogFn;
};
declare type LogLevel = 'off' | 'error' | 'warn' | 'info' | 'debug';
/** The log probabilities of the tokens generated by the model. This is useful for understanding the model's confidence in its predictions and for debugging. For example, you can use log probabilities to identify when the model is making a less confident prediction or to explore alternative responses that the model considered. A low log probability can also indicate that the model is "hallucinating" or generating factually incorrect information. */
export declare interface LogprobsResult {
/** A list of the chosen candidate tokens at each decoding step. The length of this list is equal to the total number of decoding steps. Note that the chosen candidate might not be in `top_candidates`. */
chosenCandidates?: LogprobsResultCandidate[];
/** A list of the top candidate tokens at each decoding step. The length of this list is equal to the total number of decoding steps. */
topCandidates?: LogprobsResultTopCandidates[];
}
/** A single token and its associated log probability. */
export declare interface LogprobsResultCandidate {
/** The log probability of this token. A higher value indicates that the model was more confident in this token. The log probability can be used to assess the relative likelihood of different tokens and to identify when the model is uncertain. */
logProbability?: number;
/** The token's string representation. */
token?: string;
/** The token's numerical ID. While the `token` field provides the string representation of the token, the `token_id` is the numerical representation that the model uses internally. This can be useful for developers who want to build custom logic based on the model's vocabulary. */
tokenId?: number;
}
/** A list of the top candidate tokens and their log probabilities at each decoding step. This can be used to see what other tokens the model considered. */
export declare interface LogprobsResultTopCandidates {
/** The list of candidate tokens, sorted by log probability in descending order. */
candidates?: LogprobsResultCandidate[];
}
/** Configuration for a Mask reference image. */
export declare interface MaskReferenceConfig {
/** Prompts the model to generate a mask instead of you needing to
provide one (unless MASK_MODE_USER_PROVIDED is used). */
maskMode?: MaskReferenceMode;
/** A list of up to 5 class ids to use for semantic segmentation.
Automatically creates an image mask based on specific objects. */
segmentationClasses?: number[];
/** Dilation percentage of the mask provided.
Float between 0 and 1. */
maskDilation?: number;
}
/** A mask reference image.
This encapsulates either a mask image provided by the user and configs for
the user provided mask, or only config parameters for the model to generate
a mask.
A mask image is an image whose non-zero values indicate where to edit the base
image. If the user provides a mask image, the mask must be in the same
dimensions as the raw image. */
export declare class MaskReferenceImage {
/** The reference image for the editing operation. */
referenceImage?: Image_2;
/** The id of the reference image. */
referenceId?: number;
/** The type of the reference image. Only set by the SDK. */
referenceType?: string;
/** Configuration for the mask reference image. */
config?: MaskReferenceConfig;
/** Internal method to convert to ReferenceImageAPIInternal. */
toReferenceImageAPI(): ReferenceImageAPIInternal;
}
/** Enum representing the mask mode of a mask reference image. */
export declare enum MaskReferenceMode {
MASK_MODE_DEFAULT = "MASK_MODE_DEFAULT",
MASK_MODE_USER_PROVIDED = "MASK_MODE_USER_PROVIDED",
MASK_MODE_BACKGROUND = "MASK_MODE_BACKGROUND",
MASK_MODE_FOREGROUND = "MASK_MODE_FOREGROUND",
MASK_MODE_SEMANTIC = "MASK_MODE_SEMANTIC"
}
/** A MCPServer is a server that can be called by the model to perform actions. It is a server that implements the MCP protocol. Next ID: 5. This data type is not supported in Vertex AI. */
export declare interface McpServer {
/** The name of the MCPServer. */
name?: string;
/** A transport that can stream HTTP requests and responses. */
streamableHttpTransport?: StreamableHttpTransport;
}
/**
* MCPServer tool call content.
*/
declare interface MCPServerToolCallContent {
/**
* A unique ID for this specific tool call.
*/
id: string;
/**
* The JSON object of arguments for the function.
*/
arguments: {
[key: string]: unknown;
};
/**
* The name of the tool which was called.
*/
name: string;
/**
* The name of the used MCP server.
*/
server_name: string;
type: 'mcp_server_tool_call';
}
/**
* MCPServer tool result content.
*/
declare interface MCPServerToolResultContent {
/**
* ID to match the ID from the MCP server tool call block.
*/
call_id: string;
/**
* The output from the MCP server call. Can be simple text or rich content.
*/
result: unknown | Array | string;
type: 'mcp_server_tool_result';
/**
* Name of the tool which is called for this specific tool call.
*/
name?: string;
/**
* The name of the used MCP server.
*/
server_name?: string;
/**
* A signature hash for backend validation.
*/
signature?: string;
}
/**
* Creates a McpCallableTool from MCP clients and an optional config.
*
* The callable tool can invoke the MCP clients with given function call
* arguments. (often for automatic function calling).
* Use the config to modify tool parameters such as behavior.
*
* @experimental Built-in MCP support is an experimental feature, may change in future
* versions.
*/
export declare function mcpToTool(...args: [...Client[], CallableToolConfig | Client]): CallableTool;
/** Server content modalities. */
export declare enum MediaModality {
/**
* The modality is unspecified.
*/
MODALITY_UNSPECIFIED = "MODALITY_UNSPECIFIED",
/**
* Plain text.
*/
TEXT = "TEXT",
/**
* Images.
*/
IMAGE = "IMAGE",
/**
* Video.
*/
VIDEO = "VIDEO",
/**
* Audio.
*/
AUDIO = "AUDIO",
/**
* Document, e.g. PDF.
*/
DOCUMENT = "DOCUMENT"
}
/** The media resolution to use. */
export declare enum MediaResolution {
/**
* Media resolution has not been set
*/
MEDIA_RESOLUTION_UNSPECIFIED = "MEDIA_RESOLUTION_UNSPECIFIED",
/**
* Media resolution set to low (64 tokens).
*/
MEDIA_RESOLUTION_LOW = "MEDIA_RESOLUTION_LOW",
/**
* Media resolution set to medium (256 tokens).
*/
MEDIA_RESOLUTION_MEDIUM = "MEDIA_RESOLUTION_MEDIUM",
/**
* Media resolution set to high (zoomed reframing with 256 tokens).
*/
MEDIA_RESOLUTION_HIGH = "MEDIA_RESOLUTION_HIGH"
}
/**
* This type contains `RequestInit` options that may be available on the current runtime,
* including per-platform extensions like `dispatcher`, `agent`, `client`, etc.
*/
declare type MergedRequestInit = RequestInits &
/** We don't include these in the types as they'll be overridden for every request. */
Partial>;
/** Server content modalities. */
export declare enum Modality {
/**
* The modality is unspecified.
*/
MODALITY_UNSPECIFIED = "MODALITY_UNSPECIFIED",
/**
* Indicates the model should return text
*/
TEXT = "TEXT",
/**
* Indicates the model should return images.
*/
IMAGE = "IMAGE",
/**
* Indicates the model should return audio.
*/
AUDIO = "AUDIO"
}
/** Represents token counting info for a single modality. */
export declare interface ModalityTokenCount {
/** The modality associated with this token count. */
modality?: MediaModality;
/** The number of tokens counted for this modality. */
tokenCount?: number;
}
/** A trained machine learning model. */
export declare interface Model {
/** Resource name of the model. */
name?: string;
/** Display name of the model. */
displayName?: string;
/** Description of the model. */
description?: string;
/** Version ID of the model. A new version is committed when a new
model version is uploaded or trained under an existing model ID. The
version ID is an auto-incrementing decimal number in string
representation. */
version?: string;
/** List of deployed models created from this base model. Note that a
model could have been deployed to endpoints in different locations. */
endpoints?: Endpoint[];
/** Labels with user-defined metadata to organize your models. */
labels?: Record;
/** Information about the tuned model from the base model. */
tunedModelInfo?: TunedModelInfo;
/** The maximum number of input tokens that the model can handle. */
inputTokenLimit?: number;
/** The maximum number of output tokens that the model can generate. */
outputTokenLimit?: number;
/** List of actions that are supported by the model. */
supportedActions?: string[];
/** The default checkpoint id of a model version.
*/
defaultCheckpointId?: string;
/** The checkpoints of the model. */
checkpoints?: Checkpoint[];
/** Temperature value used for sampling set when the dataset was saved.
This value is used to tune the degree of randomness. */
temperature?: number;
/** The maximum temperature value used for sampling set when the
dataset was saved. This value is used to tune the degree of randomness. */
maxTemperature?: number;
/** Optional. Specifies the nucleus sampling threshold. The model
considers only the smallest set of tokens whose cumulative probability is
at least `top_p`. This helps generate more diverse and less repetitive
responses. For example, a `top_p` of 0.9 means the model considers tokens
until the cumulative probability of the tokens to select from reaches 0.9.
It's recommended to adjust either temperature or `top_p`, but not both. */
topP?: number;
/** Optional. Specifies the top-k sampling threshold. The model
considers only the top k most probable tokens for the next token. This can
be useful for generating more coherent and less random text. For example,
a `top_k` of 40 means the model will choose the next word from the 40 most
likely words. */
topK?: number;
/** Whether the model supports thinking features. If true, thoughts are
returned only if the model supports thought and thoughts are available. */
thinking?: boolean;
}
/**
* The model that will complete your prompt.\n\nSee [models](https://ai.google.dev/gemini-api/docs/models) for additional details.
*/
declare type Model_2 = 'gemini-2.5-flash' | 'gemini-2.5-flash-image' | 'gemini-2.5-flash-lite' | 'gemini-2.5-flash-lite-preview-09-2025' | 'gemini-2.5-flash-native-audio-preview-12-2025' | 'gemini-2.5-flash-preview-09-2025' | 'gemini-2.5-flash-preview-tts' | 'gemini-2.5-pro' | 'gemini-2.5-pro-preview-tts' | 'gemini-3-flash-preview' | 'gemini-3-pro-image-preview' | 'gemini-3-pro-preview' | 'gemini-3.1-pro-preview' | 'gemini-3.1-flash-image-preview' | (string & {});
/** Configuration for Model Armor. Model Armor is a Google Cloud service that provides safety and security filtering for prompts and responses. It helps protect your AI applications from risks such as harmful content, sensitive data leakage, and prompt injection attacks. This data type is not supported in Gemini API. */
export declare interface ModelArmorConfig {
/** Optional. The resource name of the Model Armor template to use for prompt screening. A Model Armor template is a set of customized filters and thresholds that define how Model Armor screens content. If specified, Model Armor will use this template to check the user's prompt for safety and security risks before it is sent to the model. The name must be in the format `projects/{project}/locations/{location}/templates/{template}`. */
promptTemplateName?: string;
/** Optional. The resource name of the Model Armor template to use for response screening. A Model Armor template is a set of customized filters and thresholds that define how Model Armor screens content. If specified, Model Armor will use this template to check the model's response for safety and security risks before it is returned to the user. The name must be in the format `projects/{project}/locations/{location}/templates/{template}`. */
responseTemplateName?: string;
}
export declare class Models extends BaseModule {
private readonly apiClient;
constructor(apiClient: ApiClient);
/**
* Calculates embeddings for the given contents.
*
* @param params - The parameters for embedding contents.
* @return The response from the API.
*
* @example
* ```ts
* const response = await ai.models.embedContent({
* model: 'text-embedding-004',
* contents: [
* 'What is your name?',
* 'What is your favorite color?',
* ],
* config: {
* outputDimensionality: 64,
* },
* });
* console.log(response);
* ```
*/
embedContent: (params: types.EmbedContentParameters) => Promise;
/**
* Makes an API request to generate content with a given model.
*
* For the `model` parameter, supported formats for Vertex AI API include:
* - The Gemini model ID, for example: 'gemini-2.0-flash'
* - The full resource name starts with 'projects/', for example:
* 'projects/my-project-id/locations/us-central1/publishers/google/models/gemini-2.0-flash'
* - The partial resource name with 'publishers/', for example:
* 'publishers/google/models/gemini-2.0-flash' or
* 'publishers/meta/models/llama-3.1-405b-instruct-maas'
* - `/` separated publisher and model name, for example:
* 'google/gemini-2.0-flash' or 'meta/llama-3.1-405b-instruct-maas'
*
* For the `model` parameter, supported formats for Gemini API include:
* - The Gemini model ID, for example: 'gemini-2.0-flash'
* - The model name starts with 'models/', for example:
* 'models/gemini-2.0-flash'
* - For tuned models, the model name starts with 'tunedModels/',
* for example:
* 'tunedModels/1234567890123456789'
*
* Some models support multimodal input and output.
*
* @param params - The parameters for generating content.
* @return The response from generating content.
*
* @example
* ```ts
* const response = await ai.models.generateContent({
* model: 'gemini-2.0-flash',
* contents: 'why is the sky blue?',
* config: {
* candidateCount: 2,
* }
* });
* console.log(response);
* ```
*/
generateContent: (params: types.GenerateContentParameters) => Promise;
/**
* This logic is needed for GenerateContentConfig only.
* Previously we made GenerateContentConfig.responseSchema field to accept
* unknown. Since v1.9.0, we switch to use backend JSON schema support.
* To maintain backward compatibility, we move the data that was treated as
* JSON schema from the responseSchema field to the responseJsonSchema field.
*/
private maybeMoveToResponseJsonSchem;
/**
* Makes an API request to generate content with a given model and yields the
* response in chunks.
*
* For the `model` parameter, supported formats for Vertex AI API include:
* - The Gemini model ID, for example: 'gemini-2.0-flash'
* - The full resource name starts with 'projects/', for example:
* 'projects/my-project-id/locations/us-central1/publishers/google/models/gemini-2.0-flash'
* - The partial resource name with 'publishers/', for example:
* 'publishers/google/models/gemini-2.0-flash' or
* 'publishers/meta/models/llama-3.1-405b-instruct-maas'
* - `/` separated publisher and model name, for example:
* 'google/gemini-2.0-flash' or 'meta/llama-3.1-405b-instruct-maas'
*
* For the `model` parameter, supported formats for Gemini API include:
* - The Gemini model ID, for example: 'gemini-2.0-flash'
* - The model name starts with 'models/', for example:
* 'models/gemini-2.0-flash'
* - For tuned models, the model name starts with 'tunedModels/',
* for example:
* 'tunedModels/1234567890123456789'
*
* Some models support multimodal input and output.
*
* @param params - The parameters for generating content with streaming response.
* @return The response from generating content.
*
* @example
* ```ts
* const response = await ai.models.generateContentStream({
* model: 'gemini-2.0-flash',
* contents: 'why is the sky blue?',
* config: {
* maxOutputTokens: 200,
* }
* });
* for await (const chunk of response) {
* console.log(chunk);
* }
* ```
*/
generateContentStream: (params: types.GenerateContentParameters) => Promise>;
/**
* Transforms the CallableTools in the parameters to be simply Tools, it
* copies the params into a new object and replaces the tools, it does not
* modify the original params. Also sets the MCP usage header if there are
* MCP tools in the parameters.
*/
private processParamsMaybeAddMcpUsage;
private initAfcToolsMap;
private processAfcStream;
/**
* Generates an image based on a text description and configuration.
*
* @param params - The parameters for generating images.
* @return The response from the API.
*
* @example
* ```ts
* const response = await client.models.generateImages({
* model: 'imagen-4.0-generate-001',
* prompt: 'Robot holding a red skateboard',
* config: {
* numberOfImages: 1,
* includeRaiReason: true,
* },
* });
* console.log(response?.generatedImages?.[0]?.image?.imageBytes);
* ```
*/
generateImages: (params: types.GenerateImagesParameters) => Promise;
list: (params?: types.ListModelsParameters) => Promise>;
/**
* Edits an image based on a prompt, list of reference images, and configuration.
*
* @param params - The parameters for editing an image.
* @return The response from the API.
*
* @example
* ```ts
* const response = await client.models.editImage({
* model: 'imagen-3.0-capability-001',
* prompt: 'Generate an image containing a mug with the product logo [1] visible on the side of the mug.',
* referenceImages: [subjectReferenceImage]
* config: {
* numberOfImages: 1,
* includeRaiReason: true,
* },
* });
* console.log(response?.generatedImages?.[0]?.image?.imageBytes);
* ```
*/
editImage: (params: types.EditImageParameters) => Promise;
/**
* Upscales an image based on an image, upscale factor, and configuration.
* Only supported in Vertex AI currently.
*
* @param params - The parameters for upscaling an image.
* @return The response from the API.
*
* @example
* ```ts
* const response = await client.models.upscaleImage({
* model: 'imagen-4.0-upscale-preview',
* image: image,
* upscaleFactor: 'x2',
* config: {
* includeRaiReason: true,
* },
* });
* console.log(response?.generatedImages?.[0]?.image?.imageBytes);
* ```
*/
upscaleImage: (params: types.UpscaleImageParameters) => Promise;
/**
* Generates videos based on a text description and configuration.
*
* @param params - The parameters for generating videos.
* @return A Promise which allows you to track the progress and eventually retrieve the generated videos using the operations.get method.
*
* @example
* ```ts
* const operation = await ai.models.generateVideos({
* model: 'veo-2.0-generate-001',
* source: {
* prompt: 'A neon hologram of a cat driving at top speed',
* },
* config: {
* numberOfVideos: 1
* });
*
* while (!operation.done) {
* await new Promise(resolve => setTimeout(resolve, 10000));
* operation = await ai.operations.getVideosOperation({operation: operation});
* }
*
* console.log(operation.response?.generatedVideos?.[0]?.video?.uri);
* ```
*/
generateVideos: (params: types.GenerateVideosParameters) => Promise;
private generateContentInternal;
private generateContentStreamInternal;
/**
* Calculates embeddings for the given contents. Only text is supported.
*
* @param params - The parameters for embedding contents.
* @return The response from the API.
*
* @example
* ```ts
* const response = await ai.models.embedContent({
* model: 'text-embedding-004',
* contents: [
* 'What is your name?',
* 'What is your favorite color?',
* ],
* config: {
* outputDimensionality: 64,
* },
* });
* console.log(response);
* ```
*/
private embedContentInternal;
/**
* Private method for generating images.
*/
private generateImagesInternal;
/**
* Private method for editing an image.
*/
private editImageInternal;
/**
* Private method for upscaling an image.
*/
private upscaleImageInternal;
/**
* Recontextualizes an image.
*
* There are two types of recontextualization currently supported:
* 1) Imagen Product Recontext - Generate images of products in new scenes
* and contexts.
* 2) Virtual Try-On: Generate images of persons modeling fashion products.
*
* @param params - The parameters for recontextualizing an image.
* @return The response from the API.
*
* @example
* ```ts
* const response1 = await ai.models.recontextImage({
* model: 'imagen-product-recontext-preview-06-30',
* source: {
* prompt: 'In a modern kitchen setting.',
* productImages: [productImage],
* },
* config: {
* numberOfImages: 1,
* },
* });
* console.log(response1?.generatedImages?.[0]?.image?.imageBytes);
*
* const response2 = await ai.models.recontextImage({
* model: 'virtual-try-on-001',
* source: {
* personImage: personImage,
* productImages: [productImage],
* },
* config: {
* numberOfImages: 1,
* },
* });
* console.log(response2?.generatedImages?.[0]?.image?.imageBytes);
* ```
*/
recontextImage(params: types.RecontextImageParameters): Promise;
/**
* Segments an image, creating a mask of a specified area.
*
* @param params - The parameters for segmenting an image.
* @return The response from the API.
*
* @example
* ```ts
* const response = await ai.models.segmentImage({
* model: 'image-segmentation-001',
* source: {
* image: image,
* },
* config: {
* mode: 'foreground',
* },
* });
* console.log(response?.generatedMasks?.[0]?.mask?.imageBytes);
* ```
*/
segmentImage(params: types.SegmentImageParameters): Promise;
/**
* Fetches information about a model by name.
*
* @example
* ```ts
* const modelInfo = await ai.models.get({model: 'gemini-2.0-flash'});
* ```
*/
get(params: types.GetModelParameters): Promise;
private listInternal;
/**
* Updates a tuned model by its name.
*
* @param params - The parameters for updating the model.
* @return The response from the API.
*
* @example
* ```ts
* const response = await ai.models.update({
* model: 'tuned-model-name',
* config: {
* displayName: 'New display name',
* description: 'New description',
* },
* });
* ```
*/
update(params: types.UpdateModelParameters): Promise;
/**
* Deletes a tuned model by its name.
*
* @param params - The parameters for deleting the model.
* @return The response from the API.
*
* @example
* ```ts
* const response = await ai.models.delete({model: 'tuned-model-name'});
* ```
*/
delete(params: types.DeleteModelParameters): Promise;
/**
* Counts the number of tokens in the given contents. Multimodal input is
* supported for Gemini models.
*
* @param params - The parameters for counting tokens.
* @return The response from the API.
*
* @example
* ```ts
* const response = await ai.models.countTokens({
* model: 'gemini-2.0-flash',
* contents: 'The quick brown fox jumps over the lazy dog.'
* });
* console.log(response);
* ```
*/
countTokens(params: types.CountTokensParameters): Promise;
/**
* Given a list of contents, returns a corresponding TokensInfo containing
* the list of tokens and list of token ids.
*
* This method is not supported by the Gemini Developer API.
*
* @param params - The parameters for computing tokens.
* @return The response from the API.
*
* @example
* ```ts
* const response = await ai.models.computeTokens({
* model: 'gemini-2.0-flash',
* contents: 'What is your name?'
* });
* console.log(response);
* ```
*/
computeTokens(params: types.ComputeTokensParameters): Promise;
/**
* Private method for generating videos.
*/
private generateVideosInternal;
}
/** Config for model selection. */
export declare interface ModelSelectionConfig {
/** Options for feature selection preference. */
featureSelectionPreference?: FeatureSelectionPreference;
}
/** Configuration for a multi-speaker text-to-speech request. */
export declare interface MultiSpeakerVoiceConfig {
/** Required. A list of configurations for the voices of the speakers. Exactly two speaker voice configurations must be provided. */
speakerVoiceConfigs?: SpeakerVoiceConfig[];
}
/** The mode of music generation. */
export declare enum MusicGenerationMode {
/**
* Rely on the server default generation mode.
*/
MUSIC_GENERATION_MODE_UNSPECIFIED = "MUSIC_GENERATION_MODE_UNSPECIFIED",
/**
* Steer text prompts to regions of latent space with higher quality
music.
*/
QUALITY = "QUALITY",
/**
* Steer text prompts to regions of latent space with a larger
diversity of music.
*/
DIVERSITY = "DIVERSITY",
/**
* Steer text prompts to regions of latent space more likely to
generate music with vocals.
*/
VOCALIZATION = "VOCALIZATION"
}
/**
* @license
* Copyright 2025 Google LLC
* SPDX-License-Identifier: Apache-2.0
*/
///
/**
* Shims for types that we can't always rely on being available globally.
*
* Note: these only exist at the type-level, there is no corresponding runtime
* version for any of these symbols.
*/
declare type NeverToAny = T extends never ? any : T;
/** @ts-ignore For users with node-fetch@2 */
declare type NodeFetch2RequestInit = NotAny | NotAny | NotAny | NotAny | NotAny | NotAny | NotAny | NotAny | NotAny | NotAny;
/** @ts-ignore For users with node-fetch@3, doesn't need file extension because types are at ./@types/index.d.ts */
declare type NodeFetch3RequestInit = NotAny | NotAny | NotAny | NotAny | NotAny | NotAny | NotAny | NotAny | NotAny | NotAny;
/** @ts-ignore */
declare type _NodeReadableStream = ReadableStream_2;
declare type NotAny = [0] extends [1 & T] ? never : T;
declare class NotFoundError extends APIError<404, Headers> {
}
/**
* @internal
* Users can pass explicit nulls to unset default headers. When we parse them
* into a standard headers type we need to preserve that information.
*/
declare type NullableHeaders = {
/** Brand check, prevent users from creating a NullableHeaders. */
[brand_privateNullableHeaders]: true;
/** Parsed headers. */
values: Headers;
/** Set of lowercase header names explicitly set to null. */
nulls: Set;
};
/** A long-running operation. */
export declare interface Operation {
/** The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. */
name?: string;
/** Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. */
metadata?: Record;
/** If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. */
done?: boolean;
/** The error result of the operation in case of failure or cancellation. */
error?: Record;
/** The response if the operation is successful. */
response?: T;
/**
* Instantiates an Operation of the same type as the one being called with the fields set from the API response.
* @internal
*/
_fromAPIResponse({ apiResponse, _isVertexAI, }: OperationFromAPIResponseParameters): Operation;
}
/** Parameters of the fromAPIResponse method of the Operation class. */
export declare interface OperationFromAPIResponseParameters {
/** The API response to be converted to an Operation. */
apiResponse: Record;
/** Whether the API response is from Vertex AI. */
_isVertexAI: boolean;
}
/** Parameters for the get method of the operations module. */
export declare interface OperationGetParameters> {
/** Used to override the default configuration. */
config?: GetOperationConfig;
/** The operation to be retrieved. */
operation: U;
}
export declare class Operations extends BaseModule {
private readonly apiClient;
constructor(apiClient: ApiClient);
/**
* Gets the status of a long-running operation.
*
* @param parameters The parameters for the get operation request.
* @return The updated Operation object, with the latest status or result.
*/
getVideosOperation(parameters: types.OperationGetParameters): Promise;
/**
* Gets the status of a long-running operation.
*
* @param parameters The parameters for the get operation request.
* @return The updated Operation object, with the latest status or result.
*/
get>(parameters: types.OperationGetParameters): Promise>;
private getVideosOperationInternal;
private fetchPredictVideosOperationInternal;
}
/** Outcome of the code execution. */
export declare enum Outcome {
/**
* Unspecified status. This value should not be used.
*/
OUTCOME_UNSPECIFIED = "OUTCOME_UNSPECIFIED",
/**
* Code execution completed successfully.
*/
OUTCOME_OK = "OUTCOME_OK",
/**
* Code execution finished but with a failure. `stderr` should contain the reason.
*/
OUTCOME_FAILED = "OUTCOME_FAILED",
/**
* Code execution ran for too long, and was cancelled. There may or may not be a partial output present.
*/
OUTCOME_DEADLINE_EXCEEDED = "OUTCOME_DEADLINE_EXCEEDED"
}
/** Describes the info for output of EvaluationService. This data type is not supported in Gemini API. */
export declare interface OutputInfo {
/** Output only. The full path of the Cloud Storage directory created, into which the evaluation results and aggregation results are written. */
gcsOutputDirectory?: string;
}
/**
* Some environments overload the global fetch function, and Parameters only gets the last signature.
*/
declare type OverloadedParameters = T extends ({
(...args: infer A): unknown;
(...args: infer B): unknown;
(...args: infer C): unknown;
(...args: infer D): unknown;
}) ? A | B | C | D : T extends ({
(...args: infer A): unknown;
(...args: infer B): unknown;
(...args: infer C): unknown;
}) ? A | B | C : T extends ({
(...args: infer A): unknown;
(...args: infer B): unknown;
}) ? A | B : T extends (...args: infer A) => unknown ? A : never;
export declare enum PagedItem {
PAGED_ITEM_BATCH_JOBS = "batchJobs",
PAGED_ITEM_MODELS = "models",
PAGED_ITEM_TUNING_JOBS = "tuningJobs",
PAGED_ITEM_FILES = "files",
PAGED_ITEM_CACHED_CONTENTS = "cachedContents",
PAGED_ITEM_FILE_SEARCH_STORES = "fileSearchStores",
PAGED_ITEM_DOCUMENTS = "documents"
}
declare interface PagedItemConfig {
parent?: string;
config?: {
pageToken?: string;
pageSize?: number;
};
}
declare interface PagedItemResponse {
nextPageToken?: string;
sdkHttpResponse?: types.HttpResponse;
batchJobs?: T[];
models?: T[];
tuningJobs?: T[];
files?: T[];
cachedContents?: T[];
fileSearchStores?: T[];
documents?: T[];
}
/**
* Pager class for iterating through paginated results.
*/
export declare class Pager implements AsyncIterable {
private nameInternal;
private pageInternal;
private paramsInternal;
private pageInternalSize;
private sdkHttpResponseInternal?;
protected requestInternal: (params: PagedItemConfig) => Promise>;
protected idxInternal: number;
constructor(name: PagedItem, request: (params: PagedItemConfig) => Promise>, response: PagedItemResponse, params: PagedItemConfig);
private init;
private initNextPage;
/**
* Returns the current page, which is a list of items.
*
* @remarks
* The first page is retrieved when the pager is created. The returned list of
* items could be a subset of the entire list.
*/
get page(): T[];
/**
* Returns the type of paged item (for example, ``batch_jobs``).
*/
get name(): PagedItem;
/**
* Returns the length of the page fetched each time by this pager.
*
* @remarks
* The number of items in the page is less than or equal to the page length.
*/
get pageSize(): number;
/**
* Returns the headers of the API response.
*/
get sdkHttpResponse(): types.HttpResponse | undefined;
/**
* Returns the parameters when making the API request for the next page.
*
* @remarks
* Parameters contain a set of optional configs that can be
* used to customize the API request. For example, the `pageToken` parameter
* contains the token to request the next page.
*/
get params(): PagedItemConfig;
/**
* Returns the total number of items in the current page.
*/
get pageLength(): number;
/**
* Returns the item at the given index.
*/
getItem(index: number): T;
/**
* Returns an async iterator that support iterating through all items
* retrieved from the API.
*
* @remarks
* The iterator will automatically fetch the next page if there are more items
* to fetch from the API.
*
* @example
*
* ```ts
* const pager = await ai.files.list({config: {pageSize: 10}});
* for await (const file of pager) {
* console.log(file.name);
* }
* ```
*/
[Symbol.asyncIterator](): AsyncIterator;
/**
* Fetches the next page of items. This makes a new API request.
*
* @throws {Error} If there are no more pages to fetch.
*
* @example
*
* ```ts
* const pager = await ai.files.list({config: {pageSize: 10}});
* let page = pager.page;
* while (true) {
* for (const file of page) {
* console.log(file.name);
* }
* if (!pager.hasNextPage()) {
* break;
* }
* page = await pager.nextPage();
* }
* ```
*/
nextPage(): Promise;
/**
* Returns true if there are more pages to fetch from the API.
*/
hasNextPage(): boolean;
}
/** Output only. Pairwise metric choice. This enum is not supported in Gemini API. */
export declare enum PairwiseChoice {
/**
* Unspecified prediction choice.
*/
PAIRWISE_CHOICE_UNSPECIFIED = "PAIRWISE_CHOICE_UNSPECIFIED",
/**
* Baseline prediction wins
*/
BASELINE = "BASELINE",
/**
* Candidate prediction wins
*/
CANDIDATE = "CANDIDATE",
/**
* Winner cannot be determined
*/
TIE = "TIE"
}
/** Spec for pairwise metric result. This data type is not supported in Gemini API. */
export declare interface PairwiseMetricResult {
/** Output only. Spec for custom output. */
customOutput?: CustomOutput;
/** Output only. Explanation for pairwise metric score. */
explanation?: string;
/** Output only. Pairwise metric choice. */
pairwiseChoice?: PairwiseChoice;
}
/** A datatype containing media content.
Exactly one field within a Part should be set, representing the specific type
of content being conveyed. Using multiple fields within the same `Part`
instance is considered invalid. */
export declare interface Part {
/** Media resolution for the input media.
*/
mediaResolution?: PartMediaResolution;
/** Optional. The result of executing the ExecutableCode. */
codeExecutionResult?: CodeExecutionResult;
/** Optional. Code generated by the model that is intended to be executed. */
executableCode?: ExecutableCode;
/** Optional. The URI-based data of the part. This can be used to include files from Google Cloud Storage. */
fileData?: FileData;
/** Optional. A predicted function call returned from the model. This contains the name of the function to call and the arguments to pass to the function. */
functionCall?: FunctionCall;
/** Optional. The result of a function call. This is used to provide the model with the result of a function call that it predicted. */
functionResponse?: FunctionResponse;
/** Optional. The inline data content of the part. This can be used to include images, audio, or video in a request. */
inlineData?: Blob_2;
/** Optional. The text content of the part. When sent from the VSCode Gemini Code Assist extension, references to @mentioned items will be converted to markdown boldface text. For example `@my-repo` will be converted to and sent as `**my-repo**` by the IDE agent. */
text?: string;
/** Optional. Indicates whether the `part` represents the model's thought process or reasoning. */
thought?: boolean;
/** Optional. An opaque signature for the thought so it can be reused in subsequent requests.
* @remarks Encoded as base64 string. */
thoughtSignature?: string;
/** Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. */
videoMetadata?: VideoMetadata;
}
/** Partial argument value of the function call. This data type is not supported in Gemini API. */
export declare interface PartialArg {
/** Optional. Represents a boolean value. */
boolValue?: boolean;
/** Required. A JSON Path (RFC 9535) to the argument being streamed. https://datatracker.ietf.org/doc/html/rfc9535. e.g. "$.foo.bar[0].data". */
jsonPath?: string;
/** Optional. Represents a null value. */
nullValue?: 'NULL_VALUE';
/** Optional. Represents a double value. */
numberValue?: number;
/** Optional. Represents a string value. */
stringValue?: string;
/** Optional. Whether this is not the last part of the same json_path. If true, another PartialArg message for the current json_path is expected to follow. */
willContinue?: boolean;
}
export declare type PartListUnion = PartUnion[] | PartUnion;
/** Media resolution for the input media. */
export declare interface PartMediaResolution {
/** The tokenization quality used for given media.
*/
level?: PartMediaResolutionLevel;
/** Specifies the required sequence length for media tokenization.
*/
numTokens?: number;
}
/** The tokenization quality used for given media. */
export declare enum PartMediaResolutionLevel {
/**
* Media resolution has not been set.
*/
MEDIA_RESOLUTION_UNSPECIFIED = "MEDIA_RESOLUTION_UNSPECIFIED",
/**
* Media resolution set to low.
*/
MEDIA_RESOLUTION_LOW = "MEDIA_RESOLUTION_LOW",
/**
* Media resolution set to medium.
*/
MEDIA_RESOLUTION_MEDIUM = "MEDIA_RESOLUTION_MEDIUM",
/**
* Media resolution set to high.
*/
MEDIA_RESOLUTION_HIGH = "MEDIA_RESOLUTION_HIGH",
/**
* Media resolution set to ultra high.
*/
MEDIA_RESOLUTION_ULTRA_HIGH = "MEDIA_RESOLUTION_ULTRA_HIGH"
}
/** Tuning spec for Partner models. This data type is not supported in Gemini API. */
export declare interface PartnerModelTuningSpec {
/** Hyperparameters for tuning. The accepted hyper_parameters and their valid range of values will differ depending on the base model. */
hyperParameters?: Record;
/** Required. Cloud Storage path to file containing training dataset for tuning. The dataset must be formatted as a JSONL file. */
trainingDatasetUri?: string;
/** Optional. Cloud Storage path to file containing validation dataset for tuning. The dataset must be formatted as a JSONL file. */
validationDatasetUri?: string;
}
export declare type PartUnion = Part | string;
declare class PermissionDeniedError extends APIError<403, Headers> {
}
/** Enum that controls the generation of people. */
export declare enum PersonGeneration {
/**
* Block generation of images of people.
*/
DONT_ALLOW = "DONT_ALLOW",
/**
* Generate images of adults, but not children.
*/
ALLOW_ADULT = "ALLOW_ADULT",
/**
* Generate images that include adults and children.
*/
ALLOW_ALL = "ALLOW_ALL"
}
/** Sites with confidence level chosen & above this value will be blocked from the search results. This enum is not supported in Gemini API. */
export declare enum PhishBlockThreshold {
/**
* Defaults to unspecified.
*/
PHISH_BLOCK_THRESHOLD_UNSPECIFIED = "PHISH_BLOCK_THRESHOLD_UNSPECIFIED",
/**
* Blocks Low and above confidence URL that is risky.
*/
BLOCK_LOW_AND_ABOVE = "BLOCK_LOW_AND_ABOVE",
/**
* Blocks Medium and above confidence URL that is risky.
*/
BLOCK_MEDIUM_AND_ABOVE = "BLOCK_MEDIUM_AND_ABOVE",
/**
* Blocks High and above confidence URL that is risky.
*/
BLOCK_HIGH_AND_ABOVE = "BLOCK_HIGH_AND_ABOVE",
/**
* Blocks Higher and above confidence URL that is risky.
*/
BLOCK_HIGHER_AND_ABOVE = "BLOCK_HIGHER_AND_ABOVE",
/**
* Blocks Very high and above confidence URL that is risky.
*/
BLOCK_VERY_HIGH_AND_ABOVE = "BLOCK_VERY_HIGH_AND_ABOVE",
/**
* Blocks Extremely high confidence URL that is risky.
*/
BLOCK_ONLY_EXTREMELY_HIGH = "BLOCK_ONLY_EXTREMELY_HIGH"
}
/** Spec for pointwise metric result. This data type is not supported in Gemini API. */
export declare interface PointwiseMetricResult {
/** Output only. Spec for custom output. */
customOutput?: CustomOutput;
/** Output only. Explanation for pointwise metric score. */
explanation?: string;
/** Output only. Pointwise metric score. */
score?: number;
}
/** Configuration for a prebuilt voice. */
export declare interface PrebuiltVoiceConfig {
/** The name of the prebuilt voice to use. */
voiceName?: string;
}
/** Statistics computed for datasets used for preference optimization. This data type is not supported in Gemini API. */
export declare interface PreferenceOptimizationDataStats {
/** Output only. A partial sample of the indices (starting from 1) of the dropped examples. */
droppedExampleIndices?: string[];
/** Output only. For each index in `dropped_example_indices`, the user-facing reason why the example was dropped. */
droppedExampleReasons?: string[];
/** Output only. Dataset distributions for scores variance per example. */
scoreVariancePerExampleDistribution?: DatasetDistribution;
/** Output only. Dataset distributions for scores. */
scoresDistribution?: DatasetDistribution;
/** Output only. Number of billable tokens in the tuning dataset. */
totalBillableTokenCount?: string;
/** Output only. Number of examples in the tuning dataset. */
tuningDatasetExampleCount?: string;
/** Output only. Number of tuning steps for this Tuning Job. */
tuningStepCount?: string;
/** Output only. Sample user examples in the training dataset. */
userDatasetExamples?: GeminiPreferenceExample[];
/** Output only. Dataset distributions for the user input tokens. */
userInputTokenDistribution?: DatasetDistribution;
/** Output only. Dataset distributions for the user output tokens. */
userOutputTokenDistribution?: DatasetDistribution;
}
/** Hyperparameters for Preference Optimization. This data type is not supported in Gemini API. */
export declare interface PreferenceOptimizationHyperParameters {
/** Optional. Adapter size for preference optimization. */
adapterSize?: AdapterSize;
/** Optional. Weight for KL Divergence regularization. */
beta?: number;
/** Optional. Number of complete passes the model makes over the entire training dataset during training. */
epochCount?: string;
/** Optional. Multiplier for adjusting the default learning rate. */
learningRateMultiplier?: number;
}
/** Preference optimization tuning spec for tuning. */
export declare interface PreferenceOptimizationSpec {
/** Optional. If set to true, disable intermediate checkpoints for Preference Optimization and only the last checkpoint will be exported. Otherwise, enable intermediate checkpoints for Preference Optimization. Default is false. */
exportLastCheckpointOnly?: boolean;
/** Optional. Hyperparameters for Preference Optimization. */
hyperParameters?: PreferenceOptimizationHyperParameters;
/** Required. Cloud Storage path to file containing training dataset for preference optimization tuning. The dataset must be formatted as a JSONL file. */
trainingDatasetUri?: string;
/** Optional. Cloud Storage path to file containing validation dataset for preference optimization tuning. The dataset must be formatted as a JSONL file. */
validationDatasetUri?: string;
}
/** A pre-tuned model for continuous tuning. This data type is not supported in Gemini API. */
export declare interface PreTunedModel {
/** Output only. The name of the base model this PreTunedModel was tuned from. */
baseModel?: string;
/** Optional. The source checkpoint id. If not specified, the default checkpoint will be used. */
checkpointId?: string;
/** The resource name of the Model. E.g., a model resource name with a specified version id or alias: `projects/{project}/locations/{location}/models/{model}@{version_id}` `projects/{project}/locations/{location}/models/{model}@{alias}` Or, omit the version id to use the default version: `projects/{project}/locations/{location}/models/{model}` */
tunedModelName?: string;
}
/** Config for proactivity features. */
export declare interface ProactivityConfig {
/** If enabled, the model can reject responding to the last prompt. For
example, this allows the model to ignore out of context speech or to stay
silent if the user did not make a request, yet. */
proactiveAudio?: boolean;
}
/** An image of the product. */
export declare interface ProductImage {
/** An image of the product to be recontextualized. */
productImage?: Image_2;
}
/** Enum for controlling whether the model can generate images of prominent people (celebrities). */
export declare enum ProminentPeople {
/**
* Unspecified value. The model will proceed with the default behavior, which is to allow generation of prominent people.
*/
PROMINENT_PEOPLE_UNSPECIFIED = "PROMINENT_PEOPLE_UNSPECIFIED",
/**
* Allows the model to generate images of prominent people.
*/
ALLOW_PROMINENT_PEOPLE = "ALLOW_PROMINENT_PEOPLE",
/**
* Prevents the model from generating images of prominent people.
*/
BLOCK_PROMINENT_PEOPLE = "BLOCK_PROMINENT_PEOPLE"
}
/**
* @license
* Copyright 2025 Google LLC
* SPDX-License-Identifier: Apache-2.0
*/
declare type PromiseOrValue = T | Promise;
/** A RagChunk includes the content of a chunk of a RagFile, and associated metadata. This data type is not supported in Gemini API. */
export declare interface RagChunk {
/** If populated, represents where the chunk starts and ends in the document. */
pageSpan?: RagChunkPageSpan;
/** The content of the chunk. */
text?: string;
}
/** Represents where the chunk starts and ends in the document. This data type is not supported in Gemini API. */
export declare interface RagChunkPageSpan {
/** Page where chunk starts in the document. Inclusive. 1-indexed. */
firstPage?: number;
/** Page where chunk ends in the document. Inclusive. 1-indexed. */
lastPage?: number;
}
/** Specifies the context retrieval config. This data type is not supported in Gemini API. */
export declare interface RagRetrievalConfig {
/** Optional. Config for filters. */
filter?: RagRetrievalConfigFilter;
/** Optional. Config for Hybrid Search. */
hybridSearch?: RagRetrievalConfigHybridSearch;
/** Optional. Config for ranking and reranking. */
ranking?: RagRetrievalConfigRanking;
/** Optional. The number of contexts to retrieve. */
topK?: number;
}
/** Config for filters. This data type is not supported in Gemini API. */
export declare interface RagRetrievalConfigFilter {
/** Optional. String for metadata filtering. */
metadataFilter?: string;
/** Optional. Only returns contexts with vector distance smaller than the threshold. */
vectorDistanceThreshold?: number;
/** Optional. Only returns contexts with vector similarity larger than the threshold. */
vectorSimilarityThreshold?: number;
}
/** Config for Hybrid Search. This data type is not supported in Gemini API. */
export declare interface RagRetrievalConfigHybridSearch {
/** Optional. Alpha value controls the weight between dense and sparse vector search results. The range is [0, 1], while 0 means sparse vector search only and 1 means dense vector search only. The default value is 0.5 which balances sparse and dense vector search equally. */
alpha?: number;
}
/** Config for ranking and reranking. This data type is not supported in Gemini API. */
export declare interface RagRetrievalConfigRanking {
/** Optional. Config for LlmRanker. */
llmRanker?: RagRetrievalConfigRankingLlmRanker;
/** Optional. Config for Rank Service. */
rankService?: RagRetrievalConfigRankingRankService;
}
/** Config for LlmRanker. This data type is not supported in Gemini API. */
export declare interface RagRetrievalConfigRankingLlmRanker {
/** Optional. The model name used for ranking. See [Supported models](https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/inference#supported-models). */
modelName?: string;
}
/** Config for Rank Service. This data type is not supported in Gemini API. */
export declare interface RagRetrievalConfigRankingRankService {
/** Optional. The model name of the rank service. Format: `semantic-ranker-512@latest` */
modelName?: string;
}
declare class RateLimitError extends APIError<429, Headers> {
}
/** Raw output. This data type is not supported in Gemini API. */
export declare interface RawOutput {
/** Output only. Raw output string. */
rawOutput?: string[];
}
/** A raw reference image.
A raw reference image represents the base image to edit, provided by the user.
It can optionally be provided in addition to a mask reference image or
a style reference image. */
export declare class RawReferenceImage {
/** The reference image for the editing operation. */
referenceImage?: Image_2;
/** The id of the reference image. */
referenceId?: number;
/** The type of the reference image. Only set by the SDK. */
referenceType?: string;
/** Internal method to convert to ReferenceImageAPIInternal. */
toReferenceImageAPI(): ReferenceImageAPIInternal;
}
declare type _ReadableStream_2 = NeverToAny<([0] extends [1 & _DOMReadableStream] ? never : _DOMReadableStream) | ([0] extends [1 & _ConditionalNodeReadableStream] ? never : _ConditionalNodeReadableStream)>;
/** Marks the end of user activity.
This can only be sent if automatic (i.e. server-side) activity detection is
disabled. */
export declare interface RealtimeInputConfig {
/** If not set, automatic activity detection is enabled by default. If automatic voice detection is disabled, the client must send activity signals. */
automaticActivityDetection?: AutomaticActivityDetection;
/** Defines what effect activity has. */
activityHandling?: ActivityHandling;
/** Defines which input is included in the user's turn. */
turnCoverage?: TurnCoverage;
}
/** Configuration for recontextualizing an image. */
export declare interface RecontextImageConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
/** Number of images to generate. */
numberOfImages?: number;
/** The number of sampling steps. A higher value has better image
quality, while a lower value has better latency. */
baseSteps?: number;
/** Cloud Storage URI used to store the generated images. */
outputGcsUri?: string;
/** Random seed for image generation. */
seed?: number;
/** Filter level for safety filtering. */
safetyFilterLevel?: SafetyFilterLevel;
/** Whether allow to generate person images, and restrict to specific
ages. */
personGeneration?: PersonGeneration;
/** Whether to add a SynthID watermark to the generated images. */
addWatermark?: boolean;
/** MIME type of the generated image. */
outputMimeType?: string;
/** Compression quality of the generated image (for ``image/jpeg``
only). */
outputCompressionQuality?: number;
/** Whether to use the prompt rewriting logic. */
enhancePrompt?: boolean;
/** User specified labels to track billing usage. */
labels?: Record;
}
/** The parameters for recontextualizing an image. */
export declare interface RecontextImageParameters {
/** ID of the model to use. For a list of models, see `Google models
`_. */
model: string;
/** A set of source input(s) for image recontextualization. */
source: RecontextImageSource;
/** Configuration for image recontextualization. */
config?: RecontextImageConfig;
}
/** The output images response. */
export declare class RecontextImageResponse {
/** List of generated images. */
generatedImages?: GeneratedImage[];
}
/** A set of source input(s) for image recontextualization. */
export declare interface RecontextImageSource {
/** A text prompt for guiding the model during image
recontextualization. Not supported for Virtual Try-On. */
prompt?: string;
/** Image of the person or subject who will be wearing the
product(s). */
personImage?: Image_2;
/** A list of product images. */
productImages?: ProductImage[];
}
export declare type ReferenceImage = RawReferenceImage | MaskReferenceImage | ControlReferenceImage | StyleReferenceImage | SubjectReferenceImage | ContentReferenceImage;
/** Private class that represents a Reference image that is sent to API. */
declare interface ReferenceImageAPIInternal {
/** The reference image for the editing operation. */
referenceImage?: types.Image;
/** The id of the reference image. */
referenceId?: number;
/** The type of the reference image. Only set by the SDK. */
referenceType?: string;
/** Configuration for the mask reference image. */
maskImageConfig?: types.MaskReferenceConfig;
/** Configuration for the control reference image. */
controlImageConfig?: types.ControlReferenceConfig;
/** Configuration for the style reference image. */
styleImageConfig?: types.StyleReferenceConfig;
/** Configuration for the subject reference image. */
subjectImageConfig?: types.SubjectReferenceConfig;
}
/** Used to override the default configuration. */
export declare interface RegisterFilesConfig {
/** Used to override HTTP request options. */
httpOptions?: HttpOptions;
/** Abort signal which can be used to cancel the request.
NOTE: AbortSignal is a client-only operation. Using it to cancel an
operation will not cancel the request in the service. You will still
be charged usage for any applicable operations.
*/
abortSignal?: AbortSignal;
}
/** Generates the parameters for the private registerFiles method. */
export declare interface RegisterFilesParameters {
/**
* The authentication object.
*/
auth?: any;
/**
* The Google Cloud Storage URIs to register. Example: `gs://bucket/object`.
*/
uris: string[];
/** Used to override the default configuration. */
config?: RegisterFilesConfig;
}
/** Response for the _register file method. */
export declare class RegisterFilesResponse {
/** Used to retain the full HTTP response. */
sdkHttpResponse?: HttpResponse;
/** The registered files. */
files?: File_2[];
}
/** Represents a recorded session. */
export declare interface ReplayFile {
replayId?: string;
interactions?: ReplayInteraction[];
}
/** Represents a single interaction, request and response in a replay. */
export declare interface ReplayInteraction {
request?: ReplayRequest;
response?: ReplayResponse;
}
/** Represents a single request in a replay. */
export declare interface ReplayRequest {
method?: string;
url?: string;
headers?: Record;
bodySegments?: Record[];
}
/** Represents a single response in a replay. */
export declare class ReplayResponse {
statusCode?: number;
headers?: Record;
bodySegments?: Record[];
sdkResponseSegments?: Record[];
}
/** ReplicatedVoiceConfig is used to configure replicated voice. */
export declare interface ReplicatedVoiceConfig {
/** The mime type of the replicated voice.
*/
mimeType?: string;
/** The sample audio of the replicated voice.
* @remarks Encoded as base64 string. */
voiceSampleAudio?: string;
}
/**
* The type for the first argument to `fetch`.
*
* https://developer.mozilla.org/docs/Web/API/Window/fetch#resource
*/
declare type _RequestInfo = Request | URL | string;
/**
* An alias to the builtin `RequestInit` type so we can
* easily alias it in import statements if there are name clashes.
*
* https://developer.mozilla.org/docs/Web/API/RequestInit
*/
declare type _RequestInit = RequestInit;
declare type RequestInits = NotAny