/** * AI SDK Telemetry Attributes * Based on https://ai-sdk.dev/docs/ai-sdk-core/telemetry#collected-data */ /** * Common attribute for operation name across all functions and spans * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#collected-data */ export declare const OPERATION_NAME_ATTRIBUTE = "operation.name"; /** * Common attribute for AI operation ID across all functions and spans * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#collected-data */ export declare const AI_OPERATION_ID_ATTRIBUTE = "ai.operationId"; /** * `generateText` function - `ai.generateText` span * `streamText` function - `ai.streamText` span * * The prompt that was used when calling the function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generatetext-function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamtext-function */ export declare const AI_PROMPT_ATTRIBUTE = "ai.prompt"; /** * `generateObject` function - `ai.generateObject` span * `streamObject` function - `ai.streamObject` span * * The JSON schema version of the schema that was passed into the function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generateobject-function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamobject-function */ export declare const AI_SCHEMA_ATTRIBUTE = "ai.schema"; /** * `generateObject` function - `ai.generateObject` span * `streamObject` function - `ai.streamObject` span * * The name of the schema that was passed into the function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generateobject-function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamobject-function */ export declare const AI_SCHEMA_NAME_ATTRIBUTE = "ai.schema.name"; /** * `generateObject` function - `ai.generateObject` span * `streamObject` function - `ai.streamObject` span * * The description of the schema that was passed into the function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generateobject-function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamobject-function */ export declare const AI_SCHEMA_DESCRIPTION_ATTRIBUTE = "ai.schema.description"; /** * `generateObject` function - `ai.generateObject` span * `streamObject` function - `ai.streamObject` span * * The object that was generated (stringified JSON) * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generateobject-function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamobject-function */ export declare const AI_RESPONSE_OBJECT_ATTRIBUTE = "ai.response.object"; /** * `generateObject` function - `ai.generateObject` span * `streamObject` function - `ai.streamObject` span * * The object generation mode, e.g. `json` * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generateobject-function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamobject-function */ export declare const AI_SETTINGS_MODE_ATTRIBUTE = "ai.settings.mode"; /** * `generateObject` function - `ai.generateObject` span * `streamObject` function - `ai.streamObject` span * * The output type that was used, e.g. `object` or `no-schema` * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generateobject-function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamobject-function */ export declare const AI_SETTINGS_OUTPUT_ATTRIBUTE = "ai.settings.output"; /** * `embed` function - `ai.embed.doEmbed` span * `embedMany` function - `ai.embedMany` span * * The values that were passed into the function (array) * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#embed-function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#embedmany-function */ export declare const AI_VALUES_ATTRIBUTE = "ai.values"; /** * `embed` function - `ai.embed.doEmbed` span * `embedMany` function - `ai.embedMany` span * * An array of JSON-stringified embeddings * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#embed-function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#embedmany-function */ export declare const AI_EMBEDDINGS_ATTRIBUTE = "ai.embeddings"; /** * `generateText` function - `ai.generateText` span * * The text that was generated * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generatetext-function */ export declare const AI_RESPONSE_TEXT_ATTRIBUTE = "ai.response.text"; /** * `generateText` function - `ai.generateText` span * * The tool calls that were made as part of the generation (stringified JSON) * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generatetext-function */ export declare const AI_RESPONSE_TOOL_CALLS_ATTRIBUTE = "ai.response.toolCalls"; /** * `generateText` function - `ai.generateText` span * * The reason why the generation finished * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generatetext-function */ export declare const AI_RESPONSE_FINISH_REASON_ATTRIBUTE = "ai.response.finishReason"; /** * `generateText` function - `ai.generateText` span * * The maximum number of steps that were set * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generatetext-function */ export declare const AI_SETTINGS_MAX_STEPS_ATTRIBUTE = "ai.settings.maxSteps"; /** * `generateText` function - `ai.generateText.doGenerate` span * * The format of the prompt * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generatetext-function */ export declare const AI_PROMPT_FORMAT_ATTRIBUTE = "ai.prompt.format"; /** * `generateText` function - `ai.generateText.doGenerate` span * * The messages that were passed into the provider * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generatetext-function */ export declare const AI_PROMPT_MESSAGES_ATTRIBUTE = "ai.prompt.messages"; /** * `generateText` function - `ai.generateText.doGenerate` span * * Array of stringified tool definitions * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generatetext-function */ export declare const AI_PROMPT_TOOLS_ATTRIBUTE = "ai.prompt.tools"; /** * `generateText` function - `ai.generateText.doGenerate` span * * The stringified tool choice setting (JSON) * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generatetext-function */ export declare const AI_PROMPT_TOOL_CHOICE_ATTRIBUTE = "ai.prompt.toolChoice"; /** * `streamText` function - `ai.streamText.doStream` span * * The time it took to receive the first chunk in milliseconds * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamtext-function */ export declare const AI_RESPONSE_MS_TO_FIRST_CHUNK_ATTRIBUTE = "ai.response.msToFirstChunk"; /** * `streamText` function - `ai.streamText.doStream` span * * The time it took to receive the finish part of the LLM stream in milliseconds * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamtext-function */ export declare const AI_RESPONSE_MS_TO_FINISH_ATTRIBUTE = "ai.response.msToFinish"; /** * `streamText` function - `ai.streamText.doStream` span * * The average completion tokens per second * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamtext-function */ export declare const AI_RESPONSE_AVG_COMPLETION_TOKENS_PER_SECOND_ATTRIBUTE = "ai.response.avgCompletionTokensPerSecond"; /** * `embed` function - `ai.embed` span * * The value that was passed into the `embed` function * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#embed-function */ export declare const AI_VALUE_ATTRIBUTE = "ai.value"; /** * `embed` function - `ai.embed` span * * A JSON-stringified embedding * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#embed-function */ export declare const AI_EMBEDDING_ATTRIBUTE = "ai.embedding"; /** * Basic LLM span information * Multiple spans * * The functionId that was set through `telemetry.functionId` * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#basic-llm-span-information */ export declare const RESOURCE_NAME_ATTRIBUTE = "resource.name"; /** * Basic LLM span information * Multiple spans * * The id of the model * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#basic-llm-span-information */ export declare const AI_MODEL_ID_ATTRIBUTE = "ai.model.id"; /** * Basic LLM span information * Multiple spans * * The provider of the model * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#basic-llm-span-information */ export declare const AI_MODEL_PROVIDER_ATTRIBUTE = "ai.model.provider"; /** * Basic LLM span information * Multiple spans * * The request headers that were passed in through `headers` * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#basic-llm-span-information */ export declare const AI_REQUEST_HEADERS_ATTRIBUTE = "ai.request.headers"; /** * Basic LLM span information * Multiple spans * * Provider specific metadata returned with the generation response * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#basic-llm-span-information */ export declare const AI_RESPONSE_PROVIDER_METADATA_ATTRIBUTE = "ai.response.providerMetadata"; /** * Basic LLM span information * Multiple spans * * The maximum number of retries that were set * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#basic-llm-span-information */ export declare const AI_SETTINGS_MAX_RETRIES_ATTRIBUTE = "ai.settings.maxRetries"; /** * Basic LLM span information * Multiple spans * * The functionId that was set through `telemetry.functionId` * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#basic-llm-span-information */ export declare const AI_TELEMETRY_FUNCTION_ID_ATTRIBUTE = "ai.telemetry.functionId"; /** * Basic LLM span information * Multiple spans * * The metadata that was passed in through `telemetry.metadata` * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#basic-llm-span-information */ export declare const AI_TELEMETRY_METADATA_ATTRIBUTE = "ai.telemetry.metadata"; /** * Basic LLM span information * Multiple spans * * The number of completion tokens that were used * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#basic-llm-span-information */ export declare const AI_USAGE_COMPLETION_TOKENS_ATTRIBUTE = "ai.usage.completionTokens"; /** * Basic LLM span information * Multiple spans * * The number of prompt tokens that were used * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#basic-llm-span-information */ export declare const AI_USAGE_PROMPT_TOKENS_ATTRIBUTE = "ai.usage.promptTokens"; /** * Call LLM span information * Individual LLM call spans * * The model that was used to generate the response * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const AI_RESPONSE_MODEL_ATTRIBUTE = "ai.response.model"; /** * Call LLM span information * Individual LLM call spans * * The id of the response * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const AI_RESPONSE_ID_ATTRIBUTE = "ai.response.id"; /** * Call LLM span information * Individual LLM call spans * * The timestamp of the response * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const AI_RESPONSE_TIMESTAMP_ATTRIBUTE = "ai.response.timestamp"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The provider that was used * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_SYSTEM_ATTRIBUTE = "gen_ai.system"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The model that was requested * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_REQUEST_MODEL_ATTRIBUTE = "gen_ai.request.model"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The temperature that was set * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_REQUEST_TEMPERATURE_ATTRIBUTE = "gen_ai.request.temperature"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The maximum number of tokens that were set * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_REQUEST_MAX_TOKENS_ATTRIBUTE = "gen_ai.request.max_tokens"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The frequency penalty that was set * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_REQUEST_FREQUENCY_PENALTY_ATTRIBUTE = "gen_ai.request.frequency_penalty"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The presence penalty that was set * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_REQUEST_PRESENCE_PENALTY_ATTRIBUTE = "gen_ai.request.presence_penalty"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The topK parameter value that was set * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_REQUEST_TOP_K_ATTRIBUTE = "gen_ai.request.top_k"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The topP parameter value that was set * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_REQUEST_TOP_P_ATTRIBUTE = "gen_ai.request.top_p"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The stop sequences * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_REQUEST_STOP_SEQUENCES_ATTRIBUTE = "gen_ai.request.stop_sequences"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The finish reasons that were returned by the provider * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_RESPONSE_FINISH_REASONS_ATTRIBUTE = "gen_ai.response.finish_reasons"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The model that was used to generate the response * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_RESPONSE_MODEL_ATTRIBUTE = "gen_ai.response.model"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The id of the response * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_RESPONSE_ID_ATTRIBUTE = "gen_ai.response.id"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The number of prompt tokens that were used * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_USAGE_INPUT_TOKENS_ATTRIBUTE = "gen_ai.usage.input_tokens"; /** * Semantic Conventions for GenAI operations * Individual LLM call spans * * The number of completion tokens that were used * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#call-llm-span-information */ export declare const GEN_AI_USAGE_OUTPUT_TOKENS_ATTRIBUTE = "gen_ai.usage.output_tokens"; /** * Basic embedding span information * Embedding spans * * The number of tokens that were used * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#basic-embedding-span-information */ export declare const AI_USAGE_TOKENS_ATTRIBUTE = "ai.usage.tokens"; /** * Tool call spans * `ai.toolCall` span * * The name of the tool * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#tool-call-spans */ export declare const AI_TOOL_CALL_NAME_ATTRIBUTE = "ai.toolCall.name"; /** * Tool call spans * `ai.toolCall` span * * The id of the tool call * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#tool-call-spans */ export declare const AI_TOOL_CALL_ID_ATTRIBUTE = "ai.toolCall.id"; /** * Tool call spans * `ai.toolCall` span * * The parameters of the tool call * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#tool-call-spans */ export declare const AI_TOOL_CALL_ARGS_ATTRIBUTE = "ai.toolCall.args"; /** * Tool call spans * `ai.toolCall` span * * The result of the tool call * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#tool-call-spans */ export declare const AI_TOOL_CALL_RESULT_ATTRIBUTE = "ai.toolCall.result"; /** * Attributes collected for `ai.generateText` span * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generatetext-function */ export declare const AI_GENERATE_TEXT_SPAN_ATTRIBUTES: { readonly OPERATION_NAME: "operation.name"; readonly AI_OPERATION_ID: "ai.operationId"; readonly AI_PROMPT: "ai.prompt"; readonly AI_RESPONSE_TEXT: "ai.response.text"; readonly AI_RESPONSE_TOOL_CALLS: "ai.response.toolCalls"; readonly AI_RESPONSE_FINISH_REASON: "ai.response.finishReason"; readonly AI_SETTINGS_MAX_STEPS: "ai.settings.maxSteps"; readonly RESOURCE_NAME: "resource.name"; readonly AI_MODEL_ID: "ai.model.id"; readonly AI_MODEL_PROVIDER: "ai.model.provider"; readonly AI_REQUEST_HEADERS: "ai.request.headers"; readonly AI_SETTINGS_MAX_RETRIES: "ai.settings.maxRetries"; readonly AI_TELEMETRY_FUNCTION_ID: "ai.telemetry.functionId"; readonly AI_TELEMETRY_METADATA: "ai.telemetry.metadata"; readonly AI_USAGE_COMPLETION_TOKENS: "ai.usage.completionTokens"; readonly AI_USAGE_PROMPT_TOKENS: "ai.usage.promptTokens"; }; /** * Attributes collected for `ai.generateText.doGenerate` span * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generatetext-function */ export declare const AI_GENERATE_TEXT_DO_GENERATE_SPAN_ATTRIBUTES: { readonly OPERATION_NAME: "operation.name"; readonly AI_OPERATION_ID: "ai.operationId"; readonly AI_PROMPT_FORMAT: "ai.prompt.format"; readonly AI_PROMPT_MESSAGES: "ai.prompt.messages"; readonly AI_PROMPT_TOOLS: "ai.prompt.tools"; readonly AI_PROMPT_TOOL_CHOICE: "ai.prompt.toolChoice"; readonly RESOURCE_NAME: "resource.name"; readonly AI_MODEL_ID: "ai.model.id"; readonly AI_MODEL_PROVIDER: "ai.model.provider"; readonly AI_REQUEST_HEADERS: "ai.request.headers"; readonly AI_SETTINGS_MAX_RETRIES: "ai.settings.maxRetries"; readonly AI_TELEMETRY_FUNCTION_ID: "ai.telemetry.functionId"; readonly AI_TELEMETRY_METADATA: "ai.telemetry.metadata"; readonly AI_USAGE_COMPLETION_TOKENS: "ai.usage.completionTokens"; readonly AI_USAGE_PROMPT_TOKENS: "ai.usage.promptTokens"; readonly AI_RESPONSE_MODEL: "ai.response.model"; readonly AI_RESPONSE_ID: "ai.response.id"; readonly AI_RESPONSE_TIMESTAMP: "ai.response.timestamp"; readonly GEN_AI_SYSTEM: "gen_ai.system"; readonly GEN_AI_REQUEST_MODEL: "gen_ai.request.model"; readonly GEN_AI_REQUEST_TEMPERATURE: "gen_ai.request.temperature"; readonly GEN_AI_REQUEST_MAX_TOKENS: "gen_ai.request.max_tokens"; readonly GEN_AI_REQUEST_FREQUENCY_PENALTY: "gen_ai.request.frequency_penalty"; readonly GEN_AI_REQUEST_PRESENCE_PENALTY: "gen_ai.request.presence_penalty"; readonly GEN_AI_REQUEST_TOP_K: "gen_ai.request.top_k"; readonly GEN_AI_REQUEST_TOP_P: "gen_ai.request.top_p"; readonly GEN_AI_REQUEST_STOP_SEQUENCES: "gen_ai.request.stop_sequences"; readonly GEN_AI_RESPONSE_FINISH_REASONS: "gen_ai.response.finish_reasons"; readonly GEN_AI_RESPONSE_MODEL: "gen_ai.response.model"; readonly GEN_AI_RESPONSE_ID: "gen_ai.response.id"; readonly GEN_AI_USAGE_INPUT_TOKENS: "gen_ai.usage.input_tokens"; readonly GEN_AI_USAGE_OUTPUT_TOKENS: "gen_ai.usage.output_tokens"; }; /** * Attributes collected for `ai.streamText` span * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamtext-function */ export declare const AI_STREAM_TEXT_SPAN_ATTRIBUTES: { readonly OPERATION_NAME: "operation.name"; readonly AI_OPERATION_ID: "ai.operationId"; readonly AI_PROMPT: "ai.prompt"; readonly RESOURCE_NAME: "resource.name"; readonly AI_MODEL_ID: "ai.model.id"; readonly AI_MODEL_PROVIDER: "ai.model.provider"; readonly AI_REQUEST_HEADERS: "ai.request.headers"; readonly AI_SETTINGS_MAX_RETRIES: "ai.settings.maxRetries"; readonly AI_TELEMETRY_FUNCTION_ID: "ai.telemetry.functionId"; readonly AI_TELEMETRY_METADATA: "ai.telemetry.metadata"; readonly AI_USAGE_COMPLETION_TOKENS: "ai.usage.completionTokens"; readonly AI_USAGE_PROMPT_TOKENS: "ai.usage.promptTokens"; }; /** * Attributes collected for `ai.streamText.doStream` span * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamtext-function */ export declare const AI_STREAM_TEXT_DO_STREAM_SPAN_ATTRIBUTES: { readonly OPERATION_NAME: "operation.name"; readonly AI_OPERATION_ID: "ai.operationId"; readonly AI_RESPONSE_MS_TO_FIRST_CHUNK: "ai.response.msToFirstChunk"; readonly AI_RESPONSE_MS_TO_FINISH: "ai.response.msToFinish"; readonly AI_RESPONSE_AVG_COMPLETION_TOKENS_PER_SECOND: "ai.response.avgCompletionTokensPerSecond"; readonly RESOURCE_NAME: "resource.name"; readonly AI_MODEL_ID: "ai.model.id"; readonly AI_MODEL_PROVIDER: "ai.model.provider"; readonly AI_REQUEST_HEADERS: "ai.request.headers"; readonly AI_SETTINGS_MAX_RETRIES: "ai.settings.maxRetries"; readonly AI_TELEMETRY_FUNCTION_ID: "ai.telemetry.functionId"; readonly AI_TELEMETRY_METADATA: "ai.telemetry.metadata"; readonly AI_USAGE_COMPLETION_TOKENS: "ai.usage.completionTokens"; readonly AI_USAGE_PROMPT_TOKENS: "ai.usage.promptTokens"; readonly AI_RESPONSE_MODEL: "ai.response.model"; readonly AI_RESPONSE_ID: "ai.response.id"; readonly AI_RESPONSE_TIMESTAMP: "ai.response.timestamp"; readonly GEN_AI_SYSTEM: "gen_ai.system"; readonly GEN_AI_REQUEST_MODEL: "gen_ai.request.model"; readonly GEN_AI_REQUEST_TEMPERATURE: "gen_ai.request.temperature"; readonly GEN_AI_REQUEST_MAX_TOKENS: "gen_ai.request.max_tokens"; readonly GEN_AI_REQUEST_FREQUENCY_PENALTY: "gen_ai.request.frequency_penalty"; readonly GEN_AI_REQUEST_PRESENCE_PENALTY: "gen_ai.request.presence_penalty"; readonly GEN_AI_REQUEST_TOP_K: "gen_ai.request.top_k"; readonly GEN_AI_REQUEST_TOP_P: "gen_ai.request.top_p"; readonly GEN_AI_REQUEST_STOP_SEQUENCES: "gen_ai.request.stop_sequences"; readonly GEN_AI_RESPONSE_FINISH_REASONS: "gen_ai.response.finish_reasons"; readonly GEN_AI_RESPONSE_MODEL: "gen_ai.response.model"; readonly GEN_AI_RESPONSE_ID: "gen_ai.response.id"; readonly GEN_AI_USAGE_INPUT_TOKENS: "gen_ai.usage.input_tokens"; readonly GEN_AI_USAGE_OUTPUT_TOKENS: "gen_ai.usage.output_tokens"; }; /** * Attributes collected for `ai.generateObject` span * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#generateobject-function */ export declare const AI_GENERATE_OBJECT_SPAN_ATTRIBUTES: { readonly OPERATION_NAME: "operation.name"; readonly AI_OPERATION_ID: "ai.operationId"; readonly AI_SCHEMA: "ai.schema"; readonly AI_SCHEMA_NAME: "ai.schema.name"; readonly AI_SCHEMA_DESCRIPTION: "ai.schema.description"; readonly AI_RESPONSE_OBJECT: "ai.response.object"; readonly AI_SETTINGS_MODE: "ai.settings.mode"; readonly AI_SETTINGS_OUTPUT: "ai.settings.output"; readonly RESOURCE_NAME: "resource.name"; readonly AI_MODEL_ID: "ai.model.id"; readonly AI_MODEL_PROVIDER: "ai.model.provider"; readonly AI_REQUEST_HEADERS: "ai.request.headers"; readonly AI_SETTINGS_MAX_RETRIES: "ai.settings.maxRetries"; readonly AI_TELEMETRY_FUNCTION_ID: "ai.telemetry.functionId"; readonly AI_TELEMETRY_METADATA: "ai.telemetry.metadata"; readonly AI_USAGE_COMPLETION_TOKENS: "ai.usage.completionTokens"; readonly AI_USAGE_PROMPT_TOKENS: "ai.usage.promptTokens"; }; /** * Attributes collected for `ai.streamObject` span * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#streamobject-function */ export declare const AI_STREAM_OBJECT_SPAN_ATTRIBUTES: { readonly OPERATION_NAME: "operation.name"; readonly AI_OPERATION_ID: "ai.operationId"; readonly AI_SCHEMA: "ai.schema"; readonly AI_SCHEMA_NAME: "ai.schema.name"; readonly AI_SCHEMA_DESCRIPTION: "ai.schema.description"; readonly AI_RESPONSE_OBJECT: "ai.response.object"; readonly AI_SETTINGS_MODE: "ai.settings.mode"; readonly AI_SETTINGS_OUTPUT: "ai.settings.output"; readonly RESOURCE_NAME: "resource.name"; readonly AI_MODEL_ID: "ai.model.id"; readonly AI_MODEL_PROVIDER: "ai.model.provider"; readonly AI_REQUEST_HEADERS: "ai.request.headers"; readonly AI_SETTINGS_MAX_RETRIES: "ai.settings.maxRetries"; readonly AI_TELEMETRY_FUNCTION_ID: "ai.telemetry.functionId"; readonly AI_TELEMETRY_METADATA: "ai.telemetry.metadata"; readonly AI_USAGE_COMPLETION_TOKENS: "ai.usage.completionTokens"; readonly AI_USAGE_PROMPT_TOKENS: "ai.usage.promptTokens"; }; /** * Attributes collected for `ai.embed` span * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#embed-function */ export declare const AI_EMBED_SPAN_ATTRIBUTES: { readonly OPERATION_NAME: "operation.name"; readonly AI_OPERATION_ID: "ai.operationId"; readonly AI_VALUE: "ai.value"; readonly AI_EMBEDDING: "ai.embedding"; readonly RESOURCE_NAME: "resource.name"; readonly AI_MODEL_ID: "ai.model.id"; readonly AI_MODEL_PROVIDER: "ai.model.provider"; readonly AI_REQUEST_HEADERS: "ai.request.headers"; readonly AI_SETTINGS_MAX_RETRIES: "ai.settings.maxRetries"; readonly AI_TELEMETRY_FUNCTION_ID: "ai.telemetry.functionId"; readonly AI_TELEMETRY_METADATA: "ai.telemetry.metadata"; readonly AI_USAGE_TOKENS: "ai.usage.tokens"; }; /** * Attributes collected for `ai.embed.doEmbed` span * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#embed-function */ export declare const AI_EMBED_DO_EMBED_SPAN_ATTRIBUTES: { readonly OPERATION_NAME: "operation.name"; readonly AI_OPERATION_ID: "ai.operationId"; readonly AI_VALUES: "ai.values"; readonly AI_EMBEDDINGS: "ai.embeddings"; readonly RESOURCE_NAME: "resource.name"; readonly AI_MODEL_ID: "ai.model.id"; readonly AI_MODEL_PROVIDER: "ai.model.provider"; readonly AI_REQUEST_HEADERS: "ai.request.headers"; readonly AI_SETTINGS_MAX_RETRIES: "ai.settings.maxRetries"; readonly AI_TELEMETRY_FUNCTION_ID: "ai.telemetry.functionId"; readonly AI_TELEMETRY_METADATA: "ai.telemetry.metadata"; readonly AI_USAGE_TOKENS: "ai.usage.tokens"; }; /** * Attributes collected for `ai.embedMany` span * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#embedmany-function */ export declare const AI_EMBED_MANY_SPAN_ATTRIBUTES: { readonly OPERATION_NAME: "operation.name"; readonly AI_OPERATION_ID: "ai.operationId"; readonly AI_VALUES: "ai.values"; readonly AI_EMBEDDINGS: "ai.embeddings"; readonly RESOURCE_NAME: "resource.name"; readonly AI_MODEL_ID: "ai.model.id"; readonly AI_MODEL_PROVIDER: "ai.model.provider"; readonly AI_REQUEST_HEADERS: "ai.request.headers"; readonly AI_SETTINGS_MAX_RETRIES: "ai.settings.maxRetries"; readonly AI_TELEMETRY_FUNCTION_ID: "ai.telemetry.functionId"; readonly AI_TELEMETRY_METADATA: "ai.telemetry.metadata"; readonly AI_USAGE_TOKENS: "ai.usage.tokens"; }; /** * Attributes collected for `ai.toolCall` span * @see https://ai-sdk.dev/docs/ai-sdk-core/telemetry#tool-call-spans */ export declare const AI_TOOL_CALL_SPAN_ATTRIBUTES: { readonly OPERATION_NAME: "operation.name"; readonly AI_OPERATION_ID: "ai.operationId"; readonly AI_TOOL_CALL_NAME: "ai.toolCall.name"; readonly AI_TOOL_CALL_ID: "ai.toolCall.id"; readonly AI_TOOL_CALL_ARGS: "ai.toolCall.args"; readonly AI_TOOL_CALL_RESULT: "ai.toolCall.result"; readonly RESOURCE_NAME: "resource.name"; readonly AI_MODEL_ID: "ai.model.id"; readonly AI_MODEL_PROVIDER: "ai.model.provider"; readonly AI_REQUEST_HEADERS: "ai.request.headers"; readonly AI_SETTINGS_MAX_RETRIES: "ai.settings.maxRetries"; readonly AI_TELEMETRY_FUNCTION_ID: "ai.telemetry.functionId"; readonly AI_TELEMETRY_METADATA: "ai.telemetry.metadata"; }; /** * OpenAI Provider Metadata * @see https://ai-sdk.dev/providers/ai-sdk-providers/openai * @see https://github.com/vercel/ai/blob/65e042afde6aad4da9d7a62526ece839eb34f9a5/packages/openai/src/openai-chat-language-model.ts#L397-L416 * @see https://github.com/vercel/ai/blob/65e042afde6aad4da9d7a62526ece839eb34f9a5/packages/openai/src/responses/openai-responses-language-model.ts#L377C7-L384 */ interface OpenAiProviderMetadata { /** * The number of predicted output tokens that were accepted. * @see https://ai-sdk.dev/providers/ai-sdk-providers/openai#predicted-outputs */ acceptedPredictionTokens?: number; /** * The number of predicted output tokens that were rejected. * @see https://ai-sdk.dev/providers/ai-sdk-providers/openai#predicted-outputs */ rejectedPredictionTokens?: number; /** * The number of reasoning tokens that the model generated. * @see https://ai-sdk.dev/providers/ai-sdk-providers/openai#responses-models */ reasoningTokens?: number; /** * The number of prompt tokens that were a cache hit. * @see https://ai-sdk.dev/providers/ai-sdk-providers/openai#responses-models */ cachedPromptTokens?: number; /** * @see https://ai-sdk.dev/providers/ai-sdk-providers/openai#responses-models * * The ID of the response. Can be used to continue a conversation. */ responseId?: string; } /** * Anthropic Provider Metadata * @see https://ai-sdk.dev/providers/ai-sdk-providers/anthropic * @see https://github.com/vercel/ai/blob/65e042afde6aad4da9d7a62526ece839eb34f9a5/packages/anthropic/src/anthropic-messages-language-model.ts#L346-L352 */ interface AnthropicProviderMetadata { /** * The number of tokens that were used to create the cache. * @see https://ai-sdk.dev/providers/ai-sdk-providers/anthropic#cache-control */ cacheCreationInputTokens?: number; /** * The number of tokens that were read from the cache. * @see https://ai-sdk.dev/providers/ai-sdk-providers/anthropic#cache-control */ cacheReadInputTokens?: number; } /** * Amazon Bedrock Provider Metadata * @see https://ai-sdk.dev/providers/ai-sdk-providers/amazon-bedrock * @see https://github.com/vercel/ai/blob/65e042afde6aad4da9d7a62526ece839eb34f9a5/packages/amazon-bedrock/src/bedrock-chat-language-model.ts#L263-L280 */ interface AmazonBedrockProviderMetadata { /** * @see https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_ConverseTrace.html */ trace?: { /** * The guardrail trace object. * @see https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_GuardrailTraceAssessment.html * * This was purposely left as unknown as it's a complex object. This can be typed in the future * if the SDK decides to support bedrock in a more advanced way. */ guardrail?: unknown; /** * The request's prompt router. * @see https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_PromptRouterTrace.html */ promptRouter?: { /** * The ID of the invoked model. */ invokedModelId?: string; }; }; usage?: { /** * The number of tokens that were read from the cache. * @see https://ai-sdk.dev/providers/ai-sdk-providers/amazon-bedrock#cache-points */ cacheReadInputTokens?: number; /** * The number of tokens that were written to the cache. * @see https://ai-sdk.dev/providers/ai-sdk-providers/amazon-bedrock#cache-points */ cacheWriteInputTokens?: number; }; } /** * Google Generative AI Provider Metadata * @see https://ai-sdk.dev/providers/ai-sdk-providers/google-generative-ai */ export interface GoogleGenerativeAIProviderMetadata { /** * @see https://github.com/vercel/ai/blob/65e042afde6aad4da9d7a62526ece839eb34f9a5/packages/google/src/google-generative-ai-prompt.ts#L28-L30 */ groundingMetadata: null | { /** * Array of search queries used to retrieve information * @example ["What's the weather in Chicago this weekend?"] * * @see https://ai-sdk.dev/providers/ai-sdk-providers/google-generative-ai#search-grounding */ webSearchQueries: string[] | null; /** * Contains the main search result content used as an entry point * The `renderedContent` field contains the formatted content * @see https://ai-sdk.dev/providers/ai-sdk-providers/google-generative-ai#search-grounding */ searchEntryPoint?: { renderedContent: string; } | null; /** * Contains details about how specific response parts are supported by search results * @see https://ai-sdk.dev/providers/ai-sdk-providers/google-generative-ai#search-grounding */ groundingSupports: Array<{ /** * Information about the grounded text segment. */ segment: { /** * The start index of the text segment. */ startIndex?: number | null; /** * The end index of the text segment. */ endIndex?: number | null; /** * The actual text segment. */ text?: string | null; }; /** * References to supporting search result chunks. */ groundingChunkIndices?: number[] | null; /** * Confidence scores (0-1) for each supporting chunk. */ confidenceScores?: number[] | null; }> | null; }; /** * @see https://github.com/vercel/ai/blob/65e042afde6aad4da9d7a62526ece839eb34f9a5/packages/google/src/google-generative-ai-language-model.ts#L620-L627 * @see https://cloud.google.com/vertex-ai/generative-ai/docs/multimodal/configure-safety-filters */ safetyRatings?: null | unknown; } /** * DeepSeek Provider Metadata * @see https://ai-sdk.dev/providers/ai-sdk-providers/deepseek */ interface DeepSeekProviderMetadata { /** * The number of tokens that were cache hits. * @see https://ai-sdk.dev/providers/ai-sdk-providers/deepseek#cache-token-usage */ promptCacheHitTokens?: number; /** * The number of tokens that were cache misses. * @see https://ai-sdk.dev/providers/ai-sdk-providers/deepseek#cache-token-usage */ promptCacheMissTokens?: number; } /** * Perplexity Provider Metadata * @see https://ai-sdk.dev/providers/ai-sdk-providers/perplexity */ interface PerplexityProviderMetadata { /** * Object containing citationTokens and numSearchQueries metrics */ usage?: { citationTokens?: number; numSearchQueries?: number; }; /** * Array of image URLs when return_images is enabled. * * You can enable image responses by setting return_images: true in the provider options. * This feature is only available to Perplexity Tier-2 users and above. */ images?: Array<{ imageUrl?: string; originUrl?: string; height?: number; width?: number; }>; } export interface ProviderMetadata { openai?: OpenAiProviderMetadata; anthropic?: AnthropicProviderMetadata; bedrock?: AmazonBedrockProviderMetadata; google?: GoogleGenerativeAIProviderMetadata; deepseek?: DeepSeekProviderMetadata; perplexity?: PerplexityProviderMetadata; } export {}; //# sourceMappingURL=vercel-ai-attributes.d.ts.map