Gen AI
Gen AI Attributes
This document defines the attributes used to describe telemetry in the context of LLM (Large Language Models) requests and responses.
| Attribute |
Type |
Description |
Examples |
Stability |
gen_ai.completion |
string |
The full response received from the LLM. [1] |
[{'role': 'assistant', 'content': 'The capital of France is Paris.'}] |
 |
gen_ai.operation.name |
string |
The name of the operation being performed. |
chat; completion |
 |
gen_ai.prompt |
string |
The full prompt sent to an LLM. [2] |
[{'role': 'user', 'content': 'What is the capital of France?'}] |
 |
gen_ai.request.max_tokens |
int |
The maximum number of tokens the LLM generates for a request. |
100 |
 |
gen_ai.request.model |
string |
The name of the LLM a request is being made to. |
gpt-4 |
 |
gen_ai.request.temperature |
double |
The temperature setting for the LLM request. |
0.0 |
 |
gen_ai.request.top_p |
double |
The top_p sampling setting for the LLM request. |
1.0 |
 |
gen_ai.response.finish_reasons |
string[] |
Array of reasons the model stopped generating tokens, corresponding to each generation received. |
["stop"] |
 |
gen_ai.response.id |
string |
The unique identifier for the completion. |
chatcmpl-123 |
 |
gen_ai.response.model |
string |
The name of the LLM a response was generated from. |
gpt-4-0613 |
 |
gen_ai.system |
string |
The Generative AI product as identified by the client instrumentation. [3] |
openai |
 |
gen_ai.token.type |
string |
The type of token being counted. |
input; output |
 |
gen_ai.usage.completion_tokens |
int |
The number of tokens used in the LLM response (completion). |
180 |
 |
gen_ai.usage.prompt_tokens |
int |
The number of tokens used in the LLM prompt. |
100 |
 |
[1]: It's RECOMMENDED to format completions as JSON string matching OpenAI messages format
[2]: It's RECOMMENDED to format prompts as JSON string matching OpenAI messages format
[3]: The actual GenAI product may differ from the one identified by the client. For example, when using OpenAI client libraries to communicate with Mistral, the gen_ai.system is set to openai based on the instrumentation's best knowledge.
gen_ai.system has the following list of well-known values. If one of them applies, then the respective value MUST be used; otherwise, a custom value MAY be used.
| Value |
Description |
Stability |
anthropic |
Anthropic |
 |
cohere |
Cohere |
 |
openai |
OpenAI |
 |
vertex_ai |
Vertex AI |
 |
gen_ai.token.type has the following list of well-known values. If one of them applies, then the respective value MUST be used; otherwise, a custom value MAY be used.
| Value |
Description |
Stability |
input |
Input tokens (prompt, input, etc.) |
 |
output |
Output tokens (completion, response, etc.) |
 |