Skip to content

Commit 85dc005

Browse files
author
ci.datadog-api-spec
committed
Regenerate client from commit 9d88080 of spec repo
1 parent 71b9eac commit 85dc005

31 files changed

+5727
-3033
lines changed

.generator/schemas/v2/openapi.yaml

Lines changed: 613 additions & 0 deletions
Large diffs are not rendered by default.

features/v2/llm_observability.feature

Lines changed: 84 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -175,6 +175,42 @@ Feature: LLM Observability
175175
When the request is sent
176176
Then the response status is 200 OK
177177

178+
@generated @skip @team:DataDog/ml-observability
179+
Scenario: Create or update a custom evaluator configuration returns "Bad Request" response
180+
Given operation "UpdateLLMObsCustomEvalConfig" enabled
181+
And new "UpdateLLMObsCustomEvalConfig" request
182+
And request contains "eval_name" parameter from "REPLACE.ME"
183+
And body with value {"data": {"attributes": {"category": "Custom", "eval_name": "my-custom-evaluator", "llm_judge_config": {"assessment_criteria": {"max_threshold": 1.0, "min_threshold": 0.7, "pass_values": ["pass", "yes"], "pass_when": true}, "inference_params": {"frequency_penalty": 0.0, "max_tokens": 1024, "presence_penalty": 0.0, "temperature": 0.7, "top_k": 50, "top_p": 1.0}, "last_used_library_prompt_template_name": "sentiment-analysis-v1", "modified_library_prompt_template": false, "output_schema": null, "parsing_type": "structured_output", "prompt_template": [{"content": "Rate the quality of the following response:", "contents": [{"type": "text", "value": {"text": "What is the sentiment of this review?", "tool_call": {"arguments": "{\"location\": \"San Francisco\"}", "id": "call_abc123", "name": "get_weather", "type": "function"}, "tool_call_result": {"name": "get_weather", "result": "sunny, 72F", "tool_id": "call_abc123", "type": "function"}}}], "role": "user"}]}, "llm_provider": {"bedrock": {"region": "us-east-1"}, "integration_account_id": "my-account-id", "integration_provider": "openai", "model_name": "gpt-4o", "vertex_ai": {"location": "us-central1", "project": "my-gcp-project"}}, "target": {"application_name": "my-llm-app", "enabled": true, "eval_scope": "span", "filter": "@service:my-service", "root_spans_only": true, "sampling_percentage": 50.0}}, "id": "my-custom-evaluator", "type": "evaluator_config"}}
184+
When the request is sent
185+
Then the response status is 400 Bad Request
186+
187+
@generated @skip @team:DataDog/ml-observability
188+
Scenario: Create or update a custom evaluator configuration returns "Not Found" response
189+
Given operation "UpdateLLMObsCustomEvalConfig" enabled
190+
And new "UpdateLLMObsCustomEvalConfig" request
191+
And request contains "eval_name" parameter from "REPLACE.ME"
192+
And body with value {"data": {"attributes": {"category": "Custom", "eval_name": "my-custom-evaluator", "llm_judge_config": {"assessment_criteria": {"max_threshold": 1.0, "min_threshold": 0.7, "pass_values": ["pass", "yes"], "pass_when": true}, "inference_params": {"frequency_penalty": 0.0, "max_tokens": 1024, "presence_penalty": 0.0, "temperature": 0.7, "top_k": 50, "top_p": 1.0}, "last_used_library_prompt_template_name": "sentiment-analysis-v1", "modified_library_prompt_template": false, "output_schema": null, "parsing_type": "structured_output", "prompt_template": [{"content": "Rate the quality of the following response:", "contents": [{"type": "text", "value": {"text": "What is the sentiment of this review?", "tool_call": {"arguments": "{\"location\": \"San Francisco\"}", "id": "call_abc123", "name": "get_weather", "type": "function"}, "tool_call_result": {"name": "get_weather", "result": "sunny, 72F", "tool_id": "call_abc123", "type": "function"}}}], "role": "user"}]}, "llm_provider": {"bedrock": {"region": "us-east-1"}, "integration_account_id": "my-account-id", "integration_provider": "openai", "model_name": "gpt-4o", "vertex_ai": {"location": "us-central1", "project": "my-gcp-project"}}, "target": {"application_name": "my-llm-app", "enabled": true, "eval_scope": "span", "filter": "@service:my-service", "root_spans_only": true, "sampling_percentage": 50.0}}, "id": "my-custom-evaluator", "type": "evaluator_config"}}
193+
When the request is sent
194+
Then the response status is 404 Not Found
195+
196+
@generated @skip @team:DataDog/ml-observability
197+
Scenario: Create or update a custom evaluator configuration returns "OK" response
198+
Given operation "UpdateLLMObsCustomEvalConfig" enabled
199+
And new "UpdateLLMObsCustomEvalConfig" request
200+
And request contains "eval_name" parameter from "REPLACE.ME"
201+
And body with value {"data": {"attributes": {"category": "Custom", "eval_name": "my-custom-evaluator", "llm_judge_config": {"assessment_criteria": {"max_threshold": 1.0, "min_threshold": 0.7, "pass_values": ["pass", "yes"], "pass_when": true}, "inference_params": {"frequency_penalty": 0.0, "max_tokens": 1024, "presence_penalty": 0.0, "temperature": 0.7, "top_k": 50, "top_p": 1.0}, "last_used_library_prompt_template_name": "sentiment-analysis-v1", "modified_library_prompt_template": false, "output_schema": null, "parsing_type": "structured_output", "prompt_template": [{"content": "Rate the quality of the following response:", "contents": [{"type": "text", "value": {"text": "What is the sentiment of this review?", "tool_call": {"arguments": "{\"location\": \"San Francisco\"}", "id": "call_abc123", "name": "get_weather", "type": "function"}, "tool_call_result": {"name": "get_weather", "result": "sunny, 72F", "tool_id": "call_abc123", "type": "function"}}}], "role": "user"}]}, "llm_provider": {"bedrock": {"region": "us-east-1"}, "integration_account_id": "my-account-id", "integration_provider": "openai", "model_name": "gpt-4o", "vertex_ai": {"location": "us-central1", "project": "my-gcp-project"}}, "target": {"application_name": "my-llm-app", "enabled": true, "eval_scope": "span", "filter": "@service:my-service", "root_spans_only": true, "sampling_percentage": 50.0}}, "id": "my-custom-evaluator", "type": "evaluator_config"}}
202+
When the request is sent
203+
Then the response status is 200 OK
204+
205+
@generated @skip @team:DataDog/ml-observability
206+
Scenario: Create or update a custom evaluator configuration returns "Unprocessable Entity" response
207+
Given operation "UpdateLLMObsCustomEvalConfig" enabled
208+
And new "UpdateLLMObsCustomEvalConfig" request
209+
And request contains "eval_name" parameter from "REPLACE.ME"
210+
And body with value {"data": {"attributes": {"category": "Custom", "eval_name": "my-custom-evaluator", "llm_judge_config": {"assessment_criteria": {"max_threshold": 1.0, "min_threshold": 0.7, "pass_values": ["pass", "yes"], "pass_when": true}, "inference_params": {"frequency_penalty": 0.0, "max_tokens": 1024, "presence_penalty": 0.0, "temperature": 0.7, "top_k": 50, "top_p": 1.0}, "last_used_library_prompt_template_name": "sentiment-analysis-v1", "modified_library_prompt_template": false, "output_schema": null, "parsing_type": "structured_output", "prompt_template": [{"content": "Rate the quality of the following response:", "contents": [{"type": "text", "value": {"text": "What is the sentiment of this review?", "tool_call": {"arguments": "{\"location\": \"San Francisco\"}", "id": "call_abc123", "name": "get_weather", "type": "function"}, "tool_call_result": {"name": "get_weather", "result": "sunny, 72F", "tool_id": "call_abc123", "type": "function"}}}], "role": "user"}]}, "llm_provider": {"bedrock": {"region": "us-east-1"}, "integration_account_id": "my-account-id", "integration_provider": "openai", "model_name": "gpt-4o", "vertex_ai": {"location": "us-central1", "project": "my-gcp-project"}}, "target": {"application_name": "my-llm-app", "enabled": true, "eval_scope": "span", "filter": "@service:my-service", "root_spans_only": true, "sampling_percentage": 50.0}}, "id": "my-custom-evaluator", "type": "evaluator_config"}}
211+
When the request is sent
212+
Then the response status is 422 Unprocessable Entity
213+
178214
@generated @skip @team:DataDog/ml-observability
179215
Scenario: Delete LLM Observability dataset records returns "Bad Request" response
180216
Given operation "DeleteLLMObsDatasetRecords" enabled
@@ -264,6 +300,30 @@ Feature: LLM Observability
264300
When the request is sent
265301
Then the response status is 204 No Content
266302

303+
@generated @skip @team:DataDog/ml-observability
304+
Scenario: Delete a custom evaluator configuration returns "Bad Request" response
305+
Given operation "DeleteLLMObsCustomEvalConfig" enabled
306+
And new "DeleteLLMObsCustomEvalConfig" request
307+
And request contains "eval_name" parameter from "REPLACE.ME"
308+
When the request is sent
309+
Then the response status is 400 Bad Request
310+
311+
@generated @skip @team:DataDog/ml-observability
312+
Scenario: Delete a custom evaluator configuration returns "No Content" response
313+
Given operation "DeleteLLMObsCustomEvalConfig" enabled
314+
And new "DeleteLLMObsCustomEvalConfig" request
315+
And request contains "eval_name" parameter from "REPLACE.ME"
316+
When the request is sent
317+
Then the response status is 204 No Content
318+
319+
@generated @skip @team:DataDog/ml-observability
320+
Scenario: Delete a custom evaluator configuration returns "Not Found" response
321+
Given operation "DeleteLLMObsCustomEvalConfig" enabled
322+
And new "DeleteLLMObsCustomEvalConfig" request
323+
And request contains "eval_name" parameter from "REPLACE.ME"
324+
When the request is sent
325+
Then the response status is 404 Not Found
326+
267327
@generated @skip @team:DataDog/ml-observability
268328
Scenario: Delete an LLM Observability annotation queue returns "No Content" response
269329
Given operation "DeleteLLMObsAnnotationQueue" enabled
@@ -307,6 +367,30 @@ Feature: LLM Observability
307367
When the request is sent
308368
Then the response status is 404 Not Found
309369

370+
@generated @skip @team:DataDog/ml-observability
371+
Scenario: Get a custom evaluator configuration returns "Bad Request" response
372+
Given operation "GetLLMObsCustomEvalConfig" enabled
373+
And new "GetLLMObsCustomEvalConfig" request
374+
And request contains "eval_name" parameter from "REPLACE.ME"
375+
When the request is sent
376+
Then the response status is 400 Bad Request
377+
378+
@generated @skip @team:DataDog/ml-observability
379+
Scenario: Get a custom evaluator configuration returns "Not Found" response
380+
Given operation "GetLLMObsCustomEvalConfig" enabled
381+
And new "GetLLMObsCustomEvalConfig" request
382+
And request contains "eval_name" parameter from "REPLACE.ME"
383+
When the request is sent
384+
Then the response status is 404 Not Found
385+
386+
@generated @skip @team:DataDog/ml-observability
387+
Scenario: Get a custom evaluator configuration returns "OK" response
388+
Given operation "GetLLMObsCustomEvalConfig" enabled
389+
And new "GetLLMObsCustomEvalConfig" request
390+
And request contains "eval_name" parameter from "REPLACE.ME"
391+
When the request is sent
392+
Then the response status is 200 OK
393+
310394
@generated @skip @team:DataDog/ml-observability
311395
Scenario: Get annotated queue interactions returns "Bad Request" response
312396
Given operation "GetLLMObsAnnotatedInteractions" enabled

features/v2/undo.json

Lines changed: 25 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -111,6 +111,31 @@
111111
"type": "unsafe"
112112
}
113113
},
114+
"DeleteLLMObsCustomEvalConfig": {
115+
"tag": "LLM Observability",
116+
"undo": {
117+
"operationId": "UpdateLLMObsCustomEvalConfig",
118+
"parameters": [
119+
{
120+
"name": "eval_name",
121+
"source": "path"
122+
}
123+
],
124+
"type": "unsafe"
125+
}
126+
},
127+
"GetLLMObsCustomEvalConfig": {
128+
"tag": "LLM Observability",
129+
"undo": {
130+
"type": "safe"
131+
}
132+
},
133+
"UpdateLLMObsCustomEvalConfig": {
134+
"tag": "LLM Observability",
135+
"undo": {
136+
"type": "idempotent"
137+
}
138+
},
114139
"ListDatastores": {
115140
"tag": "Actions Datastores",
116141
"undo": {

0 commit comments

Comments
 (0)