@@ -175,6 +175,42 @@ Feature: LLM Observability
175175 When the request is sent
176176 Then the response status is 200 OK
177177
178+ @generated @skip @team:DataDog/ml-observability
179+ Scenario : Create or update a custom evaluator configuration returns "Bad Request" response
180+ Given operation "UpdateLLMObsCustomEvalConfig" enabled
181+ And new "UpdateLLMObsCustomEvalConfig" request
182+ And request contains "eval_name" parameter from "REPLACE.ME"
183+ And body with value {"data": {"attributes": {"category": "Custom", "eval_name": "my-custom-evaluator", "llm_judge_config": {"assessment_criteria": {"max_threshold": 1.0, "min_threshold": 0.7, "pass_values": ["pass", "yes"], "pass_when": true}, "inference_params": {"frequency_penalty": 0.0, "max_tokens": 1024, "presence_penalty": 0.0, "temperature": 0.7, "top_k": 50, "top_p": 1.0}, "last_used_library_prompt_template_name": "sentiment-analysis-v1", "modified_library_prompt_template": false, "output_schema": null, "parsing_type": "structured_output", "prompt_template": [{"content": "Rate the quality of the following response:", "contents": [{"type": "text", "value": {"text": "What is the sentiment of this review?", "tool_call": {"arguments": "{\"location\": \"San Francisco\"}", "id": "call_abc123", "name": "get_weather", "type": "function"}, "tool_call_result": {"name": "get_weather", "result": "sunny, 72F", "tool_id": "call_abc123", "type": "function"}}}], "role": "user"}]}, "llm_provider": {"bedrock": {"region": "us-east-1"}, "integration_account_id": "my-account-id", "integration_provider": "openai", "model_name": "gpt-4o", "vertex_ai": {"location": "us-central1", "project": "my-gcp-project"}}, "target": {"application_name": "my-llm-app", "enabled": true, "eval_scope": "span", "filter": "@service:my-service", "root_spans_only": true, "sampling_percentage": 50.0}}, "id": "my-custom-evaluator", "type": "evaluator_config"}}
184+ When the request is sent
185+ Then the response status is 400 Bad Request
186+
187+ @generated @skip @team:DataDog/ml-observability
188+ Scenario : Create or update a custom evaluator configuration returns "Not Found" response
189+ Given operation "UpdateLLMObsCustomEvalConfig" enabled
190+ And new "UpdateLLMObsCustomEvalConfig" request
191+ And request contains "eval_name" parameter from "REPLACE.ME"
192+ And body with value {"data": {"attributes": {"category": "Custom", "eval_name": "my-custom-evaluator", "llm_judge_config": {"assessment_criteria": {"max_threshold": 1.0, "min_threshold": 0.7, "pass_values": ["pass", "yes"], "pass_when": true}, "inference_params": {"frequency_penalty": 0.0, "max_tokens": 1024, "presence_penalty": 0.0, "temperature": 0.7, "top_k": 50, "top_p": 1.0}, "last_used_library_prompt_template_name": "sentiment-analysis-v1", "modified_library_prompt_template": false, "output_schema": null, "parsing_type": "structured_output", "prompt_template": [{"content": "Rate the quality of the following response:", "contents": [{"type": "text", "value": {"text": "What is the sentiment of this review?", "tool_call": {"arguments": "{\"location\": \"San Francisco\"}", "id": "call_abc123", "name": "get_weather", "type": "function"}, "tool_call_result": {"name": "get_weather", "result": "sunny, 72F", "tool_id": "call_abc123", "type": "function"}}}], "role": "user"}]}, "llm_provider": {"bedrock": {"region": "us-east-1"}, "integration_account_id": "my-account-id", "integration_provider": "openai", "model_name": "gpt-4o", "vertex_ai": {"location": "us-central1", "project": "my-gcp-project"}}, "target": {"application_name": "my-llm-app", "enabled": true, "eval_scope": "span", "filter": "@service:my-service", "root_spans_only": true, "sampling_percentage": 50.0}}, "id": "my-custom-evaluator", "type": "evaluator_config"}}
193+ When the request is sent
194+ Then the response status is 404 Not Found
195+
196+ @generated @skip @team:DataDog/ml-observability
197+ Scenario : Create or update a custom evaluator configuration returns "OK" response
198+ Given operation "UpdateLLMObsCustomEvalConfig" enabled
199+ And new "UpdateLLMObsCustomEvalConfig" request
200+ And request contains "eval_name" parameter from "REPLACE.ME"
201+ And body with value {"data": {"attributes": {"category": "Custom", "eval_name": "my-custom-evaluator", "llm_judge_config": {"assessment_criteria": {"max_threshold": 1.0, "min_threshold": 0.7, "pass_values": ["pass", "yes"], "pass_when": true}, "inference_params": {"frequency_penalty": 0.0, "max_tokens": 1024, "presence_penalty": 0.0, "temperature": 0.7, "top_k": 50, "top_p": 1.0}, "last_used_library_prompt_template_name": "sentiment-analysis-v1", "modified_library_prompt_template": false, "output_schema": null, "parsing_type": "structured_output", "prompt_template": [{"content": "Rate the quality of the following response:", "contents": [{"type": "text", "value": {"text": "What is the sentiment of this review?", "tool_call": {"arguments": "{\"location\": \"San Francisco\"}", "id": "call_abc123", "name": "get_weather", "type": "function"}, "tool_call_result": {"name": "get_weather", "result": "sunny, 72F", "tool_id": "call_abc123", "type": "function"}}}], "role": "user"}]}, "llm_provider": {"bedrock": {"region": "us-east-1"}, "integration_account_id": "my-account-id", "integration_provider": "openai", "model_name": "gpt-4o", "vertex_ai": {"location": "us-central1", "project": "my-gcp-project"}}, "target": {"application_name": "my-llm-app", "enabled": true, "eval_scope": "span", "filter": "@service:my-service", "root_spans_only": true, "sampling_percentage": 50.0}}, "id": "my-custom-evaluator", "type": "evaluator_config"}}
202+ When the request is sent
203+ Then the response status is 200 OK
204+
205+ @generated @skip @team:DataDog/ml-observability
206+ Scenario : Create or update a custom evaluator configuration returns "Unprocessable Entity" response
207+ Given operation "UpdateLLMObsCustomEvalConfig" enabled
208+ And new "UpdateLLMObsCustomEvalConfig" request
209+ And request contains "eval_name" parameter from "REPLACE.ME"
210+ And body with value {"data": {"attributes": {"category": "Custom", "eval_name": "my-custom-evaluator", "llm_judge_config": {"assessment_criteria": {"max_threshold": 1.0, "min_threshold": 0.7, "pass_values": ["pass", "yes"], "pass_when": true}, "inference_params": {"frequency_penalty": 0.0, "max_tokens": 1024, "presence_penalty": 0.0, "temperature": 0.7, "top_k": 50, "top_p": 1.0}, "last_used_library_prompt_template_name": "sentiment-analysis-v1", "modified_library_prompt_template": false, "output_schema": null, "parsing_type": "structured_output", "prompt_template": [{"content": "Rate the quality of the following response:", "contents": [{"type": "text", "value": {"text": "What is the sentiment of this review?", "tool_call": {"arguments": "{\"location\": \"San Francisco\"}", "id": "call_abc123", "name": "get_weather", "type": "function"}, "tool_call_result": {"name": "get_weather", "result": "sunny, 72F", "tool_id": "call_abc123", "type": "function"}}}], "role": "user"}]}, "llm_provider": {"bedrock": {"region": "us-east-1"}, "integration_account_id": "my-account-id", "integration_provider": "openai", "model_name": "gpt-4o", "vertex_ai": {"location": "us-central1", "project": "my-gcp-project"}}, "target": {"application_name": "my-llm-app", "enabled": true, "eval_scope": "span", "filter": "@service:my-service", "root_spans_only": true, "sampling_percentage": 50.0}}, "id": "my-custom-evaluator", "type": "evaluator_config"}}
211+ When the request is sent
212+ Then the response status is 422 Unprocessable Entity
213+
178214 @generated @skip @team:DataDog/ml-observability
179215 Scenario : Delete LLM Observability dataset records returns "Bad Request" response
180216 Given operation "DeleteLLMObsDatasetRecords" enabled
@@ -264,6 +300,30 @@ Feature: LLM Observability
264300 When the request is sent
265301 Then the response status is 204 No Content
266302
303+ @generated @skip @team:DataDog/ml-observability
304+ Scenario : Delete a custom evaluator configuration returns "Bad Request" response
305+ Given operation "DeleteLLMObsCustomEvalConfig" enabled
306+ And new "DeleteLLMObsCustomEvalConfig" request
307+ And request contains "eval_name" parameter from "REPLACE.ME"
308+ When the request is sent
309+ Then the response status is 400 Bad Request
310+
311+ @generated @skip @team:DataDog/ml-observability
312+ Scenario : Delete a custom evaluator configuration returns "No Content" response
313+ Given operation "DeleteLLMObsCustomEvalConfig" enabled
314+ And new "DeleteLLMObsCustomEvalConfig" request
315+ And request contains "eval_name" parameter from "REPLACE.ME"
316+ When the request is sent
317+ Then the response status is 204 No Content
318+
319+ @generated @skip @team:DataDog/ml-observability
320+ Scenario : Delete a custom evaluator configuration returns "Not Found" response
321+ Given operation "DeleteLLMObsCustomEvalConfig" enabled
322+ And new "DeleteLLMObsCustomEvalConfig" request
323+ And request contains "eval_name" parameter from "REPLACE.ME"
324+ When the request is sent
325+ Then the response status is 404 Not Found
326+
267327 @generated @skip @team:DataDog/ml-observability
268328 Scenario : Delete an LLM Observability annotation queue returns "No Content" response
269329 Given operation "DeleteLLMObsAnnotationQueue" enabled
@@ -307,6 +367,30 @@ Feature: LLM Observability
307367 When the request is sent
308368 Then the response status is 404 Not Found
309369
370+ @generated @skip @team:DataDog/ml-observability
371+ Scenario : Get a custom evaluator configuration returns "Bad Request" response
372+ Given operation "GetLLMObsCustomEvalConfig" enabled
373+ And new "GetLLMObsCustomEvalConfig" request
374+ And request contains "eval_name" parameter from "REPLACE.ME"
375+ When the request is sent
376+ Then the response status is 400 Bad Request
377+
378+ @generated @skip @team:DataDog/ml-observability
379+ Scenario : Get a custom evaluator configuration returns "Not Found" response
380+ Given operation "GetLLMObsCustomEvalConfig" enabled
381+ And new "GetLLMObsCustomEvalConfig" request
382+ And request contains "eval_name" parameter from "REPLACE.ME"
383+ When the request is sent
384+ Then the response status is 404 Not Found
385+
386+ @generated @skip @team:DataDog/ml-observability
387+ Scenario : Get a custom evaluator configuration returns "OK" response
388+ Given operation "GetLLMObsCustomEvalConfig" enabled
389+ And new "GetLLMObsCustomEvalConfig" request
390+ And request contains "eval_name" parameter from "REPLACE.ME"
391+ When the request is sent
392+ Then the response status is 200 OK
393+
310394 @generated @skip @team:DataDog/ml-observability
311395 Scenario : Get annotated queue interactions returns "Bad Request" response
312396 Given operation "GetLLMObsAnnotatedInteractions" enabled
0 commit comments