diff --git a/.generator/schemas/v2/openapi.yaml b/.generator/schemas/v2/openapi.yaml index 085c5aa51a..8c9919873e 100644 --- a/.generator/schemas/v2/openapi.yaml +++ b/.generator/schemas/v2/openapi.yaml @@ -28208,6 +28208,10 @@ components: description: The `add_env_vars` processor adds environment variable values to log events. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean id: description: The unique identifier for this component. Used to reference this processor in the pipeline. @@ -28267,6 +28271,10 @@ components: ObservabilityPipelineAddFieldsProcessor: description: The `add_fields` processor adds static key-value fields to logs. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean fields: description: A list of static fields (key-value pairs) that is added to each log event processed by this component. @@ -28768,6 +28776,10 @@ components: Remap Language (VRL)](https://vector.dev/docs/reference/vrl/) scripts with advanced filtering capabilities. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean id: description: The unique identifier for this processor. example: remap-vrl-processor @@ -28940,6 +28952,10 @@ components: properties: action: $ref: '#/components/schemas/ObservabilityPipelineDatadogTagsProcessorAction' + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean id: description: The unique identifier for this component. Used to reference this component in other parts of the pipeline (for example, as the `input` @@ -29025,6 +29041,10 @@ components: ObservabilityPipelineDedupeProcessor: description: The `dedupe` processor removes duplicate fields in log events. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean fields: description: A list of log field paths to check for duplicates. example: @@ -29266,6 +29286,10 @@ components: description: The `enrichment_table` processor enriches logs using a static CSV file or GeoIP database. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean file: $ref: '#/components/schemas/ObservabilityPipelineEnrichmentTableFile' geoip: @@ -29329,6 +29353,10 @@ components: on a Datadog search query. Logs that match the `include` query are passed through; others are discarded. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean id: description: The unique identifier for this component. Used to reference this component in other parts of the pipeline (for example, as the `input` @@ -29439,6 +29467,10 @@ components: ' properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean id: description: The unique identifier for this component. Used to reference this component in other parts of the pipeline. @@ -30015,6 +30047,10 @@ components: description: The `ocsf_mapper` processor transforms logs into the OCSF schema using a predefined mapping configuration. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean id: description: The unique identifier for this component. Used to reference this component in other parts of the pipeline. @@ -30149,6 +30185,10 @@ components: by Datadog. example: true type: boolean + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean id: description: A unique identifier for this processor. example: parse-grok-processor @@ -30273,6 +30313,10 @@ components: and flattens it into the event. This is useful when logs contain embedded JSON as a string. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean field: description: The name of the log field that contains a JSON string. example: message @@ -30336,6 +30380,10 @@ components: the filter query continue through the pipeline. example: false type: boolean + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean id: description: The unique identifier for this component. Used to reference this component in other parts of the pipeline (for example, as the `input` @@ -30468,6 +30516,10 @@ components: description: The `reduce` processor aggregates and merges logs based on matching keys and merge strategies. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean group_by: description: A list of fields used to group log events for merging. example: @@ -30564,6 +30616,10 @@ components: ObservabilityPipelineRemoveFieldsProcessor: description: The `remove_fields` processor deletes specified fields from logs. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean fields: description: A list of field names to be removed from each log event. example: @@ -30611,6 +30667,10 @@ components: ObservabilityPipelineRenameFieldsProcessor: description: The `rename_fields` processor changes field names. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean fields: description: A list of rename rules specifying which fields to rename in the event, what to rename them to, and whether to preserve the original @@ -30750,6 +30810,10 @@ components: description: The `sample` processor allows probabilistic sampling of logs at a fixed rate. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean id: description: The unique identifier for this component. Used to reference this component in other parts of the pipeline (for example, as the `input` @@ -30801,6 +30865,10 @@ components: description: The `sensitive_data_scanner` processor detects and optionally redacts sensitive data in log events. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean id: description: The unique identifier for this component. Used to reference this component in other parts of the pipeline (e.g., as input to downstream @@ -31806,6 +31874,10 @@ components: description: The `throttle` processor limits the number of events that pass through over a given time window. properties: + enabled: + description: The processor passes through all events if it is set to `false`. + Defaults to `true`. + type: boolean group_by: description: Optional list of fields used to group events before the threshold has been reached. diff --git a/src/datadog_api_client/v2/model/observability_pipeline_add_env_vars_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_add_env_vars_processor.py index ab11ebe286..06f36d82b6 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_add_env_vars_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_add_env_vars_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -31,6 +33,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "id": (str,), "include": (str,), "inputs": ([str],), @@ -39,6 +42,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "id": "id", "include": "include", "inputs": "inputs", @@ -53,11 +57,15 @@ def __init__( inputs: List[str], type: ObservabilityPipelineAddEnvVarsProcessorType, variables: List[ObservabilityPipelineAddEnvVarsProcessorVariable], + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ The ``add_env_vars`` processor adds environment variable values to log events. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param id: The unique identifier for this component. Used to reference this processor in the pipeline. :type id: str @@ -73,6 +81,8 @@ def __init__( :param variables: A list of environment variable mappings to apply to log fields. :type variables: [ObservabilityPipelineAddEnvVarsProcessorVariable] """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.id = id diff --git a/src/datadog_api_client/v2/model/observability_pipeline_add_fields_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_add_fields_processor.py index 023b7a6d57..93bab2adba 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_add_fields_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_add_fields_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -27,6 +29,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "fields": ([ObservabilityPipelineFieldValue],), "id": (str,), "include": (str,), @@ -35,6 +38,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "fields": "fields", "id": "id", "include": "include", @@ -49,11 +53,15 @@ def __init__( include: str, inputs: List[str], type: ObservabilityPipelineAddFieldsProcessorType, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ The ``add_fields`` processor adds static key-value fields to logs. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param fields: A list of static fields (key-value pairs) that is added to each log event processed by this component. :type fields: [ObservabilityPipelineFieldValue] @@ -69,6 +77,8 @@ def __init__( :param type: The processor type. The value should always be ``add_fields``. :type type: ObservabilityPipelineAddFieldsProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.fields = fields diff --git a/src/datadog_api_client/v2/model/observability_pipeline_config_processor_item.py b/src/datadog_api_client/v2/model/observability_pipeline_config_processor_item.py index 11201ae45d..ce6981fa7a 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_config_processor_item.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_config_processor_item.py @@ -15,6 +15,9 @@ def __init__(self, **kwargs): """ A processor for the pipeline. + :param enabled: The processor passes through all events if it is set to `false`. Defaults to `true`. + :type enabled: bool, optional + :param id: The unique identifier for this component. Used to reference this component in other parts of the pipeline (for example, as the `input` to downstream components). :type id: str diff --git a/src/datadog_api_client/v2/model/observability_pipeline_custom_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_custom_processor.py index 214d511ebc..2f508c6c7d 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_custom_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_custom_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -37,6 +39,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "id": (str,), "include": (str,), "inputs": ([str],), @@ -45,6 +48,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "id": "id", "include": "include", "inputs": "inputs", @@ -58,11 +62,15 @@ def __init__( inputs: List[str], remaps: List[ObservabilityPipelineCustomProcessorRemap], type: ObservabilityPipelineCustomProcessorType, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ The ``custom_processor`` processor transforms events using `Vector Remap Language (VRL) `_ scripts with advanced filtering capabilities. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param id: The unique identifier for this processor. :type id: str @@ -78,6 +86,8 @@ def __init__( :param type: The processor type. The value should always be ``custom_processor``. :type type: ObservabilityPipelineCustomProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) include = kwargs.get("include", "*") diff --git a/src/datadog_api_client/v2/model/observability_pipeline_datadog_tags_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_datadog_tags_processor.py index 23cedc4d69..e429dfc41d 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_datadog_tags_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_datadog_tags_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -38,6 +40,7 @@ def openapi_types(_): return { "action": (ObservabilityPipelineDatadogTagsProcessorAction,), + "enabled": (bool,), "id": (str,), "include": (str,), "inputs": ([str],), @@ -48,6 +51,7 @@ def openapi_types(_): attribute_map = { "action": "action", + "enabled": "enabled", "id": "id", "include": "include", "inputs": "inputs", @@ -65,6 +69,7 @@ def __init__( keys: List[str], mode: ObservabilityPipelineDatadogTagsProcessorMode, type: ObservabilityPipelineDatadogTagsProcessorType, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ @@ -73,6 +78,9 @@ def __init__( :param action: The action to take on tags with matching keys. :type action: ObservabilityPipelineDatadogTagsProcessorAction + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param id: The unique identifier for this component. Used to reference this component in other parts of the pipeline (for example, as the ``input`` to downstream components). :type id: str @@ -91,6 +99,8 @@ def __init__( :param type: The processor type. The value should always be ``datadog_tags``. :type type: ObservabilityPipelineDatadogTagsProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.action = action diff --git a/src/datadog_api_client/v2/model/observability_pipeline_dedupe_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_dedupe_processor.py index 9021940cfc..a5f07f4005 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_dedupe_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_dedupe_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -31,6 +33,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "fields": ([str],), "id": (str,), "include": (str,), @@ -40,6 +43,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "fields": "fields", "id": "id", "include": "include", @@ -56,11 +60,15 @@ def __init__( inputs: List[str], mode: ObservabilityPipelineDedupeProcessorMode, type: ObservabilityPipelineDedupeProcessorType, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ The ``dedupe`` processor removes duplicate fields in log events. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param fields: A list of log field paths to check for duplicates. :type fields: [str] @@ -79,6 +87,8 @@ def __init__( :param type: The processor type. The value should always be ``dedupe``. :type type: ObservabilityPipelineDedupeProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.fields = fields diff --git a/src/datadog_api_client/v2/model/observability_pipeline_enrichment_table_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_enrichment_table_processor.py index 480f484822..118d146af7 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_enrichment_table_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_enrichment_table_processor.py @@ -39,6 +39,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "file": (ObservabilityPipelineEnrichmentTableFile,), "geoip": (ObservabilityPipelineEnrichmentTableGeoIp,), "id": (str,), @@ -49,6 +50,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "file": "file", "geoip": "geoip", "id": "id", @@ -65,6 +67,7 @@ def __init__( inputs: List[str], target: str, type: ObservabilityPipelineEnrichmentTableProcessorType, + enabled: Union[bool, UnsetType] = unset, file: Union[ObservabilityPipelineEnrichmentTableFile, UnsetType] = unset, geoip: Union[ObservabilityPipelineEnrichmentTableGeoIp, UnsetType] = unset, **kwargs, @@ -72,6 +75,9 @@ def __init__( """ The ``enrichment_table`` processor enriches logs using a static CSV file or GeoIP database. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param file: Defines a static enrichment table loaded from a CSV file. :type file: ObservabilityPipelineEnrichmentTableFile, optional @@ -93,6 +99,8 @@ def __init__( :param type: The processor type. The value should always be ``enrichment_table``. :type type: ObservabilityPipelineEnrichmentTableProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled if file is not unset: kwargs["file"] = file if geoip is not unset: diff --git a/src/datadog_api_client/v2/model/observability_pipeline_filter_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_filter_processor.py index d7ada2fb81..08ea44b71e 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_filter_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_filter_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -25,6 +27,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "id": (str,), "include": (str,), "inputs": ([str],), @@ -32,6 +35,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "id": "id", "include": "include", "inputs": "inputs", @@ -39,11 +43,20 @@ def openapi_types(_): } def __init__( - self_, id: str, include: str, inputs: List[str], type: ObservabilityPipelineFilterProcessorType, **kwargs + self_, + id: str, + include: str, + inputs: List[str], + type: ObservabilityPipelineFilterProcessorType, + enabled: Union[bool, UnsetType] = unset, + **kwargs, ): """ The ``filter`` processor allows conditional processing of logs based on a Datadog search query. Logs that match the ``include`` query are passed through; others are discarded. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param id: The unique identifier for this component. Used to reference this component in other parts of the pipeline (for example, as the ``input`` to downstream components). :type id: str @@ -56,6 +69,8 @@ def __init__( :param type: The processor type. The value should always be ``filter``. :type type: ObservabilityPipelineFilterProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.id = id diff --git a/src/datadog_api_client/v2/model/observability_pipeline_generate_metrics_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_generate_metrics_processor.py index b0eee92908..39ff81896b 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_generate_metrics_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_generate_metrics_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -29,6 +31,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "id": (str,), "include": (str,), "inputs": ([str],), @@ -37,6 +40,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "id": "id", "include": "include", "inputs": "inputs", @@ -51,12 +55,16 @@ def __init__( inputs: List[str], metrics: List[ObservabilityPipelineGeneratedMetric], type: ObservabilityPipelineGenerateMetricsProcessorType, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ The ``generate_datadog_metrics`` processor creates custom metrics from logs and sends them to Datadog. Metrics can be counters, gauges, or distributions and optionally grouped by log fields. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param id: The unique identifier for this component. Used to reference this component in other parts of the pipeline. :type id: str @@ -72,6 +80,8 @@ def __init__( :param type: The processor type. Always ``generate_datadog_metrics``. :type type: ObservabilityPipelineGenerateMetricsProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.id = id diff --git a/src/datadog_api_client/v2/model/observability_pipeline_ocsf_mapper_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_ocsf_mapper_processor.py index 7917f1c81e..5a374b33d1 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_ocsf_mapper_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_ocsf_mapper_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -31,6 +33,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "id": (str,), "include": (str,), "inputs": ([str],), @@ -39,6 +42,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "id": "id", "include": "include", "inputs": "inputs", @@ -53,11 +57,15 @@ def __init__( inputs: List[str], mappings: List[ObservabilityPipelineOcsfMapperProcessorMapping], type: ObservabilityPipelineOcsfMapperProcessorType, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ The ``ocsf_mapper`` processor transforms logs into the OCSF schema using a predefined mapping configuration. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param id: The unique identifier for this component. Used to reference this component in other parts of the pipeline. :type id: str @@ -73,6 +81,8 @@ def __init__( :param type: The processor type. The value should always be ``ocsf_mapper``. :type type: ObservabilityPipelineOcsfMapperProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.id = id diff --git a/src/datadog_api_client/v2/model/observability_pipeline_parse_grok_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_parse_grok_processor.py index c7b2f65cfc..46219be1cd 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_parse_grok_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_parse_grok_processor.py @@ -34,6 +34,7 @@ def openapi_types(_): return { "disable_library_rules": (bool,), + "enabled": (bool,), "id": (str,), "include": (str,), "inputs": ([str],), @@ -43,6 +44,7 @@ def openapi_types(_): attribute_map = { "disable_library_rules": "disable_library_rules", + "enabled": "enabled", "id": "id", "include": "include", "inputs": "inputs", @@ -58,6 +60,7 @@ def __init__( rules: List[ObservabilityPipelineParseGrokProcessorRule], type: ObservabilityPipelineParseGrokProcessorType, disable_library_rules: Union[bool, UnsetType] = unset, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ @@ -66,6 +69,9 @@ def __init__( :param disable_library_rules: If set to ``true`` , disables the default Grok rules provided by Datadog. :type disable_library_rules: bool, optional + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param id: A unique identifier for this processor. :type id: str @@ -83,6 +89,8 @@ def __init__( """ if disable_library_rules is not unset: kwargs["disable_library_rules"] = disable_library_rules + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.id = id diff --git a/src/datadog_api_client/v2/model/observability_pipeline_parse_json_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_parse_json_processor.py index ec3357b625..183f4e6559 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_parse_json_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_parse_json_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -25,6 +27,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "field": (str,), "id": (str,), "include": (str,), @@ -33,6 +36,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "field": "field", "id": "id", "include": "include", @@ -47,11 +51,15 @@ def __init__( include: str, inputs: List[str], type: ObservabilityPipelineParseJSONProcessorType, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ The ``parse_json`` processor extracts JSON from a specified field and flattens it into the event. This is useful when logs contain embedded JSON as a string. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param field: The name of the log field that contains a JSON string. :type field: str @@ -67,6 +75,8 @@ def __init__( :param type: The processor type. The value should always be ``parse_json``. :type type: ObservabilityPipelineParseJSONProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.field = field diff --git a/src/datadog_api_client/v2/model/observability_pipeline_quota_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_quota_processor.py index aa4be5a8ca..9e45063e5d 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_quota_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_quota_processor.py @@ -46,6 +46,7 @@ def openapi_types(_): return { "drop_events": (bool,), + "enabled": (bool,), "id": (str,), "ignore_when_missing_partitions": (bool,), "include": (str,), @@ -60,6 +61,7 @@ def openapi_types(_): attribute_map = { "drop_events": "drop_events", + "enabled": "enabled", "id": "id", "ignore_when_missing_partitions": "ignore_when_missing_partitions", "include": "include", @@ -81,6 +83,7 @@ def __init__( limit: ObservabilityPipelineQuotaProcessorLimit, name: str, type: ObservabilityPipelineQuotaProcessorType, + enabled: Union[bool, UnsetType] = unset, ignore_when_missing_partitions: Union[bool, UnsetType] = unset, overflow_action: Union[ObservabilityPipelineQuotaProcessorOverflowAction, UnsetType] = unset, overrides: Union[List[ObservabilityPipelineQuotaProcessorOverride], UnsetType] = unset, @@ -93,6 +96,9 @@ def __init__( :param drop_events: If set to ``true`` , logs that matched the quota filter and sent after the quota has been met are dropped; only logs that did not match the filter query continue through the pipeline. :type drop_events: bool + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param id: The unique identifier for this component. Used to reference this component in other parts of the pipeline (for example, as the ``input`` to downstream components). :type id: str @@ -127,6 +133,8 @@ def __init__( :param type: The processor type. The value should always be ``quota``. :type type: ObservabilityPipelineQuotaProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled if ignore_when_missing_partitions is not unset: kwargs["ignore_when_missing_partitions"] = ignore_when_missing_partitions if overflow_action is not unset: diff --git a/src/datadog_api_client/v2/model/observability_pipeline_reduce_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_reduce_processor.py index bcddc2844d..bcc2514ca0 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_reduce_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_reduce_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -31,6 +33,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "group_by": ([str],), "id": (str,), "include": (str,), @@ -40,6 +43,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "group_by": "group_by", "id": "id", "include": "include", @@ -56,11 +60,15 @@ def __init__( inputs: List[str], merge_strategies: List[ObservabilityPipelineReduceProcessorMergeStrategy], type: ObservabilityPipelineReduceProcessorType, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ The ``reduce`` processor aggregates and merges logs based on matching keys and merge strategies. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param group_by: A list of fields used to group log events for merging. :type group_by: [str] @@ -79,6 +87,8 @@ def __init__( :param type: The processor type. The value should always be ``reduce``. :type type: ObservabilityPipelineReduceProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.group_by = group_by diff --git a/src/datadog_api_client/v2/model/observability_pipeline_remove_fields_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_remove_fields_processor.py index c8a0d89bc0..7c0e3013da 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_remove_fields_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_remove_fields_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -25,6 +27,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "fields": ([str],), "id": (str,), "include": (str,), @@ -33,6 +36,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "fields": "fields", "id": "id", "include": "include", @@ -47,11 +51,15 @@ def __init__( include: str, inputs: List[str], type: ObservabilityPipelineRemoveFieldsProcessorType, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ The ``remove_fields`` processor deletes specified fields from logs. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param fields: A list of field names to be removed from each log event. :type fields: [str] @@ -67,6 +75,8 @@ def __init__( :param type: The processor type. The value should always be ``remove_fields``. :type type: ObservabilityPipelineRemoveFieldsProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.fields = fields diff --git a/src/datadog_api_client/v2/model/observability_pipeline_rename_fields_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_rename_fields_processor.py index ae737aceb1..0719783d98 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_rename_fields_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_rename_fields_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -31,6 +33,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "fields": ([ObservabilityPipelineRenameFieldsProcessorField],), "id": (str,), "include": (str,), @@ -39,6 +42,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "fields": "fields", "id": "id", "include": "include", @@ -53,11 +57,15 @@ def __init__( include: str, inputs: List[str], type: ObservabilityPipelineRenameFieldsProcessorType, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ The ``rename_fields`` processor changes field names. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param fields: A list of rename rules specifying which fields to rename in the event, what to rename them to, and whether to preserve the original fields. :type fields: [ObservabilityPipelineRenameFieldsProcessorField] @@ -73,6 +81,8 @@ def __init__( :param type: The processor type. The value should always be ``rename_fields``. :type type: ObservabilityPipelineRenameFieldsProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.fields = fields diff --git a/src/datadog_api_client/v2/model/observability_pipeline_sample_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_sample_processor.py index e6f16ce99d..20ee1d6f7a 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_sample_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_sample_processor.py @@ -33,6 +33,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "id": (str,), "include": (str,), "inputs": ([str],), @@ -42,6 +43,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "id": "id", "include": "include", "inputs": "inputs", @@ -56,6 +58,7 @@ def __init__( include: str, inputs: List[str], type: ObservabilityPipelineSampleProcessorType, + enabled: Union[bool, UnsetType] = unset, percentage: Union[float, UnsetType] = unset, rate: Union[int, UnsetType] = unset, **kwargs, @@ -63,6 +66,9 @@ def __init__( """ The ``sample`` processor allows probabilistic sampling of logs at a fixed rate. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param id: The unique identifier for this component. Used to reference this component in other parts of the pipeline (for example, as the ``input`` to downstream components). :type id: str @@ -81,6 +87,8 @@ def __init__( :param type: The processor type. The value should always be ``sample``. :type type: ObservabilityPipelineSampleProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled if percentage is not unset: kwargs["percentage"] = percentage if rate is not unset: diff --git a/src/datadog_api_client/v2/model/observability_pipeline_sensitive_data_scanner_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_sensitive_data_scanner_processor.py index 9d78e97932..c2e0a261ea 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_sensitive_data_scanner_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_sensitive_data_scanner_processor.py @@ -3,11 +3,13 @@ # Copyright 2019-Present Datadog, Inc. from __future__ import annotations -from typing import List, TYPE_CHECKING +from typing import List, Union, TYPE_CHECKING from datadog_api_client.model_utils import ( ModelNormal, cached_property, + unset, + UnsetType, ) @@ -31,6 +33,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "id": (str,), "include": (str,), "inputs": ([str],), @@ -39,6 +42,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "id": "id", "include": "include", "inputs": "inputs", @@ -53,11 +57,15 @@ def __init__( inputs: List[str], rules: List[ObservabilityPipelineSensitiveDataScannerProcessorRule], type: ObservabilityPipelineSensitiveDataScannerProcessorType, + enabled: Union[bool, UnsetType] = unset, **kwargs, ): """ The ``sensitive_data_scanner`` processor detects and optionally redacts sensitive data in log events. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param id: The unique identifier for this component. Used to reference this component in other parts of the pipeline (e.g., as input to downstream components). :type id: str @@ -73,6 +81,8 @@ def __init__( :param type: The processor type. The value should always be ``sensitive_data_scanner``. :type type: ObservabilityPipelineSensitiveDataScannerProcessorType """ + if enabled is not unset: + kwargs["enabled"] = enabled super().__init__(kwargs) self_.id = id diff --git a/src/datadog_api_client/v2/model/observability_pipeline_throttle_processor.py b/src/datadog_api_client/v2/model/observability_pipeline_throttle_processor.py index aeaa56e6cb..c89b96c4e9 100644 --- a/src/datadog_api_client/v2/model/observability_pipeline_throttle_processor.py +++ b/src/datadog_api_client/v2/model/observability_pipeline_throttle_processor.py @@ -27,6 +27,7 @@ def openapi_types(_): ) return { + "enabled": (bool,), "group_by": ([str],), "id": (str,), "include": (str,), @@ -37,6 +38,7 @@ def openapi_types(_): } attribute_map = { + "enabled": "enabled", "group_by": "group_by", "id": "id", "include": "include", @@ -54,12 +56,16 @@ def __init__( threshold: int, type: ObservabilityPipelineThrottleProcessorType, window: float, + enabled: Union[bool, UnsetType] = unset, group_by: Union[List[str], UnsetType] = unset, **kwargs, ): """ The ``throttle`` processor limits the number of events that pass through over a given time window. + :param enabled: The processor passes through all events if it is set to ``false``. Defaults to ``true``. + :type enabled: bool, optional + :param group_by: Optional list of fields used to group events before the threshold has been reached. :type group_by: [str], optional @@ -81,6 +87,8 @@ def __init__( :param window: The time window in seconds over which the threshold applies. :type window: float """ + if enabled is not unset: + kwargs["enabled"] = enabled if group_by is not unset: kwargs["group_by"] = group_by super().__init__(kwargs)