diff --git a/tests/handlers/aws/test_integrations.py b/tests/handlers/aws/test_integrations.py index a957754f..e545138b 100644 --- a/tests/handlers/aws/test_integrations.py +++ b/tests/handlers/aws/test_integrations.py @@ -962,33 +962,33 @@ def test_replay(self) -> None: { "fail": { "message": "test_replay_fail_pipeline_s3", - "if": f"ctx[\"_id\"] == \"{prefix_s3_first}-000000000000\"", + "if": f'ctx["_id"] == "{prefix_s3_first}-000000000000"', } }, { "fail": { "message": "test_replay_fail_pipeline_sqs", - "if": f"ctx[\"_id\"] == \"{prefix_sqs}-000000000000\"", + "if": f'ctx["_id"] == "{prefix_sqs}-000000000000"', } }, { "fail": { "message": "test_replay_fail_pipeline_cloudwatch", - "if": f"ctx[\"_id\"] == \"{prefix_cloudwatch_logs}-000000000000\"", + "if": f'ctx["_id"] == "{prefix_cloudwatch_logs}-000000000000"', } }, { "fail": { "message": "test_replay_fail_pipeline_kinesis", - "if": f"ctx[\"_id\"] == \"{prefix_kinesis}-000000000000\"", + "if": f'ctx["_id"] == "{prefix_kinesis}-000000000000"', } }, ] } self.elasticsearch.put_pipeline(id="test_replay_fail_pipeline", body=processors) - - self.elasticsearch.create_data_stream(index='logs-generic-default') + + self.elasticsearch.create_data_stream(index="logs-generic-default") self.elasticsearch.put_settings( index="logs-generic-default", body={"index.default_pipeline": "test_replay_fail_pipeline"} ) @@ -1120,9 +1120,7 @@ def test_replay(self) -> None: self.elasticsearch.refresh(index="logs-generic-default") # Remove pipeline processors - processors = { - "processors": [] - } + processors = {"processors": []} self.elasticsearch.put_pipeline(id="test_replay_fail_pipeline", body=processors) self.elasticsearch.refresh(index="logs-generic-default") diff --git a/tests/testcontainers/es.py b/tests/testcontainers/es.py index 149b2c07..57f6fe82 100644 --- a/tests/testcontainers/es.py +++ b/tests/testcontainers/es.py @@ -209,7 +209,7 @@ def index(self, **kwargs: Any) -> dict[str, Any]: self._index_indices.add(kwargs["index"]) return self.es_client.index(**kwargs) - + def create_data_stream(self, **kwargs: Any) -> dict[str, Any]: if "index" in kwargs: self._index_indices.add(kwargs["index"])