build(ingest): support flake8 6.0.0 (#6540)

This commit is contained in:
Harshal Sheth 2022-11-23 17:40:55 -05:00 committed by GitHub
parent 0eb54c6387
commit ce3f663a57
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
4 changed files with 3 additions and 6 deletions

View File

@ -631,7 +631,6 @@ timestamp < "{end_time}"
return None return None
def test_capability(self, project_id: str) -> None: def test_capability(self, project_id: str) -> None:
lineage_metadata: Dict[str, Set[str]]
if self.config.use_exported_bigquery_audit_metadata: if self.config.use_exported_bigquery_audit_metadata:
bigquery_client: BigQueryClient = BigQueryClient(project=project_id) bigquery_client: BigQueryClient = BigQueryClient(project=project_id)
entries = self._get_exported_bigquery_audit_metadata( entries = self._get_exported_bigquery_audit_metadata(

View File

@ -4,7 +4,7 @@ import textwrap
import time import time
from dataclasses import dataclass from dataclasses import dataclass
from datetime import datetime from datetime import datetime
from typing import Any, Dict, Iterable, List, MutableMapping, Optional, Set, Union, cast from typing import Any, Dict, Iterable, List, MutableMapping, Optional, Union, cast
import cachetools import cachetools
from google.cloud.bigquery import Client as BigQueryClient from google.cloud.bigquery import Client as BigQueryClient
@ -443,7 +443,6 @@ class BigQueryUsageExtractor:
) -> Optional[OperationalDataMeta]: ) -> Optional[OperationalDataMeta]:
# If we don't have Query object that means this is a queryless read operation or a read operation which was not executed as JOB # If we don't have Query object that means this is a queryless read operation or a read operation which was not executed as JOB
# https://cloud.google.com/bigquery/docs/reference/auditlogs/rest/Shared.Types/BigQueryAuditMetadata.TableDataRead.Reason/ # https://cloud.google.com/bigquery/docs/reference/auditlogs/rest/Shared.Types/BigQueryAuditMetadata.TableDataRead.Reason/
operation_meta: OperationalDataMeta
if not event.query_event and event.read_event: if not event.query_event and event.read_event:
return OperationalDataMeta( return OperationalDataMeta(
statement_type=OperationTypeClass.CUSTOM, statement_type=OperationTypeClass.CUSTOM,
@ -839,7 +838,6 @@ class BigQueryUsageExtractor:
) )
def test_capability(self, project_id: str) -> None: def test_capability(self, project_id: str) -> None:
lineage_metadata: Dict[str, Set[str]]
for entry in self._get_parsed_bigquery_log_events(project_id, limit=1): for entry in self._get_parsed_bigquery_log_events(project_id, limit=1):
logger.debug(f"Connection test got one {entry}") logger.debug(f"Connection test got one {entry}")
return return

View File

@ -141,7 +141,8 @@ class ConfluentSchemaRegistry(KafkaSchemaRegistryBase):
if schema_seen is None: if schema_seen is None:
schema_seen = set() schema_seen = set()
for schema_ref in schema.references: # type: SchemaReference schema_ref: SchemaReference
for schema_ref in schema.references:
ref_subject: str = schema_ref["subject"] ref_subject: str = schema_ref["subject"]
if ref_subject in schema_seen: if ref_subject in schema_seen:
continue continue

View File

@ -83,7 +83,6 @@ class DataHubOperationCircuitBreakerOperator(BaseSensorOperator):
else: else:
raise Exception(f"urn parameter has invalid type {type(self.urn)}") raise Exception(f"urn parameter has invalid type {type(self.urn)}")
partition: Optional[str]
for urn in urns: for urn in urns:
self.log.info(f"Checking if dataset {self.urn} is ready to be consumed") self.log.info(f"Checking if dataset {self.urn} is ready to be consumed")
ret = self.circuit_breaker.is_circuit_breaker_active( ret = self.circuit_breaker.is_circuit_breaker_active(