* added trino integration test
* - removed warnings for classes which are not real tests
- removed "helpers" as its being used
* use a docker network instead of host
* print logs for hive failure
* removed superset unit tests
* try pinning requests for test
* try pinning requests for test
* wait for hive to be ready
* fix trino fixture
* - reduced testcontainers_config.max_tries to 5
- remove intermediate containers
* print with logs
* disable capture logging
* updated db host
* removed debug stuff
* removed debug stuff
* removed version pin for requests
* reverted superset
* ignore trino integration on python 3.8
* MINOR - Add Integration Test for S3 Storage
* MINOR - Add Integration Test for S3 Storage
* MINOR - Add Integration Test for S3 Storage
* format
* format
* feat(nosql-profiler): row count
1. Implemented the NoSQLProfilerInterface as an entrypoint for the nosql profiler.
2. Added the NoSQLMetric as an abstract class.
3. Implemented the interface for the MongoDB database source.
4. Implemented an e2e test using testcontainers.
* added profiler support for mongodb connection
* doc
* use int_admin_ometa in test setup
* - fixed linting issue in gx
- removed unused inheritance
* moved the nosql function into the metric class
* feat(profiler): add dynamodb row count
* feat(profiler): add dynamodb row count
* formatting
* validate_compose: raise exception for bad status code.
* fixed import
* format
* feat(nosql-profiler): added sample data
1. Implemented the NoSQL sampler.
2. Some naming changes to the NoSQL adaptor to avoid fixing names with the profiler interface.
3. Tests.
* added default sample limit
* formatting
* fixed import
* feat(profiler): dynamodb sample data
* tests for dynamo db sample data
* format
* format
* use service connection for nosql adaptor factory
* fixed tests
* format
* fixed after merge
* Changed the MLModelStore storage type to string
* fix checkstyle
* remove unused files
* Update requirements
* fix checkstyle
* Skipping MLFlow intergration on python 3.8
* Hack to allow pytest to parse the mlflow integrations test on python 3.8
* Fix checkstyle
* feat(nosql-profiler): row count
1. Implemented the NoSQLProfilerInterface as an entrypoint for the nosql profiler.
2. Added the NoSQLMetric as an abstract class.
3. Implemented the interface for the MongoDB database source.
4. Implemented an e2e test using testcontainers.
* added profiler support for mongodb connection
* doc
* use int_admin_ometa in test setup
* - fixed linting issue in gx
- removed unused inheritance
* moved the nosql function into the metric class
* feat(profiler): add dynamodb row count
* feat(profiler): add dynamodb row count
* formatting
* fixed import
* format
* dded dynamodb row count
* format
* removed unused factory file
* removed "validate"
* migrations
* removed validations
* format
* linting
* fixed: test_amundsen.py
* Update schemaChanges.sql
* WIP - MINOR - Clean ingestion dependencies
* test
* test
* Clean imports
* add pyiceberg for test
* Revert "add pyiceberg for test"
This reverts commit ab26942736586f089a57a644ffd727aca200db62.
* add pyiceberg for test
* Remove docker dep
* clean local docker sh
* MINOR - AKS Airflow troubleshooting docs
* Fix action
* clean local docker sh
* feat(nosql-profiler): row count
1. Implemented the NoSQLProfilerInterface as an entrypoint for the nosql profiler.
2. Added the NoSQLMetric as an abstract class.
3. Implemented the interface for the MongoDB database source.
4. Implemented an e2e test using testcontainers.
* added profiler support for mongodb connection
* doc
* use int_admin_ometa in test setup
* - fixed linting issue in gx
- removed unused inheritance
* moved the nosql function into the metric class
* formatting
* validate_compose: raise exception for bad status code.
* fixed import
* format
* feat(connector): add BigTable
* bigtable work
1. docstrings
2. tests
3. created a Row BaseModel
4. implemented a ClassConverter
* docs moved to separate PR
* format files
* minor cosmetic
- removed TODO
- changed headers' year to 2024 for new files
- fixed typos
* format
* formatting and comments
1. added missing docstrings.
2. abstracted the _find_instance method.
3. aliased the IDs used in the BigTable connection
* added comment regarding private key
* added comments regarding column families
* enclose get_schema_name_list in `try/except/else`
* format
* streamlined get_schema_name_list to include all logic in the try block
* Create the iceberg connection schema
* Link the IcebergConnection configuration with the forms on the UI
* Add the pyiceberg dependency on the ingestion package
* Create the get_connection and test_connection functions
* First iteration on the iceberg ingestion logic
* Add A more comprehensive implementation of the Iceberg Source
* Add UnitTests
* Update icebergConnection definition
* Update the iceberg souce code based on new schema
* Updated icebergConnecgtion schema for simplicity and to be able to configure Converters
* Updated setup dependencies to be more flexible
* Updated get_owner_ref logic
* Fix formatting
* Changed the icebergConnection json schema structure to enable the ClassConverters
* Add the IcebergCatalog and IcebergFileSystem ClassConverters
* Refactor the code to take into account the new jsonSchema structure
* Fix formatting
* Add Documentation for the Iceberg Connector
* Fix Menu order for Iceberg
* ui: add Iceberg service icon and constant
* Fix DynamoDb Catalog issue due to how PyIceberg instantes it
* Changed uri title to URI
* Fix ClassConverter for Iceberg
* Fix GetSecretValue for password types
* Fix formatting
* Fix formatting
* Add Iceberg Connector Images for the docs
* Add pylint disable for Hacky super() call
* Add Iceberg.md for the UI docs
* Fix pylint complaint
* Fix pylint complaint
* Fix UnitTests
* fix type error and unit tests
* update pipeline type checks
* Fix Sonar Cloud complaints
---------
Co-authored-by: Sachin Chaurasiya <sachinchaurasiyachotey87@gmail.com>
* feat: SAS Viya connector
* refactor SASCatalog to SAS
* add SAS logo to UI and connection documentation
* doc changes
* modify ingestion logic
* revert original changes
* added support for dataflow & perfect logic for reports/datatables
* add filter doc
* more updates to perfect ingestion for each asset type
* fix a bug with table lineage not created properly
* Delete ingestion/pipelines/sasCatalog.yaml
* precomit fix
* Conversion to database connector
* minor fixes
* make custom properties type generic
* Add SAS javaEnum
* add dummy variable for sas.yaml
---------
Co-authored-by: lizmc <liz.mcintosh@sas.com>
Co-authored-by: Shiyang Xiao <Shiyang.Xiao@sas.com>