Ingestion Process
Requirement | Evaluation | Remarks |
---|---|---|
Change data capture | medium | Integration with thirdparty tools like Debezium can be provided |
Scheduled ingestion | very high | |
Minimise fields | very high | Field minimization can be defined through the Data Preparation tool, and published to be used in the live Data Pipelines. |
Filter by lookup | very high | Yes, it is a standard component. |
Filter by consent | very high | It can be achieved via API integration with consent system, or through consent database lookup. |
Anonymise fields | very high | Standard anonymization available via Data Preparation option or the Spark SQL component. |
Compose Ingestion Processors | very high | Drag and drop based low-code platform |
Ingestion Fault tolerant | very high | Ability to track faults and initiate sub process |
Bootstrap + updates | very high | Can define pipeline to load historic data and subsequent updates as per the data load strategy |
Reports + Metrics | high | |
performance impact threshold | high | Configurable compute resource allocation and instances to scale up |
Secrety Management integration | very high | All secrets are stored in the Kubernestes secrets, platform provides direct integration with this. |
Data Catalogue integration | very high | Platform automatically generate data catalog from the underlying meta data. |
Visual interface | very high | |
Ingestion Manifest file | very high | It is achievable via internal metadata. |
CI/CD Pipelines Integration | high | Yes, it provides facility to check-in and check-out Pipeline definitions and metadata to GIT Lab. |
Ingestion Access Management | very high | Data Pipeline supports RBAC. |
Ingestion Audit Logs | very high | Logs can be pushed to thirdparty log monitoring systems like Datadog, Promethues, etc. |
Last updated