Data Sync Events
This page offers a detailed overview of the connecting components within the BDB Data Pipeline, specifically focusing on Data Sync Events.
What is a Data Sync Event?
A Data Sync Event in the Data Pipeline module is designed to simplify and accelerate the process of writing data directly to a configured database without the need to introduce additional components in the pipeline flow. This streamlines the process by reducing the number of writer components within the pipeline, offering a centralized and seamless method for data writing.
Benefits of using the Data Sync Event
Pre-requisites: Before creating the Data Sync Event, the user has to configure the driver settings in the Data Sync section under the Pipeline Settings admin module.
Once the Data Sync settings are configured, multiple Data Sync events can be created for the same driver configuration, and the data can be written to multiple tables.
Creating and Configuring a Data Sync Event (from the Events Panel)
Follow the steps below to create and configure a Data Sync Event in the Pipeline Workflow Editor.
Create a Data Sync Event
The Create Sync Event window opens.
Provide the required details:
Display Name: Enter a name for the new Data Sync.
Driver: Select a driver from the drop-down list (only pre-configured drivers appear).
Is Failover: Check this option if you want to create a failover Data Sync. (In this example, leave it unchecked.)
Click Add Sync Event.
A new Data Sync is created and appears under the DB Sync section in the Events tab.
Configure Data Sync Properties
Select the Data Sync Event on the canvas.
Open the Basic Info tab.
Provide the following information:
Table Name: Specify the database table name.
Driver: Select a driver using the dropdown menu. All the configured drivers for the data sync event will be listed.
Save Mode: Select one of the following from the drop-down list:
Append: Adds new records without affecting existing ones.
Upsert: Updates existing records based on a key and inserts new ones.
Composite Key: (Optional) Appears only when Upsert is selected as the save mode. Use it to define the key fields for record updates.
Click the Save Data Sync icon to save your configuration.
A success message confirms that the Data Sync information has been saved.
Auto Connecting Data Sync
This feature automatically connects the Kafka Event to the component when it is dragged from the Events panel. To use this feature, users must ensure that the "Auto connect components on drag" option is enabled in the Events panel (it comes enabled in the default setting).
Mapping Failover Data Sync to a Component
The Failover Data Sync Event component is a safeguard mechanism in resilient pipeline designs. It ensures data reliability and consistency across systems during failovers, enabling smooth transitions and uninterrupted operations in mission-critical environments.
Prerequisites
A pipeline workflow is available in the Pipeline Workflow Editor.
The Failover Data Sync Event is already available under the DB Sync section in the Events tab.
Follow the steps below to configure and map a Failover Data Sync Event:
Configure Failover Data Sync
Click on the Failover Data Sync event in the workflow editor.
In the Basic Info tab, provide the following details:
Table Name: Specify the table where failed data should be written.
Save Mode: Choose from the available options:
Append – Adds failed records to the existing table.
Upsert – Inserts or updates records in the target table.
Click the Save Data Sync icon.
A success message confirms that the Failover Data Sync has been configured.
Map Failover Data Sync with Pipeline Component
Open the pipeline component in the existing workflow that requires failover handling.
In the Basic Information tab of the selected component, locate the Failover Event drop-down menu.
Select the configured Failover Data Sync Event.
Save the component.
Upon success, update the pipeline.
Activate the pipeline workflow once the update is successful.
Validate Configuration & Logs
Open the Log Panel after pipeline activation.
Confirm pipeline activation via the success message.
Navigate to the Logs tab within the panel.
Verify notification messages indicating that the Failover Data Sync Event has been triggered.
For example, if mapped with a ClickHouse Reader component, any failure will result in the failed records being written to the designated failover table.
The logs explicitly confirm when data is written to the failover table.
Adding a Data Sync Event to a Pipeline Workflow
The Data Sync Event feature allows users to establish a direct connection between a pipeline component and a synchronization event. This can be performed quickly by using the context menu on the pipeline workspace canvas.
Steps to Add a Data Sync Event