DB Writer
Last updated
Last updated
The DB reader is a spark-based writer component which gives you capability to write data to multiple database sources.
All component configurations are classified broadly into the following sections:
Meta Information
Please check out the given demonstration to configure the component.
Please Note:
The ClickHouse driver in the Spark components will use the HTTP Port and not the TCP port.
It is always recommended to create the table before activating the pipeline to avoid errors as RDBMS has a strict schema and can result in errors.
When using the Redshift driver with a Boolean datatype in JDBC, the table is not created unless you pass the create table query. Alternatively, you can use a column filter to convert a Boolean value to a String for the desired operation.
The RDBMS writer supports 3 save modes:
As the name suggests it adds all the records without any validations.
This mode truncates the table and adds fresh records. after every run you will get records that are part of the batch process.
This operation allows the users to insert a new record or update existing data into a table. For configuring this we need to provide the Composite Key.
The BDB Data Pipeline supports composite key based upsert, in case of composite key, we can specify the second key by using comma separator e.g., key1, key2. It has now an option to upload the spark schema in JSON format. This can greatly improve the speed of the write operation as the component will ignore inferring schema and go with the provided schema.
Sort Column: This field will appear only when Upsert is selected as Save mode. If there are multiple records with the same composite key but different values in the batch, the system identifies the record with the latest value based on the Sort column. The Sort column defines the ordering of records, and the record with the highest value in the sort column is considered the latest.
Please Note:
Currently, the Sort column field is only available for the following drivers in the DB Writer: MSSQL, PostgreSQL, Oracle, Snowflake, and ClickHouse.
Database name: Enter the Database name.
Table name: Provide a table name where the data has to be written.
Enable SSL: Check this box to enable SSL for this components. Enable SSL feature in DB reader component will appear only for three(3) drivers: MongoDB, PostgreSQL and ClickHouse.
Certificate Folder: This option will appear when the Enable SSL field is checked-in. The user has to select the certificate folder from drop down which contains the files which has been uploaded to the admin settings. Please refer the below given images for the reference.
Schema File Name: Upload the Spark Schema in JSON format.
Query: In this field, we can write a DDL for creating the table in database where the in-event data has to be written. For example, please refer the below image:
Please Note:
In DB Writer component, the Save Mode for ClickHouse driver is as follows:
Append: It will create a table in ClickHouse database with a table engine Stripelog.
Upsert: It will create a table in ClickHouse database with a table engine ReplacingMergeTree.