Data Pipeline
  • Data Pipeline
    • About Data Pipeline
    • Design Philosophy
    • Low Code Visual Authoring
    • Real-time and Batch Orchestration
    • Event based Process Orchestration
    • ML and Data Ops
    • Distributed Compute
    • Fault Tolerant and Auto-recovery
    • Extensibility via Custom Scripting
  • Getting Started
    • Homepage
      • List Pipelines
      • Creating a New Pipeline
        • Adding Components to Canvas
        • Connecting Components
          • Events [Kafka and Data Sync]
        • Memory and CPU Allocations
      • List Jobs
      • Create Job
        • Job Editor Page
        • Task Components
          • Readers
            • HDFS Reader
            • MongoDB Reader
            • DB Reader
            • S3 Reader
            • Azure Blob Reader
            • ES Reader
            • Sandbox Reader
          • Writers
            • HDFS Writer
            • Azure Writer
            • DB Writer
            • ES Writer
            • S3 Writer
            • Sandbox Writer
            • Mongodb Writer
            • Kafka Producer
          • Transformations
        • PySpark Job
        • Python Job
      • List Components
      • Delete Orphan Pods
      • Scheduler
      • Data Channel
      • Cluster Event
      • Trash
      • Settings
    • Pipeline Workflow Editor
      • Pipeline Toolbar
        • Pipeline Overview
        • Pipeline Testing
        • Search Component in Pipelines
        • Push Pipeline (to VCS/GIT)
        • Pull Pipeline
        • Full Screen
        • Log Panel
        • Event Panel
        • Activate/Deactivate Pipeline
        • Update Pipeline
        • Failure Analysis
        • Pipeline Monitoring
        • Delete Pipeline
      • Component Panel
      • Right-side Panel
    • Testing Suite
    • Activating Pipeline
    • Monitoring Pipeline
  • Components
    • Adding Components to Workflow
    • Component Architecture
    • Component Base Configuration
    • Resource Configuration
    • Intelligent Scaling
    • Connection Validation
    • Readers
      • S3 Reader
      • HDFS Reader
      • DB Reader
      • ES Reader
      • SFTP Stream Reader
      • SFTP Reader
      • Mongo DB Reader
        • MongoDB Reader Lite (PyMongo Reader)
        • MongoDB Reader
      • Azure Blob Reader
      • Azure Metadata Reader
      • ClickHouse Reader (Docker)
      • Sandbox Reader
      • Azure Blob Reader
    • Writers
      • S3 Writer
      • DB Writer
      • HDFS Writer
      • ES Writer
      • Video Writer
      • Azure Writer
      • ClickHouse Writer (Docker)
      • Sandbox Writer
      • MongoDB Writers
        • MongoDB Writer
        • MongoDB Writer Lite (PyMongo Writer)
    • Machine Learning
      • DSLab Runner
      • AutoML Runner
    • Consumers
      • SFTP Monitor
      • MQTT Consumer
      • Video Stream Consumer
      • Eventhub Subscriber
      • Twitter Scrapper
      • Mongo ChangeStream
      • Rabbit MQ Consumer
      • AWS SNS Monitor
      • Kafka Consumer
      • API Ingestion and Webhook Listener
    • Producers
      • WebSocket Producer
      • Eventhub Publisher
      • EventGrid Producer
      • RabbitMQ Producer
      • Kafka Producer
    • Transformations
      • SQL Component
      • Dateprep Script Runner
      • File Splitter
      • Rule Splitter
      • Stored Producer Runner
      • Flatten JSON
      • Email Component
      • Pandas Query Component
      • Enrichment Component
      • Mongo Aggregation
      • Data Loss Protection
      • Data Preparation (Docker)
      • Rest Api Component
      • Schema Validator
    • Scripting
      • Script Runner
      • Python Script
        • Keeping Different Versions of the Python Script in VCS
    • Scheduler
  • Custom Components
  • Advance Configuration & Monitoring
    • Configuration
      • Default Component Configuration
      • Logger
    • Data Channel
    • Cluster Events
    • System Component Status
  • Version Control
  • Use Cases
Powered by GitBook
On this page
  • Steps to Use Test Framework
  • Creating a Test Case
  • Running Test Cases
  • Test Cases
  • Reports
  • Logs
  • Advance Logs
  1. Getting Started
  2. Pipeline Workflow Editor
  3. Pipeline Toolbar

Pipeline Testing

PreviousPipeline OverviewNextSearch Component in Pipelines

Last updated 2 years ago

The Test suite module helps the developers to create a unit test for every component on the pipeline. We can upload input data and expected output data for every test case this will be then compared with the actual output generated by the component.

Check out the below-given walk through to understand the Pipeline Testing functionality.

Using Test Framework

The Test suite provides the following comparisons:

  • Compare no. of rows generated with the given output.

  • Compare no. of columns with the given output

  • Compare actual data with the given output

  • Schema validation with given schema.

Steps to Use Test Framework

  • Navigate to the Pipeline List page.

Or

  • Navigate to the Pipeline Workflow Editor page. Click the Test Pipeline icon on the Header panel.

  • The Test Framework page opens displaying details of the selected pipeline.

  • Search Component: A Search bar is provided to search all components associated with that pipeline. It helps to find a specific component by inserting the name in the Search Bar.

  • Component Panel: It displays all the components associated with that pipeline.

  • Create Test Case: By clicking on the Create Test Case icon.

Creating a Test Case

  • Click the Create Test Case icon.

  • The Test Case opens. Here, the user can create a new test case for the selected component from the component panel.

    • Test Name: Enter the test case name.

    • Test Description: Enter the description of the test case.

    • Input Data File: Upload the input data file. It is required for transformations and writer’s components.

    • Output Data File: Upload the expected output data file.

    • Schema File: Upload the schema of expected output data.

    • Input Data Type: It supports the JSON type.

    • Assertion Method: It supports the equals assertion method.

    • Sort: It orders the column values in Actual Output. A User can sort the string and integer column values.

    • Comparison Logic: It contains four type of comparison logic:

      • Compare Number of Columns: It compares input data columns with output data columns.

      • Compare Data: It compares input data with output data.

      • Compare Number of Rows: It compares input data rows with output data rows.

      • Check Data Matches Schema: It checks the uploaded schema with expected output data.​

Please Note:

  • Cross: This icon will close the create test case pop-up.

  • Cancel: This button will close the create test case pop-up.

  • Save: This button will save the test case and the test case will list in the Test Case list.

Running Test Cases

  • Run Test Cases: It will run single and multiple test cases for the selected component by clicking on the Run Test Cases button.

Test Cases

  • Test Cases: It displays the created test cases list for the selected component.

Test Case List:

It displays the following the details:

  • Checkbox: The User can select multiple or single test cases while running the test cases.

  • Test Name: It displays name of the test case.

  • Test Case Description: It displays description of the test case.

  • Created Date: It displays the created date of the test case.

  • Delete Icon: The User can remove the test case by clicking on the delete icon.

Test Case Tab

The user can update the test case details under this tab.

  • Test Case Name: The User can change the test case name.

  • Test Description: The User can change the description of the test case.

  • Output Schema File: A User can change the schema of expected output data by clicking on the upload icon, view the schema by clicking on the view icon, remove the schema by clicking on cross icon.

  • Sort Column Name: A User can change the sorting column name.

  • Update: By clicking on the Update button user can update the test case details.

  • Last Updated Date: It displays the last updated date of the test case.

Input Data Tab

The user can check the existing input data by clicking on this tab. It contains Shrink, Expand, Upload, Remove icon.

  • Shrink: It Shrink the input data rows.

  • Expand: It expand the input data rows.

  • Upload: User can upload input data file by clicking on upload button.

  • Remove: User can remove input data file by clicking on remove button.

Expected Output

The user can check existing expected output data. It contains Shrink, Expand, Upload, Remove icon.

  • Shrink: It Shrink the expected output data rows.

  • Expand: It expand the expected output data rows.

  • Upload: User can upload expected output data file by clicking on upload button.

  • Remove: User can remove expected output data file by clicking on remove button.

Report Tab

It displays the selected latest and previous test case reports.

Reports

Reports: It displays the latest report of each test cases for selected component. It displays each Test case name, Component Version, Comparison Logics and Run date.

Logs

It displays the log details of the component.

Advance Logs

It displays the component pods, if user run test case.

Click the Test Pipeline icon.

​

Pipeline Testing page
Failed Status Report
Failure Reason tab
Output Comparison