Python Job
Last updated
Last updated
This feature allows users to write their own Python script and run their script in the Jobs section of Data Pipeline module.
Before creating the Python Job, the user has to create a project in the DS Lab module under Python Environment. Please refer the below image for reference:
After creating the project, the user needs to activate it and create a Notebook where they can write their own Python script. Once the script is written, the user must save it and export it to be able to use it in Python Jobs.
Click on the Data Pipeline module from the homepage.
Click on the Create Job icon on the list pipeline page.
The New Job dialog box appears redirecting the user to create a new Job.
Enter a name for the new Job.
Describe the Job (Optional).
Job Baseinfo: Select Python Job from the drop-down.
Trigger By: There are 2 options for triggering a job on success or failure of a job: Success Job: On successful execution of the selected job the current job will be triggered. Failure Job: On failure of the selected job the current job will be triggered.
Is Scheduled?
A job can be scheduled for a particular timestamp. Every time at the same timestamp the job will be triggered.
Job must be scheduled according to UTC.
Docker Configuration
Select a resource allocation option using the radio button. The given choices are:
Low
Medium
High
Provide the resources required to run the python Job in the limit and Request section. 1. Limit: Enter max CPU and Memory required for the Python Job. 2. Request: Enter the CPU and Memory required for the job at the start. 3. Instances: Enter the number of instances for the Python Job.
Click the Save option to save the Python Job.
The Python Job gets saved, and it will redirect the user to the Job Editor workspace.
Check out the below given demonstration configure a Python Job.
Once the Python Job is created, follow the below given steps to configure the Meta Information tab of the Python Job.
Project Name: Select the same Project using the drop-down menu where the Notebook has been created.
Script Name: This field will list the exported Notebook names which are exported from the Data Science Lab module to Data Pipeline.
External Library: If any external libraries are used in the script the user can mention it here. The user can mention multiple libraries by giving comma (,) in between the names.
Start Function: Select the function name in which the script has been written.
Script: The Exported script appears under this space.
Input Data: If any parameter has been given in the function, then the name of the parameter is provided as Key, and value of the parameters has to be provided as value in this field.