Creating and managing jobs in an analytics project
You create jobs to run operational assets or files, such as Data Refinery flows, SPSS Modeler flows, Notebooks, and scripts, in a project.
When you create a job you define the properties for the job, such as the name, definition, environment runtime, schedule and notification specifications on different pages. You can run a job immediately or wait for the job to run at the next scheduled interval.
Each time a job is started, a job run is created, which you can monitor and use to compare with the job run history of previous runs. You can view detailed information about each job run, job state changes, and job failures in the job run log.
How a job is created depends on the operational asset or file and the type of project you are working in.
Operational asset/file | Type of project | Create job in tool | Create job from the Assets page | Create job automatically | More information |
---|---|---|---|---|---|
Data Refinery flow | All types | ✓ | ✓ | Creating jobs in Data Refinery | |
SPSS Modeler flow | All types | ✓ | ✓ | Creating jobs in SPSS Modeler | |
DataStage flow | Default Git integration and Deprecated Git integration | ||||
Empty and From file | ✓ | ✓ | ✓ | Creating jobs in DataStage | |
Notebook created in the Notebook editor | All types | ✓ | ✓ | Creating jobs in the Notebook editor | |
Files and scripts created in JupyterLab and RStudio | Default Git integration | ✓ | Creating code-based jobs | ||
Deprecated Git integration | ✓ | ✓ | Creating jobs in projects with deprecated Git integration | ||
Metadata import | Empty and From file | ✓ | Creating a metadata import job | ||
Masking flow | All types | ✓ | Creating a masking flow job from the Assets page |
Creating jobs automatically
Some jobs are created automatically at the time the asset is created in a project. These jobs are listed on the Jobs page of the project. You can view the job run details, change job settings, run the job manually, and delete the job from the Jobs page. Note that you can't edit the job settings for a metadata import job from the Job's page. You can only do this from the project's Assets page.
In Cloud Pak for Data, jobs are created for:
- DataStage flows. See Creating jobs in DataStage.
- Metadata import assets. See Creating a metadata import job.
Creating jobs for files in a project with deprecated Git integration
You can't create jobs directly in JupyterLab or RStudio. To create jobs for Notebooks or scripts that are created in JupyterLab or RStudio in a project with deprecated Git integration, you must push the files from the IDE to the Git repository associated with your project and then sync the repository files with the project. Any Notebooks, scripts, or RShiny apps that are pushed to a GIT repository with a size of zero bytes are considered invalid and are not synced with the project.
You can create jobs after you have synced your GIT files to create project assets:
- From the Notebook viewer for Notebooks. See Creating jobs in the Notebook viewer.
- From the project Assets page for Notebooks and scripts. See Creating jobs for the Assets page.
Creating jobs from the Assets page
You can create a job to run an asset from the project's Assets page.
To create jobs from the Assets page of a project:
- Select the asset from the section for your asset type and choose Create job from the menu icon with the lists of options () at the end of the table row.
- Define the job details by entering a name and a description (optional).
- If you can select Setting, specify the settings that you want for the job.
-
If you can select Configure, choose an environment runtime for the job. Depending on the asset type, you can optionally configure more settings, for example environment variables or script arguments.
To avoid accumulating too many finished job runs and job run artifacts, set how long to retain finished job runs and job run artifacts like logs or notebook results. You can either select the number of days to retain the job runs or the last number of job runs to keep.
-
On the Schedule page, you can optionally add a one-time or repeating schedule.
If you define a start day and time without selecting Repeat, the job will run exactly one time at the specified day and time. If you define a start date and time and you select Repeat, the job will run for the first time at the timestamp indicated in the Repeat section.
You can't change the time zone; the schedule uses your web browser's time zone setting. If you exclude certain weekdays, the job might not run as you would expect. The reason might be due to a discrepancy between the time zone of the user who creates the schedule, and the time zone of the compute node where the job runs.
- Optionally set to see notifications for the job. You can select the type of alerts to receive.
- Review the job settings. Then, create the job and run it immediately, or create the job and run it later.
Managing jobs
You can view all of the jobs that exist for your project from the project's Jobs page. With Admin or Editor role for the project, you can view and edit the job details. You can run jobs manually and you can delete jobs. With Viewer role for the project, you can only view the job details. You can't run or delete jobs with Viewer role.
To view the details of a specific job, click the job. From the job's details page, you can:
- View the runs for that job and the status of each run. If a run failed, you can select the run and view the log tail or download the entire log file to help you troubleshoot the run. A failed run might be related to a temporary connection or environment problem. Try running the job again. If the job still fails, you can send the log to Customer Support.
- Edit job settings by clicking Edit job, for example to change schedule settings or to pick another environment definition.
- Run the job manually by clicking from the job's action bar. You can start a scheduled job based on the schedule and on demand.
- Delete the job by clicking from the job's action bar.
Viewing and editing jobs in a tool
You can view and edit job settings associated with an asset directly in the following tools:
- Data Refinery
- Notebook editor or viewer
-
SPSS Modeler
To view and change job settings in these tools:
- In the tool, click the Jobs icon from the toolbar and select Save and view jobs. This action lists the jobs that exist for the asset.
- Select a job to see its details. You can change job settings by clicking Edit job.
-
DataStage
To view or edit runtime settings in a DataStage flow:
- Opening the flow and click the Settings icon, which looks like a gear.
- Click Run on the Settings page.
Learn more
- Creating jobs in Data Refinery
- Creating jobs in DataStage
- Creating jobs with Data Privacy
- Creating jobs in SPSS Modeler
- Creating code-based jobs
Parent topic: Working in projects