Jobs in a project

A job is a way of running operational assets, such as Data Refinery flows or Notebooks in a project in Watson Studio. You can also create jobs for a promoted asset in deployment spaces.

You can create jobs for the following assets:

  • Data Refinery flows in projects and spaces
  • DataStage flows in projects
  • Jupyter Notebooks in projects
  • Metadata import assets in projects

You can create a job in one of several ways:

From the Jobs tab of your project, you can:

  • See the list of the jobs in your project.
  • View the details of each job, including editing settings. You can also start a job manually from here. See Viewing jobs in a project.
  • Monitor job runs
  • Delete jobs

Creating jobs in Data Refinery

You can create a job to run a Data Refinery flow directly in Data Refinery.

  1. In Data Refinery, click the Jobs icon  the jobs icon from the Data Refinery toolbar and select Save and create a job.
  2. Define the job details by entering a name and a description (optional).
  3. On the Configure page, you can:

    • View which input data file is used and what the output file will be called.
    • Select an environment runtime for the job.
  4. On the Schedule page:

    • Optional: Add a one-time or repeating schedule. Ensure that you define meaningful date and time ranges. You can’t change the time zone; the schedule uses your web browser’s time zone setting. You can select to exclude days in schedules running by minutes, hours, and days. Excluding days is not available for schedules that run by the week, where you must select a specific day in the week, and by the month, where again you must select a day in the month.

    If you exclude certain weekdays, the job might not run as you would expect. The reason might be due to a discrepancy between the time zone of the user who creates the schedule, and the time zone of the compute node where the job runs.

  5. Review the job settings. Then, create the job and run it immediately, or create the job and run it later.

You can track the status of a job’s run and view the logs on the job’s run description page. See Viewing jobs in a project.

Creating jobs in DataStage

In DataStage, a job is a platform runtime asset that is related and associated with a DataStage flow. Multiple jobs can be associated with the same DataStage flow. Jobs can be scheduled or run as needed.

Jobs are automatically created for you when you edit or work with a DataStage flow on the canvas. When you click Run on the canvas, a job is created and started. Jobs maintain their past runs and logs, which you can view on the jobs dashboard.

Because any number of jobs can be associated with a single DataStage flow, a DataStage flow has a one-to-many relationship with the job asset type.

To manually create a job, complete the following steps:

  1. Open the project that contains the DataStage flow that you want to work with, then click the Assets tab.
  2. Go to the DataStage flows section and click the vertical ellipsis icon by the DataStage flow that you want to work with. Then, click Create job.
  3. Continue through the next configuration steps by following the job creation wizard. Then, click Create or Create and run.

Creating jobs in the Notebook editor or viewer

You can create a job to run a Notebook directly from the Notebook editor or the Notebook viewer by clicking the Jobs icon the jobs icon from the Notebook’s menu bar. See Create a Notebook job.

Creating a metadata import job

A metadata import job is created at the time you create a metadata import asset. See Creating a metadata import asset and importing metadata.

Creating jobs from the Assets page

You can create jobs for Data Refinery flows and Jupyter notebooks from the Assets page of a project.

  1. Select the asset from the section for your asset type and choose Create job from the ACTIONS menu.
  2. Define the job details by entering a name and a description (optional).
  3. On the Configure page, select an environment runtime for the job. Depending on the asset, you can optionally configure more settings, for example environment variables or script arguments.
  4. On the Schedule page, you can optionally add a one-time or repeating schedule.
  5. Review the job settings. Then, create the job and run it immediately, or create the job and run it later.

Viewing jobs associated with an asset

If other jobs exist that are associated with your asset, you can view these jobs without running the job.

  1. Click the Jobs icon the jobs icon from the toolbar in Data Refinery, or in the Notebook editor or viewer, and then select Save and view jobs. This action lists the jobs that exist for the asset.
  2. Click View to see the details of a selected job.

Alternatively, you can view the details of the jobs that are associated with your asset by clicking the jobs from the Jobs tab of your project.

Viewing jobs in a project

You can view all of the jobs that exist for your project from the project’s Jobs page. You must have the Admin or Editor role to view job details and complete actions. With the Viewer role for the project, you can view only details.

To view the details of a specific job, click the job. From the job’s details page, you can:

  • View the runs for that job and the status of each run. If a run failed, you can select the run and view the log tail or download the entire log file to help you troubleshoot the run. A failed run might be related to a temporary connection or environment problem. Try running the job again. If the job still fails, you can send the log to Customer Support.
  • Edit job settings like the schedule settings or pick another environment definition.
  • Run the job manually by clicking the run icon from the jobs action bar. You must deselect the schedule to run the job manually.
  • Delete a job.