Skip to main content

User Guide

Running a Batch Job

Once created, Batch Jobs may be run from the Jobs list.

  1. Select Jobs from the Navigation sidebar. The Jobs page is displayed.

  2. Select the Batch Jobs tab.

  3. Select the Environment where the Job definition is defined, from the Environment list box.

  4. Find the Job that you want to run. Use the Filter search box to filter the Jobs if required.

  5. Click on the Play icon in the end column alongside the Job that you want to run. The Run Job window is displayed.

  6. Select the PDD that the Job will output data to from the Add to an existing Protected Data Domain list box.

    Or, select Create new PDD to create a new PDD. Enter a name for the new PDD in the edit box along with all the other required details and click Save. (For more information, see Creating a Protected Data Domain.)

  7. Click on the checkbox alongside each Schema table that you want to include in the Job.

    Not all tables need be included. Only data corresponding to the selected tables will be processed.

  8. (Optional step) If there are any values that need to be specified for either Filters (for Hive Batch Jobs) or Path replacements (for HDFS Jobs), a window is displayed asking for the values to be specified. For example:

    UUID-40ce492a-7a17-8a59-093d-67ecdf0d5170.png

    For more information, see:

  9. (Optional step) If the Privitar application.properties file has enabled the option to provide advanced diagnostic output for Jobs, an Advanced tab is displayed. Select this tab to configure the output.

    For more information about the options available from the Advanced tab, see Setting Diagnostic Options for Jobs (HDFS & Hive Batch Jobs).

  10. Click on Run to start the Job.

Advanced Options

The settings described in this section are optional and do not need to be set in normal use.

Logging properties

Additional log4j Apache logging properties may be specified for the Job run.

Maximum Partition Size

Maximum Partition Size controls how much of an input file is processed in-memory at once by a Spark Executor. This setting can be used to control memory usage in the active Spark Executors. This setting is optional and should only be set if needed.

As an approximation, this value should be set according to the algorithm:

Maximum Partition Size < Executor Memory / (2 × Executor Cores)
AWS Glue Parameters

This option is only provided for Privitar AWS deployments.

The Executors option refers to the number of Spark executors that AWS Glue Batch Jobs will request to run a Batch job.

Note

Increasing this number will increase the cost of running this job in AWS. Contact your system administrator if you are in any doubt.