Creating a Batch Job
A Batch Job contains a reference to a Policy and the location of specific input data. It lets the Policy be applied to that data, publishing the output to a specified Protected Data Domain.
In addition to the Privitar user interface and Automation API, Jobs can also be imported from a previously exported file. See, Importing an Exported Job for more information.
To create a Batch Job to process specific input files, ensure that the location of the required data is known (paths to the required files, or a Hive database name), and that the Policy already exists and is not in Draft. Then take the following steps:
Select Jobs from the Navigation sidebar. The Jobs page is displayed.
Select the Batch Jobs tab.
Choose the Environment from the Environment list box.
The chosen Environment must contain the data to be processed by the Policy, and will also be the destination of the processed data.
Click on Create New Batch Job. If asked select either Hive or HDFS from the list box. The New Batch Job window is displayed.
Please keep in mind that different types of Batch Jobs will be available, depending on the Environment in which you are creating the Job:
AWS Glue Batch Jobs can only be created in AWS Glue Environments
HDFS and Hive Jobs can only be created in Environments with an Hadoop cluster configured.