Using the SDK

The SDK can be used to run Big Data Quality SDK jobs using any one of these two approaches:

  1. On a console, directly run the module-specific JAR files and pass the various XML -format configuration properties files as arguments to the commands.

    For MapReduce jobs run the hadoop command, while for Spark jobs run the submit-spark command.

    For the steps, see Using Configuration Property Files.

  2. Create your own Java client project by importing the relevant Big Data Quality SDK module JAR file, specify all required job configurations for your desired job within your client project and run it.

    For the steps, see Creating a Java Application.