Using the SDK
The SDK can be used to run Big Data Quality SDK jobs using any one of these two approaches:
- On a console, directly run the module-specific JAR files and pass the various XML -format
configuration properties files as arguments to the commands.
For MapReduce jobs run the
hadoop
command, while for Spark jobs run thesubmit-spark
command.For the steps, see Using Configuration Property Files.
- Create your own Java client project by importing the relevant Big Data Quality SDK
module JAR file, specify all required job configurations for your desired job within your client
project and run it.
For the steps, see Creating a Java Application.