iemodel evaluate model

The iemodel evaluate command evaluates an Information Extraction Module model that has previously been trained.

Usage

iemodel evaluate model --n modelName --t testFileName --o outputFileName --c categoryCount --d trueOrfalse
RequiredArgumentDescription
Yes--n modelNameSpecifies the name and location of the model you want to evaluate. Directory paths you specify here are relative to the location where you are running the Administration Utility.
Yes--t testFileNameSpecifies the name and location of the test file used to evaluate the model.
No--o outputFileNameSpecifies the name and location of the output file that will store the evaluation results.
No--c categoryCountSpecifies the number of categories in the model; must be a numeric value.
Note: It is applicable only for Text Classification model.
No--d trueOrfalseSpecifies whether to display a table with entity wise detailed analysis; the value must be true or false, as below:
true
Detailed evaluation results are required.
false
Detailed evaluation results are not required.
The default is false.

The Model Evaluation Results table, and Confusion Matrix with its columns, as described below, display the counts per entity.

Note: If the command is run without this argument or with the argument value false, the Model Evaluation Results table and Confusion Matrix are not displayed. Only the Model Evaluation Statistics are displayed.

Output

Model Evaluation Statistics
Executing this command displays these evaluation statistics in a tabular format:
  • Precision: It is a measure of exactness. Precision defines the proportion of correctly identified tuples.
  • Recall: It is a measure of completeness of the results. Recall can be defined as a fraction of relevant instances that are retrieved.
  • F1 Measure: It is the measure of the accuracy of a test. The computation of F1 score takes into account both precision and recall of the test. It can be interpreted as the weighted average of the precision and recall, where F1 score reaches its best value at 1 and worst at 0.
  • Accuracy: It measures the degree of correctness of results. It defines the closeness of the measured value to the known value.
Model Evaluation Results
If the command is run with the argument --d true, the match counts of all the entities are displayed in a tabular format. The columns of the table are:
Input Count
The number of occurrences of the entity in the input data.
Mismatch Count
The number of times the entity match failed.
Match Count
The number of times the entity match succeeded.
Confusion Matrix
The Confusion Matrix (shown below) allows visualization of how an algorithm performs. It illustrates the performance of a classification model.
The column represents the instances in a predicted class while the row represents the instances in an actual class. Some of the terms associated with the confusion matrix are:
Actual
The number of occurrences of the entity in the actual class.
Predicted
The number of occurrences of the entity in the predicted class.
TP
True Positive: The number of entity occurrences predicted as positive and actually true as well.
TN
True Negative:The number of entity occurrences predicted as negative but actually true.
FP
False Positive: The number of entity occurrences predicted as positive but actually false.
FN
False Negative: The number of entity occurrences predicted as negative and actually false as well.

Example

This example:
  • Evaluates the model called "MyModel"
  • Uses a test file called "ModelTestFile" in the same location
  • Stores the output of the evaluation in a file called "MyModelTestOutput"
  • Specifies a category count of 4
  • Specifies that a detailed analysis of the evaluation is required
iemodel evaluate model --n MyModel --t C:\Spectrum\IEModels\ModelTestFile --o C:\Spectrum\IEModels\MyModelTestOutput --c 4 --d true