Random Forest Classifier Training Jobs
Train a random forest classifier for text classification.
This job is deprecated in 5.3.x. The Classification job, introduced in Fusion 5.2.0, provides more options and better logging. |
Legacy Product
Train a random forest classifier for text classification.
This job is deprecated in 5.3.x. The Classification job, introduced in Fusion 5.2.0, provides more options and better logging. |
Use this job when you have training data and you want to train a random forest model to classify text into groups.
LuceneTextAnalyzer schema for tokenization (JSON-encoded)
Default: { "analyzers": [{ "name": "StdTokLowerStop","charFilters": [ { "type": "htmlstrip" } ],"tokenizer": { "type": "standard" },"filters": [{ "type": "lowercase" },{ "type": "KStem" },{ "type": "length", "min": "2", "max": "32767" },{ "type": "fusionstop", "ignoreCase": "true", "format": "snowball", "words": "org/apache/lucene/analysis/snowball/english_stop.txt" }] }],"fields": [{ "regex": ".+", "analyzer": "StdTokLowerStop" } ]}
Ensure that all classes of training data have the same size
Default: true
Spark-compatible format which training data comes in (like 'solr', 'hdfs', 'file', 'parquet' etc)
>= 1 characters
Default: solr
Optimize hyperparameter search over one of [binary, multiclass, regression] metrics, or 'none'
Default: none
Allowed values: binarymulticlassregressionnone
Solr field containing text training data. Data from multiple fields with different weights can be combined by specifying them as field1:weight1,field2:weight2 etc.
>= 1 characters
Perform grid search to optimize hyperparameters
Default: false
The ID for this Spark job. Used in the API to reference this job. Allowed characters: a-z, A-Z, dash (-) and underscore (_). Maximum length: 63 characters.
<= 63 characters
Match pattern: [a-zA-Z][_\-a-zA-Z0-9]*[a-zA-Z0-9]?
Create a label class 'Other' which contains all examples not in a class large enough to train on
Default: true
Max number of bins for discretizing continuous features. Must be >=2 and >= number of categories for any categorical feature.
<= 128
exclusiveMinimum: false
exclusiveMaximum: false
Default: 32
To be kept, terms must occur in no more than this number of documents (if > 1.0), or no more than this fraction of documents (if <= 1.0)
Default: 1
Maximum depth of the tree (>= 0). E.g., depth 0 means 1 leaf node; depth 1 means 1 internal node + 2 leaf nodes.
>= 1
<= 20
exclusiveMinimum: false
exclusiveMaximum: false
Default: 5
To be kept, terms must occur in at least this number of documents (if > 1.0), or at least this fraction of documents (if <= 1.0)
Default: 0
Minimum number of Spark partitions for training job.
>= 1
exclusiveMinimum: false
Default: 200
Ensure that all classes of training data have at least this many examples
>= 1
exclusiveMinimum: false
Default: 100
Identifier for the model to be trained; uses the supplied Spark Job ID if not provided.
>= 1 characters
p-norm to normalize vectors with (choose -1 to turn normalization off)
Default: 2
Allowed values: -1012
Number of trees to train (>= 1)
>= 1
<= 1000
exclusiveMinimum: false
exclusiveMaximum: false
Default: 20
Label class name for the catch-all 'Other' class
>= 1 characters
Default: Other
Solr Collection to store model-labeled data to
If a model exists in the model store, overwrite when this job runs
Default: true
Solr field which will contain labels when classifier is applied to documents
Default: labelPredictedByFusionModel
For any deterministic pseudorandom number generation
Default: 1234
Options used when reading input from Solr or other sources.
object attributes:{key
required : {
display name: Parameter Name
type: string
}value
: {
display name: Parameter Value
type: string
}}
Solr fields to load (comma-delimited). Leave empty to allow the job to select the required fields to load at runtime.
Spark configuration settings.
object attributes:{key
required : {
display name: Parameter Name
type: string
}value
: {
display name: Parameter Value
type: string
}}
Solr Collection containing labeled training data
>= 1 characters
Solr query to use when loading training data if using Solr, Spark SQL expression for all other data sources
Default: *:*
Additional spark dataframe loading configuration options
Fraction of the training data to use
<= 1
exclusiveMaximum: false
Default: 1
Solr field containing labels for training instances (should be single-valued strings)
Default: random_forests_classifier
Allowed values: random_forests_classifier
Word-vector dimensionality to represent text (choose > 0 to use)
exclusiveMinimum: false
Default: 0
Maximum number of iterations of the word2vec training
Default: 1
Sets the maximum length (in words) of each sentence in the input data. Any sentence longer than this threshold will be divided into chunks of up to `maxSentenceLength` size.
>= 3
exclusiveMinimum: false
Default: 1000
Training parameter for word2vec convergence (change at your own peril)
>= 0.005
exclusiveMinimum: false
Default: 0.025
The window size (context words from [-window, window]) for word2vec
>= 3
exclusiveMinimum: false
Default: 5
Weight vector components based on inverse document frequency
Default: true
Options used when writing output to Solr.
object attributes:{key
required : {
display name: Parameter Name
type: string
}value
: {
display name: Parameter Value
type: string
}}