All options contained on the Find Best Model dialogs are described below.  See the Fitting the Best Model chapter within the Data Science User Guide (Data Science - Help - User Guides) for a complete walk through of how to use this feature. 

Data Dialog

The Data dialog is where the data source is listed, the input and output variables are selected and the Success Class and Probability are set.  

Data Source:  Click the down arrow to select the workbook where the Find Best Model:  Classification method will be applied.

Worksheet:  Click the down arrow to select the worksheet where the Find Best Model:  Classification method will be applied.

Data range:  Select the range where the data appears on the selected worksheet.

#Columns:  (Read-only) The number of columns in the data range.

# Rows In:  Training Set:  (Read-only) The number of rows in the training partition.

# Rows In:  Validation Set:  (Read-only) The number of rows in the validation partition.

# Rows In:  Test Set:  (Read-only) The number of rows in the test partition.

First Row Contains Headers:  Select this option if the first row of the dataset contains column headings.

Variables in Input Data:  Variables contained in the dataset.
Selected Variables: Variables appearing under Selected Variables will be treated as continuous.    
Categorical Variables:  Variables appearing under Categorical Variables will be treated as categorical.
Output Variable:  Select the output variable, or the variable to be classified, here.

Classes: Number of Classes:  (Read-only) The number of classes that exist in the output variable.  
Binary Classification:  Success Class:  This option is selected by default.  Select the class to be considered a “success” or the significant class.  This option is enabled when the number of classes in the output variable is equal to 2.
Binary Classification:  Success Probability Cutoff: Enter a value between 0 and 1 here to denote the cutoff probability for success.  If the calculated probability for success for an observation is greater than or equal to this value, than a “success” (or a 1) will be predicted for that observation.  If the calculated probability for success for an observation is less than this value, then a “non-success” (or a 0) will be predicted for that observation.  The default value is 0.5.  This option is only enabled when the # of classes is equal to 2.

Parameters Dialog

Select Find Best Model parameters here such as the learners to be applied, and their options, and the metric and partition to be used to measure the performance of each learner.    

Preprocessing

Partition Data:  Analytic Solver Data Science includes the ability to partition a dataset from within a classification or prediction method by clicking Partition Data on the Parameters dialog. Analytic Solver Data Science will partition your dataset (according to the partition options you set) immediately before running the classification method.  If partitioning has already occurred on the dataset, this option will be disabled.  For more information on partitioning, please see the Data Science Partitioning chapter.  

Rescale Data:  Use Rescaling to normalize one or more features in your data during the data preprocessing stage. Analytic Solver Data Science provides the following methods for feature scaling:  Standardization, Normalization, Adjusted Normalization and Unit Norm.  For more information on this feature, see the Rescale Continuous Data section within the Transform Continuous Data chapter that occurs earlier in this guide.

Analytic Solver Data Mining:  Notes on Rescaling and Simulation functionality

Find Best Model Fitting Learners

Select All:  Click to select all available learners.

Clear All:  Click to deselect all previously selected learners.  

Learners:  Select to run the desired Learner.  Click the Parameters button to change a learner-specific option.  

Find Best Model - Scoring: 

The model with the "best" metric in the selected partition, will be used in scoring.  

Metric for Scoring:  Select the down arrow to select the metric to determine the best fit model.  See the tables below for a description of each metric.  

Based on partition:  Select the down arrow to select the metric in the desired partition to determine the best fit model.    

Find Best Model:  Classification Learner Parameters

Logistic Regression 

For more information on each parameter, see the Logistic Regression chapter within the Analytic Solver Reference Guide.

Fit Intercept: When this option is selected, the default setting, Analytic Solver Data Science will fit the Logistic Regression intercept.  If this option is not selected, Analytic Solver Data Science will force the intercept term to 0.

Iterations (Max):  Estimating the coefficients in the Logistic Regression algorithm requires an iterative non-linear maximization procedure.  You can specify a maximum number of iterations to prevent the program from getting lost in very lengthy iterative loops.  This value must be an integer greater than 0 or less than or equal to 100 (1< value <= 100).  

Nearest Neighbors Parameter

For more information on each parameter, see the K-Nearest Neighbors Classification chapter within the Analytic Solver Reference Guide.

# Neighbors (K):  Enter a value for the parameter K in the Nearest Neighbor algorithm.  

Classification Tree Parameter

For more information on each parameter, see the Classification Tree chapter within the Analytic Solver Reference Guide. 

Tree Growth Levels, Nodes, Splits, Tree Records in Terminal Nodes:  In the Tree Growth section, select Levels, Nodes, Splits, and Records in Terminal Nodes.  Values entered for these options limit tree growth, i.e. if 10 is entered for Levels, the tree will be limited to 10 levels.   

Prune:  If a validation partition exists, this option is enabled.  When this option is selected, Analytic Solver Data Science will prune the tree using the validation set. Pruning the tree using the validation set reduces the error from over-fitting the tree to the training data. Click Tree for Scoring to click the Tree type used for scoring:  Fully Grown, Best Pruned, Minimum Error, User Specified or Number of Decision Nodes. 

Neural Network Parameter 

For more information on each parameter, see the Neural Network Classification chapter within the Analytic Solver Reference Guide.

Architecture:  Click Add Layer to add a hidden layer.  To delete a layer, click Remove Layer.  Once the layer is added, enter the desired Neurons.  

Hidden Layer: Nodes in the hidden layer receive input from the input layer.  The output of the hidden nodes is a weighted sum of the input values.  This weighted sum is computed with weights that are initially set at random values.  As the network "learns", these weights are adjusted.  This weighted sum is used to compute the hidden node's output using a transfer function.  The default selection is Sigmoid. 

Output Layer:  As in the hidden layer output calculation (explained in the above paragraph), the output layer is also computed using the same transfer function as described for Activation:  Hidden Layer.  The default selection is Sigmoid.

Training Parameters:  Click Training Parameters to open the Training Parameters dialog to specify parameters related to the training of the Neural Network algorithm.  

Stopping Rules:  Click Stopping Rules to open the Stopping Rules dialog. Here users can specify a comprehensive set of rules for stopping the algorithm early plus cross-validation on the training error.  

Linear Discriminant Parameter

For more information on this method, see the Discriminant Analysis Classification chapter within the Analytic Solver Reference Guide.

Bagging Ensemble Method Parameter: For more information on each parameter, see the Ensemble Methods Classifification chapter within the Analytic Solver Reference Guide.

Number of Weak Learners:  This option controls the number of "weak" classification models that will be created.  The ensemble method will stop when the number or classification models created reaches the value set for this option.  The algorithm will then compute the weighted sum of votes for each class and assign the "winning" classification to each record.

Weak Learner

Under Ensemble:  Classification click the down arrow beneath Weak Leaner to select one of the six featured classifiers:  Discriminant Analysis, Logistic Regression, k-NN, Naïve Bayes, Neural Networks, or Decision Trees.  The command button to the right will be enabled.  Click this command button to control various option settings for the weak leaner.

Random Seed for Bootstrapping:  Enter an integer value to specify the seed for random resampling of the training data for each weak learner.  Setting the random number seed to a nonzero value (any number of your choice is OK) ensures that the same sequence of random numbers is used each time the dataset is chosen for the classifier.  The default value is "12345".  If left blank, the random number generator is initialized from the system clock, so the sequence of random numbers will be different in each calculation. If you need the results from successive runs of the algorithm to another to be strictly comparable, you should set the seed. To do this, type the desired number you want into the box. This option accepts both positive and negative integers with up to 9 digits.  

Boosting Ensemble Method Parameter:  For more information on each parameter, see the Ensemble Methods Classifification chapter within the Analytic Solver Reference Guide.

Number of Weak Learners, Weak Learner:  See descriptions above.

Adaboost Variant:  In AdaBoost.M1 (Freund), the constant is calculated as:

Analytic Solver Data Mining: Ensemble Methods Adaboost M1 formula

In AdaBoost.M1 (Breiman), the constant is calculated as:
Analytic Solver Data Mining: Ensemble Methods Adaboost Breiman formula

In SAMME, the constant is calculated as:

Analytic Solver Data Mining: Ensemble Methods Adaboost SAMME formula

Random Seed for Resampling:  Enter an integer value to specify the seed for random resampling of the training data for each weak learner.  Setting the random number seed to a nonzero value (any number of your choice is OK) ensures that the same sequence of random numbers is used each time the dataset is chosen for the classifier.  The default value is "12345".   

Random Trees Ensemble Method Parameter:  For more information on each parameter, see the Ensemble Methods Classifification chapter within the Analytic Solver Reference Guide.

Number of Weak Learners, Random Seed for Bootstrapping, Weak Learner:  See descriptions above.

Number of Randomly Selected Features:  The Random Trees ensemble method works by training multiple "weak" classification trees using a fixed number of randomly selected features then taking the mode of each class to create a "strong" classifier.  The option Number of randomly selected features controls the fixed number of randomly selected features in the algorithm.  The default setting is 3.  

Feature Selection Random Seed:  If an integer value appears for Feature Selection Random seed, Analytic Solver Data Science will use this value to set the feature selection random number seed.  Setting the random number seed to a nonzero value (any number of your choice is OK) ensures that the same sequence of random numbers is used each time the dataset is chosen for the classifier.  The default value is "12345".  If left blank, the random number generator is initialized from the system clock, so the sequence of random numbers will be different in each calculation. If you need the results from successive runs of the algorithm to another to be strictly comparable, you should set the seed. To do this, type the desired number you want into the box. This option accepts both positive and negative integers with up to 9 digits.    

Find Best Model Classification Scoring Statistic

Accuracy:  Accuracy refers to the ability of the classifier to predict a class label correctly.

Specificity:  Specificity is defined as the proportion of negative classifications that were actually negative.

Sensitivity:  Sensitivity is defined as the proportion of positive cases there were classified correctly as positive.

Precision:  Precision is defined as the proportion of positive results that are truly positive. 

F1:  Calculated as 0.743 -2  x (Precision * Sensitivity)/(Precision + Sensitivity):  The F-1 Score provides a statistic to balance between Precision and Sensitivity, especially if an uneven class distribution exists.  The closer the F-1 score is to 1 (the upper bound) the better the precision and recall.   

Find Best Model Prediction Scoring Statistic

R2: Coefficient of Determination: Examines how differences in one variable can be explained by the difference in a second variable, when predicting the outcome of a given event.
SSE Sum of Squared Error:   The sum of the squares of the differences between the actual and predicted values.
MSE  Mean Squared Error:   The average of the squared differences between the actual and predicted values.  
RMSE:  Root Mean Squared Error:  The standard deviation of the residuals.
MAD:  Mean Absolute Deviation:  Average distance between each data value and the sample mean; describes variation in a data set. 
 

Scoring Dialog

Select the type of output for the Find Best Model method and/or a scoring method (all optional).    

Scoring Data for Training, Validation and Test

Detailed Report:  Select this option to add a Detailed Report to the output.  This report shows the classification/prediction of records by row.  

Summary Report:  Select this option to add a Summary Report to the output. This report summarizes the Detailed Report.  For classification models, the Summary Report contains a confusion matrix, error report and various metrics:  accuracy, specificity, sensitivity, precision and F1.  For regression models, the Summary Report contains five metrics:  SSE, MSE, RMSE, MAD and R2. 

Lift Charts:  Select this option to add Lift charts, gain charts, and ROC curves to the output.  For a description of each, see the Find Best Model chapter within the Data Science User Guide.  

Frequency Chart:  New in V2023, when Frequency Chart is selected, a frequency chart will be displayed when the CFBM_TrainingScore, CFBM_ValidationScore,  PFBM_TrainingScore or PFBM_ValidationScore worksheets are selected.  This chart will display an interactive application similar to the Analyze Data feature, explained in detail in the Analyze Data chapter that appears earlier in this guide.  This chart will include frequency distributions of the actual and predicted responses individually, or side-by-side, depending on the user's preference, as well as basic and advanced statistics for variables, percentiles, six sigma indices.  

Score New Data*:  

In Worksheet:  Select to score new data in the worksheet immediately after the Find Best Model method is complete.   

In Database:  Select to score new data within a database immediately after the Find Best Model method is complete.       

*See the Scoring chapter in the Data Science User Guide for more information on scoring new data.  

Simulation Dialog

All supervised algorithms in V2023 include a new Simulation tab.  This tab uses the functionality from the Generate Data feature (described earlier in this guide) to generate synthetic data based on the training partition, and uses the fitted model to produce predictions for the synthetic data.  The resulting report, CFBM_Simulation (for classification) or PFBM_Simulation (for prediction) , will contain the synthetic data, the predicted values and the Excel-calculated Expression column, if present.  In addition, frequency charts containing the Predicted, Training, and Expression (if present) sources or a combination of any pair may be viewed, if the charts are of the same type.  

Evaluation:  Select Calculate Expression to amend an Expression column onto the frequency chart displayed on the CFBM_Simulation or PFBM_Simulation output tabs.  Expression can be any valid Excel formula that references a variable and the response as [@COLUMN_NAME].  Click the Expression Hints button for more information on entering an expression.