Classification and Regression with the BigML Dashboard

7.3 Creating Evaluations

The process to create an evaluation is different if you want to create a single evaluation or a cross-validation evaluation:

  • To create a single evaluation, you need two resources: a testing dataset (different from the one used for training) and a model, an ensemble, a logistic regression, deepnet, or fusion. All three processes follow a similar logic. You can find a separate explanation of each one in the following subsections.

  • To create a cross-validation evaluation, you just need a dataset. BigML allows you to create cross-validation for models, ensembles, logistic regressions, deepnets and fusions. This process is explained in subsection 7.3.6

7.3.1 Model Evaluations

To evaluate a model, you can use any of the following options from the BigML Dashboard:

  • Click Evaluate a model in the 1-click action menu from the evaluation list view. (Figure 7.31 .)

    \includegraphics[]{images/evaluations/evaluate-model-listing-view}
    Figure 7.31 Evaluate model from evaluation list view

    This option redirects you to the New Evaluation view where you need to select a model and a testing dataset. (See Figure 7.32 .) From this view you can also select an ensemble by clicking the ensemble icon above the model selector.

    \includegraphics[]{images/evaluations/evaluation-search-model}
    Figure 7.32 Select model and dataset
  • Click Evaluate in the 1-click action menu from the model view. (See Figure 7.33 .)

    \includegraphics[]{images/evaluations/evaluate-model}
    Figure 7.33 Evaluate model from 1-click action menu

    Alternatively, click Evaluate in the pop up menu from the model list view (see Figure 7.34 ).

    \includegraphics[]{images/evaluations/evaluate-model-listing-models}
    Figure 7.34 Evaluate model from pop up menu

    By using any of these options, you will be redirected to the New Evaluation view where the model will be pre-filled in the selector and you only have to choose the testing dataset. If you previously split your original dataset into two subsets (one for training and another for testing) using the 1-click menu option from your dataset view, BigML will automatically select the corresponding testing dataset. Finally, click the Evaluate green button to perform the evaluation. (See Figure 7.35 .)

    \includegraphics[]{images/evaluations/evaluate-from-model-menu}
    Figure 7.35 Evaluation with pre-filled model and dataset information

7.3.2 Ensemble Evaluations

To evaluate an ensemble you can use the following options from the BigML Dashboard:

  • Click Evaluate an ensemble in the 1-click action menu from the evaluation list view (Figure 7.36 ).

    \includegraphics[]{images/evaluations/evaluate-ensemble-listing-view}
    Figure 7.36 Evaluate ensemble from evaluation list view

    This option takes you to the New Evaluation view where you need to select an ensemble and a testing dataset. (See Figure 7.37 .) From this view, you can also select a model by clicking the model icon above the ensemble selector.

    \includegraphics[]{images/evaluations/evaluation-search-ensemble}
    Figure 7.37 Select ensemble and dataset
  • Click Evaluate in the 1-click action menu from the ensemble view- (Figure 7.38 .)

    \includegraphics[]{images/evaluations/evaluate-ensemble}
    Figure 7.38 Evaluate ensemble from 1-click action menu

    Alternatively, click Evaluate in the pop up menu from the ensembles list view (Figure 7.39 ).

    \includegraphics[]{images/evaluations/evaluate-ensemble-listing-ensembles}
    Figure 7.39 Evaluate ensemble from pop up menu

    By using any of these two options, you will be redirected to the New Evaluation view where the ensemble will be pre-filled and you only have to choose the testing dataset. If you previously split your original dataset into two subsets (one for training and another for testing) using the 1-click menu option from your dataset view, BigML will automatically select the corresponding testing dataset. Finally, click the Evaluate green button to perform the evaluation. (See Figure 7.40 .)

    \includegraphics[]{images/evaluations/evaluate-from-ensemble-menu}
    Figure 7.40 Evaluation with pre-filled ensemble and dataset information

7.3.3 Logistic Regression Evaluations

To evaluate a logistic regression, you can use these options from the BigML Dashboard:

  • Click Evaluate a logistic regression from the 1-click action menu from the evaluation list view. (See Figure 7.41 .)

    \includegraphics[]{images/evaluations/lr-eval-listing-view}
    Figure 7.41 Evaluate logistic regression from evaluation list view

    This option redirects you to the New Evaluation view where you need to select a logistic regression and a testing dataset. (See Figure 7.42 .)

    \includegraphics[]{images/evaluations/eval-lr-search-dataset}
    Figure 7.42 Select logistic regression and dataset
  • Click Evaluate from the logistic regression 1-click action menu. (Figure 7.43 .)

    \includegraphics[]{images/evaluations/evaluate-lr-one-click}
    Figure 7.43 Evaluate logistic regression from 1-click action menu

    Alternatively, click Evaluate in the pop up menu from the logistic regression list view. (Figure 7.44 .)

    \includegraphics[]{images/evaluations/evaluate-lr-listing-view}
    Figure 7.44 Evaluate logistic regression from pop up menu

    By using any of these options, you will be redirected to the New Evaluation view where the logistic regression will be pre-filled in the selector and you only need to choose the testing dataset. If you previously split your original dataset into two subsets (one for training and another for testing) using the 1-click menu option from your dataset view, BigML will automatically select the corresponding testing dataset. Finally, click the Evaluate green button to perform the evaluation. (See Figure 7.45 .)

    \includegraphics[]{images/evaluations/eval-lr-testing-dataset}
    Figure 7.45 Evaluation with pre-filled logistic regression and dataset information

7.3.4 Deepnet Evaluations

To evaluate a deepnet, you can use these options from the BigML Dashboard:

  • Click Evaluate a deepnet from the 1-click action menu from the evaluation list view. (See Figure 7.46 .)

    \includegraphics[]{images/evaluations/deepnet-eval-listing-view}
    Figure 7.46 Evaluate deepnet from evaluation list view

    This option redirects you to the New Evaluation view where you need to select a deepnet and a testing dataset. (See Figure 7.47 .)

    \includegraphics[]{images/evaluations/eval-deepnet-search-dataset}
    Figure 7.47 Select deepnet and dataset
  • Click Evaluate from the deepnet 1-click action menu. (See Figure 7.48 .)

    \includegraphics[]{images/evaluations/evaluate-deepnet-one-click}
    Figure 7.48 Evaluate deepnet from 1-click action menu

    Alternatively, click Evaluate in the pop up menu from the deepnet list view. (Figure 7.49 .)

    \includegraphics[]{images/evaluations/evaluate-deepnet-listing-view}
    Figure 7.49 Evaluate deepnet from pop up menu

    By using any of these options, you will be redirected to the New Evaluation view where the deepnet will be pre-filled in the selector and you only need to choose the testing dataset. If you previously split your original dataset into two subsets (one for training and another for testing) using the 1-click menu option from your dataset view, BigML will automatically select the corresponding testing dataset. Finally, click the Evaluate green button to perform the evaluation. (See Figure 7.50 .)

    \includegraphics[]{images/evaluations/eval-deepnet-testing-dataset}
    Figure 7.50 Evaluation with pre-filled deepnet and dataset information

7.3.5 Fusion Evaluations

To evaluate a fusion, you can use these options from the BigML Dashboard:

  • Click Evaluate a fusion from the 1-click action menu from the evaluation list view. (See Figure 7.51 .)

    \includegraphics[]{images/evaluations/fusion-eval-listing-view}
    Figure 7.51 Evaluate fusion from evaluation list view

    This option redirects you to the New Evaluation view where you need to select a fusion and a testing dataset. (See Figure 7.52 .)

    \includegraphics[]{images/evaluations/eval-fusion-search-dataset}
    Figure 7.52 Select fusion and dataset
  • Click Evaluate from the fusion 1-click action menu. (See Figure 7.53 .)

    \includegraphics[]{images/evaluations/evaluate-fusion-one-click}
    Figure 7.53 Evaluate fusion from 1-click action menu

    Alternatively, click Evaluate in the pop up menu from the fusion list view. (Figure 7.54 .)

    \includegraphics[]{images/evaluations/evaluate-fusion-listing-view}
    Figure 7.54 Evaluate fusion from pop up menu

    By using any of these options, you will be redirected to the New Evaluation view where the fusion will be pre-filled in the selector and you only need to choose the testing dataset. Finally, click the Evaluate green button to perform the evaluation. (See Figure 7.55 .)

    \includegraphics[]{images/evaluations/eval-fusion-testing-dataset}
    Figure 7.55 Evaluate a fusion using a testing dataset

7.3.6 Cross-Validation Evaluations

In BigML, you can use k-fold cross-validation to evaluate your models, ensembles, logistic regressions, and deepnets. Cross-validation evaluations are implementing in BigML as a WhizzML script and they can be found in BigML Gallery:

  1. Go to the scripts Gallery where you will find five different scripts to perform cross-validation:

  2. By clicking the script preview you can inspect script’s details such as the source-code, the script input and the output. (See Figure 7.57 .) You can find additional documentation about WhizzML scripts here.

    \includegraphics[]{images/evaluations/cross-val-script}
    Figure 7.61 Cross-validation script view
  3. Clone your preferred script for FREE. You can clone it from the script preview by clicking the FREE button. (See Figure 7.62 .)

    \includegraphics[width=7cm]{images/evaluations/cross-val-clone}
    Figure 7.62 Clone script from preview

    Alternatively, you can clone it from the script view by clicking the FREE or Buy buttons. (See Figure 7.63 .)

    \includegraphics[]{images/evaluations/cross-val-clone2}
    Figure 7.63 Clone script from script view

    A modal window will appear asking you for confirmation. (See Figure 7.64 .)

    \includegraphics[width=10cm]{images/evaluations/cross-val-clone3}
    Figure 7.64 Confirmation message to clone script
  4. Once you clone the script, you will be redirected to the Execution view to set your inputs. You need to select a dataset and optionally, you can configure the rest of the inputs. If you do not configure them, they will take the default values. You can find an explanation of all your inputs in subsection 7.4.5 . (See Figure 7.65 .)

    \includegraphics[]{images/evaluations/cross-val-execution}
    Figure 7.65 Configure cross-validation inputs
  5. Once you have selected the dataset, click Execute . (See Figure 7.66 .)

    \includegraphics[]{images/evaluations/cross-val-execute}
    Figure 7.66 Execute cross-validation script
  6. Once you execute the script, you can check the progress of your script in the execution view where you will find the elapsed time, the total resources generated and the script log messages. (See Figure 7.67 .)

    \includegraphics[]{images/evaluations/cross-val-logs}
    Figure 7.67 Cross-validation execution progress
  7. Finally, cross-validation yields \(k\) different models and \(k\) different evaluations. The results of the single evaluations are averaged to obtain the final model performance measures. Access the final cross-validation evaluation containing the averaged measures by clicking on the evaluation ID link in the Outputs section. Read more about cross-validation measures in subsection 7.2.3 . The \(k\) intermediary resources can be found in the same view under the Resources panel. (See Figure 7.68 .)

    \includegraphics[]{images/evaluations/cross-validation-output}
    Figure 7.68 Cross-validation output and resources

You can perform again any new cross-validation by clicking on your cloned script listed under the scripts tab in the BigML Dashboard. (See Figure 7.69 .)

\includegraphics[]{images/evaluations/cross-val-script-listing}
Figure 7.69 Cross-validation script in scripts list view