Driverless AI MOJO Scoring Pipeline - C++ Runtime with Python (Supports Shapley) and R Wrappers

The C++ Scoring Pipeline is provided as R and Python packages for the protobuf-based MOJO2 protocol. Use your preferred method once the MOJO Scoring Pipeline has been built.

Notes:

  • These scoring pipelines are currently not available for RuleFit models.

  • Unlike the Java Runtime, TensorFlow/Bert are supported by C++ Runtime MOJO.

  • You can have Driverless AI attempt to reduce the size of the MOJO scoring pipeline when the experiment is being built by enabling the Reduce MOJO Size expert setting also see.

  • Shapley contributions come with the downloaded experiment MOJO scoring pipeline. See C++ MOJO runtime Shapley values support for scoring example.

  • Shapley contributions for transformed features and original features are currently available for XGBoost (GBM, GLM, RF, DART), LightGBM, Zero-Inflated, Imbalanced and DecisionTree models (and their ensemble). For ensemble with ExtraTrees meta learner (ensemble_meta_learner=’extra_trees’) models, we suggest to use the MLI Python scoring package.

  • To disable the automatic creation of this scoring pipeline, set the Make MOJO Scoring Pipeline expert setting to Off while building an experiment.

  • The Download MOJO Scoring Pipeline button appears as Build MOJO Scoring Pipeline if the MOJO Scoring Pipeline is disabled.

Prerequisites

In order to perform C++ MOJO scoring on natural language processing (NLP) models, installing libopenblas-dev is required. To install libopenblas-dev, run the following command:

sudo apt install libopenblas-dev

Downloads

This section contains download links for the C++ MOJO runtime and its Python and R wrappers.

Python:

R:

Java:

  • The Java wrapper for the C++ MOJO runtime is now available as part of the existing Java runtime (mojo2-runtime.jar). For more information, see H2O MOJO Java Wrapper.

Note

The Python and R packages can also be downloaded from within the Driverless AI application by clicking Resources, and then clicking MOJO Py Runtime or MOJO R Runtime from the drop-down menu.

Examples

The following examples show how to use the R and Python APIs of the C++ MOJO runtime.

Prerequisites

  • Linux OS (x86 or PPC)

  • Driverless AI License (either file or environment variable)

  • Rcpp (>=1.0.0)

  • data.table

Running the MOJO2 R Runtime

 # Install the R MOJO runtime using one of the methods below

 # Install the R MOJO runtime on PPC Linux
 install.packages("./daimojo_2.8.1_ppc64le-redhat-linux-gnu.tar.gz")

 # Install the R MOJO runtime on x86 Linux
 install.packages("./daimojo_2.8.2_x86_64-redhat-linux-gnu.tar.gz")

 # Load the MOJO
 library(daimojo)
 m <- load.mojo("./mojo-pipeline/pipeline.mojo")

 # retrieve the creation time of the MOJO
 create.time(m)
 ## [1] "2019-11-18 22:00:24 UTC"

 # retrieve the UUID of the experiment
 uuid(m)
 ## [1] "65875c15-943a-4bc0-a162-b8984fe8e50d"

 # Load data and make predictions
 col_class <- setNames(feature.types(m), feature.names(m))  # column names and types

 library(data.table)
 d <- fread("./mojo-pipeline/example.csv", colClasses=col_class, header=TRUE, sep=",")

 predict(m, d)
 ##       label.B    label.M
 ## 1  0.08287659 0.91712341
 ## 2  0.77655075 0.22344925
 ## 3  0.58438434 0.41561566
 ## 4  0.10570505 0.89429495
 ## 5  0.01685609 0.98314391
 ## 6  0.23656610 0.76343390
 ## 7  0.17410333 0.82589667
 ## 8  0.10157948 0.89842052
 ## 9  0.13546191 0.86453809
 ## 10 0.94778244 0.05221756

Get the prediction interval from the MOJO

To get the prediction interval from the MOJO, run using the flag --with-prediction-interval

java -Xmx5g -Dai.h2o.mojos.runtime.license.file=license.file -jar mojo2-runtime.jar --with-prediction-interval pipeline.mojo example.csv

C++ MOJO runtime Shapley values support

The C++ MOJO runtime and its Python wrapper support Shapley contributions for transformed features and original features. The following example demonstrates how to retrieve Shapley contributions for transformed and original features when making predictions:

import datatable as dt
import daimojo
X = dt.Frame("example.jay")
m = daimojo.model("pipeline.mojo")
m.predict(X)  # Prediction call that returns regular predictions
m.predict(X, pred_contribs=True)  # Prediction call that returns Shapley contributions for transformed features
m.predict(X, pred_contribs=True, pred_contribs_original=True)  # Prediction call that returns Shapley contributions for original features

Note

  • Setting pred_contribs_original=True requires that pred_contribs is also set to True.

  • Presently, Shapley contributions for transformed features and original features are available for XGBoost (GBM, GLM, RF, DART), LightGBM, Zero-Inflated, Imbalanced and DecisionTree models (and their ensemble). For ensemble with ExtraTrees meta learner (ensemble_meta_learner=’extra_trees’) models we suggest to use the Python scoring packages.

  • In MOJOs, Shapley values for original features are approximated from the accompanying Shapley values for transformed features with the Naive Shapley (even split) method.

  • The Shapley fast approximation uses only one model (from the first fold) with no more than the first 50 trees. For details see fast_approx_num_trees and fast_approx_do_one_fold_one_model config.toml settings.