Fraud Detection With Categorical XGBoost#

This example notebook shows how to train and deploy an XGBoost model with categorical features in Triton using the FIL backend. We begin by training two separate models on a fraud detection dataset with categorical variables: one small model designed to maximize runtime performance and one larger model designed to maximize accurate and precise detection of fraud. We then deploy both models on CPU and GPU and compare their performance using Triton’s perf_analyzer. Based on these results, we see that GPU deployment opens up the possibility of deploying a much larger and more accurate fraud model with higher throughput while also keeping to a tight latency budget.

Running the notebook#

In order to launch the Triton server, you will need Docker installed on your system. The rest of the notebook also requires a few Python dependencies. To easily install these additional dependencies, you may make use of the provided conda environment file as follows:

conda env create -f environment.yml

You may then activate the conda environment and run the notebook as usual:

conda activate triton_example
jupyter notebook

The Jupyter interface should now be accessible from a browser, and you can follow the instructions within the notebook itself from there.