In this example we show an end-to-end GPU-accelerated fraud detection example making use of tree-based models like XGBoost. In the first notebook 1_prep_rapids_train_xgb.ipynb we demonstrate GPU-accelerated tabular data preprocessing using RAPIDS and training of XGBoost model for fraud detection on the GPU in SageMaker. Then in second notebook 2_triton_xgb_fil_ensemble.ipynb we walkthrough the process of deploying data preprocessing + XGBoost model inference pipeline for high throughput, low-latency inference on Triton in SageMaker.
-
Launch SageMaker notebook instance with
g4dn.xlarge
instance.- In additional configuration specify on_start.sh as the lifecycle configuration start notebook script. This will create the RAPIDS kernel for us to use inside SageMaker notebook.
- 200GB storage should be fine.
- For git repository specify https://github.com/kshitizgupta21/fil_triton_sagemaker
-
Once JupyterLab is ready, launch the 1_prep_rapids_train_xgb.ipynb notebook with
rapids-2106
conda kernel and run through this notebook to do GPU-accelerated data preprocessing and XGBoost training on credit card transactions dataset for fraud detection use-case. -
Launch the 2_triton_xgb_fil_ensemble.ipynb notebook using
conda_python3
kernel and run through this notebook to deploy the ensemble data preprocessing + XGBoost model inference pipeline using the Triton's Python and FIL Backends on Triton SageMakerg4dn.xlarge
endpoint.