Champion/Challenger Model Comparison
This section describes the Champion/Challenger feature in ModelOp Center, and how to use it to help determine the best model method or approach to promote to production.
Table of Contents
Â
Introduction
Data Science requires experimentation using a variety of methods, approaches, frameworks, and potentially even coding languages. Champion/Challenger is an industry standard technique to compare the performance of competing model strategies, or the performance of a recently updated model against a previous version of the same model. The Champion/Challenger feature in ModelOp Center does a side-by side comparison of the performance of different models in order to determine which model is better to promote to production.
The following phases lay the groundwork for doing a Champion/Challenger comparison.
Define the evaluation metrics for the model. See Model Efficacy Metrics and Monitoring.
Build evaluation tests - this is done through an automated MLC Process. You can build an MLC Process to automatically execute metrics tests for a particular version of a model. See Model Lifecycle Manager Automation for more information.
Conduct a side-by-side comparison of the test results in the Champion/Challenger page of the Command Center.
Champion/Challenger Comparison
For the previously generated metrics results (see steps 1 & 2 above), use the following steps to activate the Champion/Challenger feature in the ModelOp Center to compare test results side-by-side.
In the Command Center, navigate to Models.
Choose the (two or more) models you would like to compare.
Â
Select the test results for each of the models
Â
You can view the metrics side-by-side to decide which model is performing better.
Â
Next Article: Model Efficacy Metrics and Monitoring >