List of Terminology
ModelOp Center is the the leading enterprise-grade ModelOps software, helping large companies organize their enterprise AI effortsleading AI Governance software that helps enterprises safeguard all AI initiatives — including generative AI, Large Language Models (LLMs), in-house, third-party vendor, and embedded systems — without stifling innovation.
Term | Definition |
Abstraction | The art of replacing specific details about a model with generic ones. |
Artificial intelligence (AI) | A computer engineering discipline using mathematical or logic-based techniques to uncover, capture, or code knowledge and sophisticated techniques to arrive at inferences or predictions to solve business problems. |
Asset | Any individual component that is used and required during a model’s life cycle, such as model source code, schemas, dependencies, serialized objects, training artifacts, etc. |
Associated Model | A model that is linked to another model. An associated model may be a model implementation of a use case, or another model as part of an ensemble, or it may be a monitoring model. |
Data drift | The evolution of data over time, potentially introducing previously unseen variety and/or new categories of data that deviates from a baseline data, which is often the training data set. |
Deployment (aka Productionization or Operationalization) | The process of making a model available for use by the business. |
Enterprise AI | Enterprise AI encompasses the end-to-end business processes by which organizations incorporate AI into 24x7 business functions that are accountable, manageable and governable at enterprise scale. |
Governance | The processes and policies to manage and mitigate risk in the usage of analytical models across the enterprise. |
Inferences | Descriptions of the relationship between the independent variables and the outcomes in a data set. |
Interpretability | The ability of a human to retrace how a model generates its inferences or predictions. |
Lineage | All human and system interactions (code changes, testing, promotions, approvals, etc) that have occurred throughout a model’s entire life cycle. |
Machine learning (ML) | A subset of AI that uses algorithms to parse data, capture knowledge, and develop predictions or determinations. ML models are first trained on data sets; then, once in production, use a closed-loop process to “learn” from experience and improve the accuracy of their predictions or determinations. Some ML models are both complex and opaque, making it difficult to explain how the models arrive at specific predictions or determinations. |
Model | A set of code that represents functions, actions, and predictions important to the business. |
Model debt | The implied cost of undeployed models and/or models deployed without proper monitoring and governance. |
Model decay | A change in model performance that makes it less accurate in its inferences or predictions. |
Model life cycle (MLC) | A model's journey from use case inception, to development, independent validation, productionization, monitoring, iteration, and ultimately retirement. |
ModelOps | The key strategic capability for operationalizing enterprise AI. ModelOps encompasses the systems and processes that streamline the orchestration, monitoring, governance, and continuous improvement of data science models, but its fundamental role is to improve business results. |
Monitoring | The act of observing statistical, technical, and ethical aspects of a model's performance in operation. |
Predictions | Descriptions of the relationship between the independent variables and the outcomes in a data set which are used to estimate outcomes for new data points. |
Reference (Base) Model | Typically the primary (base) model that is providing business-decisioning. The Reference model may have 1 or more associated models that refer to the Reference model. |
Schema | The definition of a model’s expected data inputs or outputs expressed in a standard way. |
Shadow AI | The implied cost and risk of deployment of AI initiatives and models in production with no accountability to IT or governance organizations. It is expected to be the biggest risk to effective and ethical decision. |
Training | Tuning model parameters to optimize performance on a particular data set, with the typical output being a trained model artifact. |
(AI) Use Case | A business problem that can be addressed by AI, ML, or Advanced Analytics |
Model-Specific Metadata:
For a given Model registered in ModelOp Center, ModelOp Center stores a substantial amount of metadata about the model. The details can be found in this article.
Specifically within the ModelOp Center model metadata, there are a few elements that do not have industry standard terminology. Therefore, the purpose of these elements within the ModelOp Center platform have been defined here:
modelName | The user-supplied name of the model |
description | The user-supplied description of the model |
group | The LDAP group to which the model is registered |
modelOrganization | The name of the “organization” or “team” that owns the model. |
modelCategory | The technical type of model: BUSINESS_MODEL, MODEL_MONITORING_MODEL |
modelUseCategory | The “class” or “type” of model, such as “credit”, “fraud”, “marketing”, etc. |
modelMethodology | The methodology used for the model (e.g. OLS, WLS, Arima, CNN, etc) |
modelRisk | The risk classification of the model, such as “high risk”, “medium risk”, or “low risk”. |
Next Article: Getting Oriented with ModelOp Center's Command Center>