# ML Model Monitoring & Alerts

## **MONITORING & ALERTS**

* [Monitor! Stop being a blind DS](https://towardsdatascience.com/monitor-stop-being-a-blind-data-scientist-ac915286075f)
* [Monitor your dependencies! Stop being a blind DS](https://towardsdatascience.com/monitor-your-dependencies-stop-being-a-blind-data-scientist-a3150bd64594)
* [Data science observability for executives](https://towardsdatascience.com/data-science-observability-for-executives-a054411faecc)
* [Production Machine Learning Monitoring: Outliers, Drift, Explainers & Statistical Performance](https://towardsdatascience.com/production-machine-learning-monitoring-outliers-drift-explainers-statistical-performance-d9b1d02ac158), [youtube](https://www.youtube.com/watch?v=QcevzK9ZuDg), uses [alibi-explain](https://docs.google.com/document/d/1dXELAcJn9KCPSRMDvZoumUyHx8K8Yn7wfFxesSpbNCM/edit#heading=h.xs1o8m3ro5iy) (see compendium) and Ali-detect (see compendium)
* [Mlflow, Hyperparameterhunter,hyperopt, concept drift, unit tests.](https://towardsdatascience.com/putting-ml-in-production-ii-logging-and-monitoring-algorithms-91f174044e4e)
* [meta anomaly over multiple models, aggregate.](https://www.anodot.com/blog/monitoring-machine-learning/)
* [Vidhya on monitoring data & models](https://www.analyticsvidhya.com/blog/2019/10/deployed-machine-learning-model-post-production-monitoring/)
* [Monitor ML features using Amazon SageMaker Feature Store and AWS Glue DataBrew](https://towardsdatascience.com/monitor-ml-features-using-amazon-sagemaker-feature-store-and-aws-glue-databrew-c530abcc479a)

### **Drift**

1. [Data & concept drifts](https://deepchecks.com/how-to-monitor-ml-models-in-production/), [2](https://www.explorium.ai/blog/understanding-and-handling-data-and-concept-drift/)
2. (good) [Inferring Concept Drift Without Labeled Data](https://concept-drift.fastforwardlabs.com/). also talks about stream-based drift by Cloudera - fast forward labs.
3. Arize.ai
   1. Data, concept, [feature drifts](https://towardsdatascience.com/using-statistical-distance-metrics-for-machine-learning-observability-4c874cded78) - various comparisons between train/prod/validation time windows, diff models, a/b testing etc.., and how to measure drifts
   2. [Model store, Feature store, evaluation store](https://towardsdatascience.com/the-only-3-ml-tools-you-need-1aa750778d33)
   3. [Monitor model performance in production](https://towardsdatascience.com/the-playbook-to-monitor-your-models-performance-in-production-ec06c1cc3245) - real- time, biased, delayed, and no ground truth.
   4. [use cases - i.e., how to use statistical differences/distances](https://towardsdatascience.com/using-statistical-distance-metrics-for-machine-learning-observability-4c874cded78)
4. [Some advice on medium](https://towardsdatascience.com/concept-drift-and-model-decay-in-machine-learning-a98a809ea8d4), relabel using latest model (can we even trust it?) retrain after.
5. [Adversarial Validation Approach to Concept Drift Problem in User Targeting Automation Systems at Uber](https://arxiv.org/abs/2004.03045) - Previous research on concept drift mostly proposed model retraining after observing performance decreases. However, this approach is suboptimal because the system fixes the problem only after suffering from poor performance on new data. Here, we introduce an adversarial validation approach to concept drift problems in user targeting automation systems. With our approach, the system detects concept drift in new data before making inference, trains a model, and produces predictions adapted to the new data.
6. Drift estimator between data sets using random forest, the formula is in the medium article above, code here at [mlBOX](https://github.com/AxeldeRomblay/MLBox/blob/811dbcb04fc7f5501e82f3e78aa6c119f426ee78/python-package/mlbox/preprocessing/drift/drift_estimator.py)
7. [Alibi-detect](https://docs.google.com/document/d/1dXELAcJn9KCPSRMDvZoumUyHx8K8Yn7wfFxesSpbNCM/edit#heading=h.y6mpsp4co5t9) - is an open-source Python library focused on outlier, adversarial, and drift detection, by Seldon.
8. [What is concept drift and why does it go undetected](https://censius.ai/blogs/what-is-concept-drift-and-why-does-it-go-undetected) Breaking down concept drit and explaining the best methods to avoid it
9. [\*\*How does data drift hamper AI performance \*\*](https://censius.ai/blogs/data-drift-barrier-to-ai-performance) Understand how data drift affect peak AI performance and how you can detect it

![Alibi Detection Drift Features](https://lh4.googleusercontent.com/sASV5qq3CTmv0gx6Tl3DiwACMnwsW9wj1yNHF5sFIFbQr4BFFgAVgfcWsnrHxNnQtQKa-b5-IdbC-OElnQIr117lxaH3TGCuz1CmpgU6mof3i9VkPR3LyzdD9S0ujTmWj7o88Iep)

### Tool Comparisons

1. [State of MLOps](https://www.stateofmlops.com) (by me), [medium](https://towardsdatascience.com/mlops-monitoring-market-review-66904f0863bb) article, open-source [AirTable](https://airtable.com/shr4rfiuOIVjMhvhL).
2. [MLOps.toys](https://mlops.toys/) - A curated list of MLOps projects by [Aporia](https://aporia.com)
3. [Neptune.AI](https://mlops.neptune.ai/) MLOPS tools landscape
4. [Twimlai](https://twimlai.com/solutions/) ML AI solutions
5. [Ambiata](https://www.ambiata.com/blog/2020-12-07-mlops-tools/) how to choose the best MLOps tools
6. [Lakefs](https://lakefs.io/the-state-of-data-engineering-in-2021/) on the state of data engineering - has monitoring and observability inside
7. [The NLP Pandec](https://github.com/ivan-bilan/The-NLP-Pandect#mlops-for-nlp) - MLOps for NLP
8. [ml-ops.org](https://ml-ops.org/)
9. [Awesome production ML](https://github.com/EthicalML/awesome-production-machine-learning/)

![Awesome production ML](https://3144294592-files.gitbook.io/~/files/v0/b/gitbook-legacy-files/o/assets%2F-Mgd48oS5_duTKOVE_Et%2F-MhH2fBLfTznt8DNZYts%2F-MhH2wJFojsekboCC0JN%2Fimage.png?alt=media\&token=7a900e49-eed2-4490-8538-d30d1622deb3)
