Running and Monitoring Distributed ML with Ray and whylogs
- open source
- whylogs
- Integration
- AI Observability
Nov 23, 2021
Running and monitoring distributed ML systems can be challenging. The need to manage multiple servers, and the fact that those servers emit different logs, means that there can be a lot of overhead involved in scaling up a distributed ML system. Fortunately, Ray makes parallelizing Python processes easy, and the open source whylogs enables users to monitor ML models in production, even if those models are running in a distributed environment.
Ray is an exciting project that allows you to parallelize pretty much anything written in Python. One of the advantages of the whylogs architecture is that it operates on mergeable profiles that can be easily generated in distributed systems and collected into a single profile downstream for analysis, enabling monitoring for distributed systems. This post will review some options that Ray users have for integrating whylogs into their architectures as a monitoring solution.
Continue reading on the Anyscale Ray Blog
Other posts
Achieving Ethical AI with Model Performance Tracing and ML Explainability
Feb 2, 2023
- WhyLabs
- ML Monitoring
BigQuery Data Monitoring with WhyLabs
Jan 17, 2023
- WhyLabs
- BigQuery
- Integration
Robust & Responsible AI Newsletter - Issue #4
Dec 22, 2022
- WhyLabs
- Newsletter
WhyLabs Private Beta: Real-time Data Monitoring on Prem
Dec 21, 2022
- whylogs
- WhyLabs
Understanding Kolmogorov-Smirnov (KS) Tests for Data Drift on Profiled Data
Dec 21, 2022
- Data Science
- Machine Learning
- MLOps
Re-imagine Data Monitoring with whylogs and Apache Spark
Nov 23, 2022
- Integration
- Apache Spark
- whylogs
ML Monitoring in Under 5 Minutes
Nov 15, 2022
- ML Monitoring
- MLOps
- Data Logging
AIShield and WhyLabs: Threat Detection and Monitoring for AI
Nov 8, 2022
- Integration
- AI Observability
Large Scale Data Profiling with whylogs and Fugue on Spark, Ray or Dask
Oct 13, 2022
- whylogs
- Integration