Running and Monitoring Distributed ML with Ray and whylogs
- open source
- whylogs
- Integration
- AI Observability
Nov 23, 2021
Running and monitoring distributed ML systems can be challenging. The need to manage multiple servers, and the fact that those servers emit different logs, means that there can be a lot of overhead involved in scaling up a distributed ML system. Fortunately, Ray makes parallelizing Python processes easy, and the open source whylogs enables users to monitor ML models in production, even if those models are running in a distributed environment.
Ray is an exciting project that allows you to parallelize pretty much anything written in Python. One of the advantages of the whylogs architecture is that it operates on mergeable profiles that can be easily generated in distributed systems and collected into a single profile downstream for analysis, enabling monitoring for distributed systems. This post will review some options that Ray users have for integrating whylogs into their architectures as a monitoring solution.
Continue reading on the Anyscale Ray Blog
Other posts
Get Early Access to the First Purpose-Built Monitoring Solution for LLMs
May 11, 2023
- WhyLabs
- AI Observability
Mind Your Models: 5 Ways to Implement ML Monitoring in Production
May 17, 2023
- WhyLabs
- ML Monitoring
Simplifying ML Deployment: A Conversation with BentoML's Founder & CEO Chaoyu Yang
Apr 4, 2023
- WhyLabs
Data Drift vs. Concept Drift and Why Monitoring for Them is Important
Mar 28, 2023
- ML Monitoring
- MLOps
Robust & Responsible AI Newsletter - Issue #5
Mar 10, 2023
- WhyLabs
- Newsletter
Detecting Financial Fraud in Real-Time: A Guide to ML Monitoring
Mar 7, 2023
- ML Monitoring
- MLOps
How to Troubleshoot Embeddings Without Eye-balling t-SNE or UMAP Plots
Feb 23, 2023
- AI Observability
- whylogs
Achieving Ethical AI with Model Performance Tracing and ML Explainability
Feb 2, 2023
- WhyLabs
- ML Monitoring
Detecting and Fixing Data Drift in Computer Vision
Jan 26, 2023
- ML Monitoring
- whylogs
- WhyLabs