Running and Monitoring Distributed ML with Ray and whylogs
- open source
- whylogs
- Integration
- AI Observability
Nov 23, 2021
Running and monitoring distributed ML systems can be challenging. The need to manage multiple servers, and the fact that those servers emit different logs, means that there can be a lot of overhead involved in scaling up a distributed ML system. Fortunately, Ray makes parallelizing Python processes easy, and the open source whylogs enables users to monitor ML models in production, even if those models are running in a distributed environment.
Ray is an exciting project that allows you to parallelize pretty much anything written in Python. One of the advantages of the whylogs architecture is that it operates on mergeable profiles that can be easily generated in distributed systems and collected into a single profile downstream for analysis, enabling monitoring for distributed systems. This post will review some options that Ray users have for integrating whylogs into their architectures as a monitoring solution.
Continue reading on the Anyscale Ray Blog
Other posts
WhyLabs Announces SCA with AWS to Accelerate Responsible Generative AI Adoption
Nov 14, 2023
- WhyLabs
- SageMaker
Understanding and Mitigating LLM Hallucinations
Oct 18, 2023
- LLMs
- AI Observability
Understanding and Monitoring Embeddings in Amazon SageMaker with WhyLabs
Sep 11, 2023
- WhyLabs
- ML Monitoring
Data Drift Monitoring and Its Importance in MLOps
Aug 29, 2023
- MLOps
- Data Logging
Glassdoor Decreases Latency Overhead and Improves Data Monitoring with WhyLabs
Aug 17, 2023
- WhyLabs
- Machine Learning
Ensuring AI Success in Healthcare: The Vital Role of ML Monitoring
Aug 10, 2023
- ML Monitoring
WhyLabs Recognized by CB Insights GenAI 50 among the Most Innovative Generative AI Startups
Aug 8, 2023
- WhyLabs
Hugging Face and LangKit: Your Solution for LLM Observability
Jul 26, 2023
- LLMs
- WhyLabs
7 Ways to Monitor Large Language Model Behavior
Jul 20, 2023
- LLMs
- WhyLabs