Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Read by thought-leaders and decision-makers around the world. Phone Number: +1-650-246-9381 Email: [email protected]
228 Park Avenue South New York, NY 10003 United States
Website: Publisher: https://towardsai.net/#publisher Diversity Policy: https://towardsai.net/about Ethics Policy: https://towardsai.net/about Masthead: https://towardsai.net/about
Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Founders: Roberto Iriondo, , Job Title: Co-founder and Advisor Works for: Towards AI, Inc. Follow Roberto: X, LinkedIn, GitHub, Google Scholar, Towards AI Profile, Medium, ML@CMU, FreeCodeCamp, Crunchbase, Bloomberg, Roberto Iriondo, Generative AI Lab, Generative AI Lab Denis Piffaretti, Job Title: Co-founder Works for: Towards AI, Inc. Louie Peters, Job Title: Co-founder Works for: Towards AI, Inc. Louis-François Bouchard, Job Title: Co-founder Works for: Towards AI, Inc. Cover:
Towards AI Cover
Logo:
Towards AI Logo
Areas Served: Worldwide Alternate Name: Towards AI, Inc. Alternate Name: Towards AI Co. Alternate Name: towards ai Alternate Name: towardsai Alternate Name: towards.ai Alternate Name: tai Alternate Name: toward ai Alternate Name: toward.ai Alternate Name: Towards AI, Inc. Alternate Name: towardsai.net Alternate Name: pub.towardsai.net
5 stars – based on 497 reviews

Frequently Used, Contextual References

TODO: Remember to copy unique IDs whenever it needs used. i.e., URL: 304b2e42315e

Resources

Take our 85+ lesson From Beginner to Advanced LLM Developer Certification: From choosing a project to deploying a working product this is the most comprehensive and practical LLM course out there!

Publication

Explainable Monitoring for Successful Impact with AI Deployments
Artificial Intelligence

Explainable Monitoring for Successful Impact with AI Deployments

Last Updated on January 29, 2021 by Editorial Team

Author(s): Anusha Sethuraman

Artificial Intelligence

Training and deploying ML models are relatively fast and cheap, but operationalizationβ€Šβ€”β€Šmaintaining, monitoring, and governing models over timeβ€Šβ€”β€Šis difficult and expensive. An Explainable ML Monitoring system extends traditional monitoring to provide deep model insights with actionable steps. As part of Fiddler’s 3rd annual Explainable AI Summit in October 2020, we brought together a panel of technical and product leaders to discuss operationalizing machine learning systems, and the key role that monitoring and explainability have to play in an organization’s AIΒ stack.

The shift to operationalization

As Natalia Burina (AI Product Leader, Facebook) noted, β€œThere’s been a shift towards operations with the rise of MLOps. A recent report gave the figure that 25% of the top 20 fastest-growing Github projects of Q2 2020 concerned ML infrastructure, tooling, and operations.” Abhishek Gupta (Engineering Lead, Facebook; ex-Head of Engineering, Hired, Inc.) predicts that over the next 2–5 years, we will see more and more tools that β€œSaaSify” aspects of ML operationalization.

These innovations are a response to more organizations tryingβ€Šβ€”β€Šand often strugglingβ€Šβ€”β€Što get their ML projects β€œout of the lab.” As Peter Skomoroch (Machine Learning Advisor) explained, due to the push around big data years ago, companies have already been investing in data infrastructure to help power analytics on their site. Now they’re trying to use this data for machine learning, but running into challenges. Traditional engineering processes are based around software that the team writes, tests, and then deploys to the site, and while it might be A/B tested for effectiveness, the software itself isn’t changing. However, the same can’t be said for machine learning. Monitoring and explainability are therefore key components of a successful AIΒ system.

Case in point:Β COVID-19

Kenny Daniel (Co-founder and CTO, Algorithmia) shared that, β€œIn the data science communities that I run in, there’s a picture of a timeseries, any time series, and it looks normal, and thenβ€Šβ€”β€ŠCOVID hit.” Moral of the story: If you don’t have a way of recognizing when the macro environment has shifted, you’re going to have problems. Airlines experienced this: at the start of the pandemic, their prices dropped dramatically, because the algorithms mistakenly thought that was the way to get people flyingΒ again.

Many companies had to rapidly retrain their models when COVID hit. Gupta described the situation at Hired as β€œsurreal” as they saw a sudden drop in hiring and surge in candidates, resulting in their models behaving in less-than-ideal ways. (Gupta has since moved on to an engineering lead role at Facebook.)

Monitoring and explainability

All the panelists agreed that monitoring is especially important for machine learning systems, and that most companies’ current tools are not sufficient. β€œYou have to assume that things will go wrong and your machine learning team will be under the gun to fix itβ€Šβ€”β€Šquickly,” said Skomoroch. β€œIf you have a model that you can’t interrogate, where you can’t determine why the accuracy is dropping, that’s a very stressful situation.”

This is even more important for high-stakes use cases where you’re dealing with fairness and vulnerable groups, Burina said, and added that β€œDebugging models is something that’s developing. We don’t have in the industry a very good way of doing this like we have in traditional software.” Skomoroch agreed: β€œThat’s why I think stuff like Fiddler is pretty exciting because a lot of this is done manually currently and ad hocβ€Šβ€”β€Šthere’s some notebooks flying around in emails. We really need to have benchmarks that we’re looking at consistently and continuously.”

Gupta said that in his opinion, β€œML monitoring and the ability to drill down and explain is inextricably linked.” When you have both of these things, you get faster detection and resolution of issues, and at the same time, ML engineers are able to develop a better intuition about which models and features need more work. Gupta explained that β€œFiddler’s tool and explainable monitoring has been a gamechanger and a step function improvement to how we monitor and react to challenges that we see in the marketplace.”

Monolithic solutions vs best-in-breed approach

The panelists unanimously agreed that the trend in the AI tooling stack is towards a more heterogeneous, β€œbest-in-breed” approach that combines open source, custom software, and various vendor solutionsβ€Šβ€”β€Šrather than one tool that does itΒ all.

According to Daniel, β€œThe more valuable and the more important the project is, the more you really want to have the best component for each bit.” In traditional software, that means combining different solutions for CI/CD, testing, monitoring, and observability, and the same logic applies for ML. After all, β€œYou can’t build the end-to-end solution and expect to succeed in an industry that’s evolving so quickly. You need to be able to switch out parts of the car while you’re driving it, because the things that were popular two years ago are notΒ today.”

Components for an ML tooling stack are increasingly out-sourced, not built in-house. The task for companies now is to pick high-quality tools that are specifically geared towards their domain and use case. β€œFor companies that are serious from the get-go,” said Burina, β€œthey should really consider best-of-breed solutions because that’s going to be their competitive advantage.”

Stakeholders forΒ AI

What are all the different personas that might care about a model and its outputs? Of course, data scientists and engineers are one group. Also, product managers care about the fit of a model with business strategy and purpose. Legal teams, regulators, and end-users will all potentially require access to this information, as well. And C-suite leadership often wants to know how models are doing at a highΒ level.

As Skomoroch put it, β€œThere’s a whole world of people who don’t really understand what you [data scientists] do day to day, and the whole team is kind of a black box to them. So there’s a side benefit to having something like Fiddler, having this observability, and monitoring happening, which is they have something to look at where they can see: what’s the progress? What’s happening with our machine learning models?” Gupta observed that having ML monitoring and explainability provides β€œa shared understanding of the levers and tradeoffsβ€Šβ€”β€Šand having a conversation at that level of abstraction goes a longΒ way.”

Algorithmic bias andΒ fairness

One of the most important use cases for explainable AI and monitoring, and one that stakeholders have a shared interest in, is preventing issues with bias and fairness. β€œUnwanted consequences can creep in at any part of the pipeline,” said Burina. β€œCompanies must think about it holistically, from design to development, and they really should have continuous monitoring for bias and fairness.”

Continuous monitoring can help teams β€œtrust but verify,” according to Gupta. With many people working asynchronously to improve the collective performance of an AI system, individual bias can over time creep in, even though no single person is controlling at the macro level how the system must behave. This is where explainable monitoring can reallyΒ help.

Who is ultimately responsible for making sure AI isn’t biased? After all, as Daniel noted, β€œJust because it’s in an AI black box doesn’t mean nobody’s responsible. Somebody still needs to be responsible.” In Skomoroch’s opinion, having a dedicated role like a chief data science officer or director focused on AI ethics can be a good choice. This person can make sure that nothing falls through the cracks when work moves from one team to the next. Burina also proposed a new industry-wide role of β€œmodel quality scientist: someone who would challenge the model, check it for robustness, including anything potentially adversarial….someone who would approve deployment, really making it a more rigorous process.”

At Fiddler we’ve heard about bias concerns from many of the customers we’ve engaged with. In response, we’ve been trying to put together a high-level framework that can showcase where there could be bias, and allow customers to take action from those insights: whether they might want to retrain a model, balance their data set, or continuously monitor over time and use those insights to adjust their applications.

Interested in listening to the full panel discussion? You can watch the live recording hereΒ . Panelists:

Peter Skomoroch, Machine LearningΒ Advisor

Abhishek Gupta, Engineering Lead, Facebook; ex-Head of Engineering, Hired

Natalia Burina, AI Product Leader,Β Facebook

Kenny Daniel, Co-Founder and CTO, Algorithmia

Moderated by Rob Harrell, Senior Product Manager,Β Fiddler

Originally published at https://blog.fiddler.ai on January 20,Β 2021.


Explainable Monitoring for Successful Impact with AI Deployments was originally published in Towards AI on Medium, where people are continuing the conversation by highlighting and responding to this story.

Published via Towards AI

Feedback ↓