(new)AML Continuous Monitoring | Mapping Ground Truth Data for Model Performance Monitoring in Azure Machine Learning
This is the new post as my first post couldn't be viewed( shown 404 error after I clicked it)
Greetings,
I am reading the document regarding MLOps and would like to continuously monitor the ML model deployed inside Azure Machine Learning. The goal of monitoring performance metrics is to send the signal to Azure Event and function triggering the training pipeline to retrain the model if the performance drift is detected.
Here’s what I’ve accomplished so far:
- Completed data ingestion, ETL, model training, and registration using MLflow.
- Successfully deployed the model and tested the endpoint by executing the POST method via Postman.
- Enabled the data collector, and I can view the input and output production data in Blob Storage.
- Configured a training pipeline that can be triggered by an Azure Function executing the POST method.
My problem is the monitoring part and sending the model's performance drift signal to Azure Event Grid. Despite reading and following the steps in the document ( https://zcusa.951200.xyz/en-us/azure/machine-learning/how-to-monitor-model-performance?view=azureml-api-2&tabs=azure-cli ) I could see the correlationID in each row of output and understand that I need to map it with ground truth data to calculate the performance. (if I understand it correctly)
I’m struggling with the following:
- How do I retrieve the ground truth data and ensure it maps correctly to the corresponding
correlationID
in the model output? - As this is a proof-of-concept (POC) project using a mock dataset, what are the best practices or recommendations for setting up the monitoring and retraining scenario effectively?
I’d greatly appreciate any advice, recommendations, or insights on how to address this issue.
Thank you for taking the time to read my question and for your help in advance!
Best regards,
Boon Hawaree