Responsible AI image dashboard in Azure Machine Learning studio (preview)
The Responsible AI image dashboards are linked to your registered computer vision models in Azure Machine Learning. While steps to view and configure the Responsible AI dashboard is similar across scenarios, some features are unique to image scenarios.
Important
Responsible AI image dashboard is currently in public preview. This preview is provided without a service-level agreement, and are not recommended for production workloads. Certain features might not be supported or might have constrained capabilities. For more information, see Supplemental Terms of Use for Microsoft Azure Previews.
Full functionality with integrated compute resource
Some features of the Responsible AI image dashboard require dynamic, on-the-fly, and real-time computation. When you connect to a compute resource you enable full functionality of unique components to the image scenario:
- For object detection, setting an Intersection of Union threshold is disabled by default, and only enabled if a compute resource is attached.
- Enable pre-computing of all model explanations when submitting a DPv2 job, instead of loading explanations on-demand.
You can also find this information on the Responsible AI dashboard page by selecting the Information icon, as shown in the following image:
Overview of features in the Responsible AI image dashboard
The Responsible AI dashboard includes a robust, rich set of visualizations and functionality to help you analyze your machine learning model or make data-driven business decisions:
- Error Analysis (Image classification & multi-classification only)
- Model overview
- Data explorer
- Model interpretability
Error analysis
Error analysis tools are available for image classification and multi-classification to accelerate detection of fairness errors and identify under/overrepresentation in your dataset. Instead of passing in tabular data, you can run error analysis on specified image metadata features by including metadata as additional columns in your mltable dataset. To learn more about error analysis, see Assess errors in machine learning models.
Model overview
The model overview component provides a comprehensive set of performance metrics for evaluating your computer vision model, along with key performance disparity metrics across specified dataset cohorts.
Note
Performance metrics will display N/A at its initial state and while metric computations are loading.
Dataset cohorts
On the Dataset cohorts pane, you can investigate your model by comparing the model performance of various user-specified dataset cohorts (accessible via the Cohort settings icon).
Multiclass classification:
Object detection:
- Help me choose metrics: Select this icon to open a panel with more information about what model performance metrics are available to be shown in the table. Easily adjust which metrics to view by using the multi-select dropdown list to select and deselect performance metrics.
- Choose aggregation: Select this button to which aggregation method to apply, affecting the calculation of Mean Average Precision.
- Choose class label: Select which class labels are used to calculate class-level metrics (for example, average precision, average recall).
- Set Intersection of Union (IOU) threshold – Object Detection only: Set an IOU threshold value (Intersection of Union between ground truth & prediction bounding box) that defines error and affects calculation of model performance metrics. For example, setting an IOU of greater than 70% means that a prediction with greater than 70% overlap with ground truth is True. This feature is disabled by default, and can be enabled by attaching a Python backend.
- Table of metrics for each dataset cohort: View columns of dataset cohorts, the sample size of each cohort, and the selected model performance metrics for each cohort – aggregated based on the selected aggregation method.
- Visualizations
- Bar graph (Image Classification, Multilabel classification): Compare aggregated performance metrics across selected dataset cohort(s).
- Confusion matrix (Image Classification, Multilabel classification): View a selected model performance metric across selected dataset cohort(s) and selected class(es).
- Choose metric (x-axis): Select this button to choose which metric to view in the visualization (confusion matrix or scatterplot).
- Choose cohorts (y-axis): Select this button to choose which cohorts to view in the confusion matrix..
Computer vision scenario | Metric |
---|---|
Image Classification | Accuracy, precision, F1 score, recall |
Image Multilabel Classification | Accuracy, precision, F1 score, recall |
Object Detection | Mean average precision, average precision, average recall |
Feature cohorts
On the Feature cohorts pane, you can investigate your model by comparing model performance across user-specified sensitive and non-sensitive features (for example, performance for cohorts across various image metadata values like gender, race, and income). To learn more about feature cohorts, see the feature cohorts section of Responsible AI dashboard.
Feature cohorts for multiclass classification:
Feature cohorts for object detection:
Data explorer
The data explorer component contains multiple panes to provide various perspectives of your dataset.
Image explorer view
The image explorer pane displays images instances of model predictions, automatically categorized by correct and incorrectly labeled predictions. This view helps you quickly identify high-level patterns of error in your data and select which instances to investigate more deeply.
For image classification and multiclassification, incorrect predictions refer to images where the predicted class label differs from ground truth. For object detection, incorrect predictions refer to images where:
- At least one object was incorrectly labeled
- Incorrectly detecting an object class when a ground truth object doesn't exist
- Failing to detect an object class when a ground truth object exists
Note
If object(s) in an image was correctly labeled but with an IOU score below the default threshold of 50%, the prediction bounding box for the object will not be visible, but the ground truth bounding box will be visible. The image instance would appear in the error instance category. Currently, it is not possible to change the default IOU threshold in the Data Explorer component.
Image explorer for multilabel classification:
Image explorer for object detection:
- Select a dataset cohort to explore: View images across all data or for specific user-defined cohorts.
- Set thumbnail size: Adjust the size of image cards displayed in this page.
- Set an Intersection of Union (IOU) threshold – Object Detection only: Changing the IOU threshold impacts which images are considered an incorrect prediction.
- Image card: Each image card displays the image, predicted class labels (top), and ground truth class labels (bottom). For object detection, bounding boxes for detected objects are also shown.
- Create a new dataset cohort with filters: Filter your dataset by index, metadata values, and classification outcome. You can add multiple filters, save the resulting filtered data with a specified cohort name, and automatically switch your image explorer view to display contents of your new cohort.
Selecting an image instance
By selecting an image card, you can access a flyout to view the following components supporting analysis of model predictions:
Explanations for multiclass classification:
Explanations for object detection:
- View predicted and ground truth outcomes: In comma-separated format, view the predicted & corresponding ground truth class label for the image or objects in the image.
- Metadata: View image metadata values for the selected instance.
- Explanation: View a visualization (SHAP feature attributions – image classification & multi-classification, D-Rise saliency map – object detection) to gain insight on model behavior leading to the execution of a computer vision task.
Table view
The Table view pane shows you a table view of your dataset with rows for each image instance in your dataset, and columns for the corresponding index, ground truth class labels, predicted class labels, and metadata features.
- Manually select images to create a new dataset cohort: Hover on each image row and select the checkbox to include images in your new dataset cohort. Keep track of the number of images selected and save the new cohort.
Table view for multiclass classification:
Table view for object detection:
Class view
The Class view pane breaks down your model predictions by class label. You can identify error patterns per class to diagnose fairness concerns and evaluate under/overrepresentation in your dataset.
- Select label type: Choose to view images by the predicted or ground truth label.
- Select labels to display: View image instances containing your selection of one or more class labels.
- View images per class label: Identify successful and error image instances per selected class label(s), and the distribution of each class label in your dataset. If a class label has "10/120 examples", out of 120 total images in the dataset, 10 images belong to that class label.
Class view for multiclass classification:
Class view for object detection:
Model interpretability
For AutoML image classification models, four kinds of explainability methods are supported, namely Guided backprop, Guided gradCAM, Integrated Gradients and XRAI. To learn more about the four explainability methods, see Generate explanations for predictions.
Note
- These four methods are specific to AutoML image classification only and will not work with other task types such as object detection, instance segmentation etc. Non-AutoML image classification models can leverage SHAP vision for model interpretability.
- The explanations are only generated for the predicted class. For multilabel classification, a threshold on confidence score is required, to select the classes for which the explanations are generated. See the parameter list for the parameter name.
Both AutoML and non-AutoML object detection models can leverage D-RISE to generate visual explanations for model predictions.
For information about vision model interpretability techniques and how to interpret visual explanations of model behavior, see Model interpretability.
Next steps
- Learn more about the concepts and techniques behind the Responsible AI dashboard.
- View sample YAML and Python notebooks to generate a Responsible AI dashboard with YAML or Python.
- Learn more about how you can use the Responsible AI image dashboard to debug image data and models and inform better decision-making in this tech community blog post.
- Learn about how the Responsible AI dashboard was used by Clearsight in a real-life customer story.
Feedback
https://aka.ms/ContentUserFeedback.
Coming soon: Throughout 2024 we will be phasing out GitHub Issues as the feedback mechanism for content and replacing it with a new feedback system. For more information see:Submit and view feedback for