# Copyright 2020 MONAI Consortium
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from typing import Callable, List, Optional, Sequence, Union
import torch
import torch.distributed as dist
from monai.metrics import compute_roc_auc
from monai.utils import Average, exact_version, optional_import
Metric, _ = optional_import("ignite.metrics", "0.4.2", exact_version, "Metric")
reinit__is_reduced, _ = optional_import("ignite.metrics.metric", "0.4.2", exact_version, "reinit__is_reduced")
[docs]class ROCAUC(Metric): # type: ignore[valid-type, misc] # due to optional_import
"""
Computes Area Under the Receiver Operating Characteristic Curve (ROC AUC).
accumulating predictions and the ground-truth during an epoch and applying `compute_roc_auc`.
Args:
to_onehot_y: whether to convert `y` into the one-hot format. Defaults to False.
softmax: whether to add softmax function to `y_pred` before computation. Defaults to False.
other_act: callable function to replace `softmax` as activation layer if needed, Defaults to ``None``.
for example: `other_act = lambda x: torch.log_softmax(x)`.
average: {``"macro"``, ``"weighted"``, ``"micro"``, ``"none"``}
Type of averaging performed if not binary classification. Defaults to ``"macro"``.
- ``"macro"``: calculate metrics for each label, and find their unweighted mean.
This does not take label imbalance into account.
- ``"weighted"``: calculate metrics for each label, and find their average,
weighted by support (the number of true instances for each label).
- ``"micro"``: calculate metrics globally by considering each element of the label
indicator matrix as a label.
- ``"none"``: the scores for each class are returned.
output_transform: a callable that is used to transform the
:class:`~ignite.engine.Engine` `process_function` output into the
form expected by the metric. This can be useful if, for example, you have a multi-output model and
you want to compute the metric with respect to one of the outputs.
device: device specification in case of distributed computation usage.
Note:
ROCAUC expects y to be comprised of 0's and 1's.
y_pred must either be probability estimates or confidence values.
"""
def __init__(
self,
to_onehot_y: bool = False,
softmax: bool = False,
other_act: Optional[Callable] = None,
average: Union[Average, str] = Average.MACRO,
output_transform: Callable = lambda x: x,
device: Optional[torch.device] = None,
) -> None:
super().__init__(output_transform, device=device)
self.to_onehot_y = to_onehot_y
self.softmax = softmax
self.other_act = other_act
self.average: Average = Average(average)
[docs] @reinit__is_reduced
def reset(self) -> None:
self._predictions: List[torch.Tensor] = []
self._targets: List[torch.Tensor] = []
[docs] @reinit__is_reduced
def update(self, output: Sequence[torch.Tensor]) -> None:
"""
Args:
output: sequence with contents [y_pred, y].
Raises:
ValueError: When ``output`` length is not 2. ROCAUC metric can only support y_pred and y.
ValueError: When ``y_pred`` dimension is not one of [1, 2].
ValueError: When ``y`` dimension is not one of [1, 2].
"""
if len(output) != 2:
raise ValueError(f"output must have length 2, got {len(output)}.")
y_pred, y = output
if y_pred.ndimension() not in (1, 2):
raise ValueError("Predictions should be of shape (batch_size, n_classes) or (batch_size, ).")
if y.ndimension() not in (1, 2):
raise ValueError("Targets should be of shape (batch_size, n_classes) or (batch_size, ).")
self._predictions.append(y_pred.clone())
self._targets.append(y.clone())
[docs] def compute(self):
_prediction_tensor = torch.cat(self._predictions, dim=0)
_target_tensor = torch.cat(self._targets, dim=0)
if dist.is_available() and dist.is_initialized() and not self._is_reduced:
# create placeholder to collect the data from all processes:
output = [torch.zeros_like(_prediction_tensor) for _ in range(dist.get_world_size())]
dist.all_gather(output, _prediction_tensor)
_prediction_tensor = torch.cat(output, dim=0)
output = [torch.zeros_like(_target_tensor) for _ in range(dist.get_world_size())]
dist.all_gather(output, _target_tensor)
_target_tensor = torch.cat(output, dim=0)
self._is_reduced = True
return compute_roc_auc(
y_pred=_prediction_tensor,
y=_target_tensor,
to_onehot_y=self.to_onehot_y,
softmax=self.softmax,
other_act=self.other_act,
average=self.average,
)