Abstract
Purpose
The quantitative detection of failure modes is important for making deep neural networks reliable and usable at scale. We consider three examples for common failure modes in image reconstruction and demonstrate the potential of uncertainty quantification as a fine-grained alarm system.
Methods
We propose a deterministic, modular and lightweight approach called Interval Neural Network (INN) that produces fast and easy to interpret uncertainty scores for deep neural networks. Importantly, INNs can be constructed post hoc for already trained prediction networks. We compare it against state-of-the-art baseline methods (MCDrop, ProbOut).
Results
We demonstrate on controlled, synthetic inverse problems the capacity of INNs to capture uncertainty due to noise as well as directional error information. On a real-world inverse problem with human CT scans, we can show that INNs produce uncertainty scores which improve the detection of all considered failure modes compared to the baseline methods.
Conclusion
Interval Neural Networks offer a promising tool to expose weaknesses of deep image reconstruction models and ultimately make them more reliable. The fact that they can be applied post hoc to equip already trained deep neural network models with uncertainty scores makes them particularly interesting for deployment.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
The reconstruction of unknown signals from indirect measurements plays an important role in many applications, including medical imaging [2, 14]. Typically, such tasks are modeled as finite-dimensional linear inverse problems
where \(\varvec{x}\in \mathbb {R}^n\) is the signal of interest, \(\varvec{A}\in \mathbb {R}^{m\times n}\) denotes the forward operator representing a physical measurement process, and \(\varvec{\eta }\in \mathbb {R}^m\) is modeling noise in the measurements. Important examples include magnetic resonance imaging and computed tomography, where \(\mathbf {A}\) is a subsampled discrete Fourier or Radon transform, respectively. Solving the inverse problem (1) requires computing an approximate reconstruction of \(\varvec{x}\) from the observed measurements \(\varvec{y}\).
Classical reconstruction methods, e.g., based on sparse regularization models, constitute the state of the art for solving (1) in many cases and are backed by theoretical guarantees [8]. Recently, data-driven deep learning methods are increasingly gaining attention and are repeatedly able to outperform traditional solvers in terms of empirical reconstruction performance or speed, see for example [2].
Despite the advantages, the use of deep learning methods in sensitive applications such as clinical diagnosis is still a concern [23], due to questions regarding the reliability and robustness of the obtained reconstructions when compared to traditional approaches [1, 13]. What is more, erroneous artifacts in the reconstructed signals can be hard to detect as they tend to “blend in” well with the rest of the signal.
Various approaches for incorporating uncertainty quantification (UQ) into deep learning have been proposed to address these issues [10, 16, 18, 22]. However, as we demonstrate, existing UQ approaches come with limitations regarding their capacity to detect failure modes or their post hoc applicability to trained deep learning models.
In this work, we consider a straight-forward approach to solving (1) by employing a neural network to post-process a standard model-based inversion as in [14]. This reconstruction is given by
where \(\varvec{\varPhi }:\mathbb {R}^n\rightarrow \mathbb {R}^n\) is a neural network trained to minimize the loss \(\Vert \varvec{x}-\varvec{\varPhi }(\varvec{A}^\dagger (\varvec{y}))\Vert _2^2\) and \(\varvec{A}^\dagger :\mathbb {R}^m\rightarrow \mathbb {R}^n\) denotes the non-learned model-based inversion (e.g., the filtered back-projection in the case of Radon measurements). We will denote \(\varvec{z}=\varvec{A}^\dagger (\varvec{y})\) in the following. Given \(\varvec{y}\) or \(\varvec{z}\), a UQ method is supposed to extend the predicted reconstruction \(\varvec{\varPhi }(\varvec{z})\) by a component-wise uncertainty score \(\varvec{u}(\varvec{z})\) that provides additional information regarding the reliability of the reconstruction. Therefore, \(\varvec{u}(\varvec{z})\) should be correlated with the component-wise error \(|\varvec{x}-\varvec{\varPhi }(\varvec{z})|\). We evaluate this for three different failure modes [7] that can arise during inference (see “Experiment B (i): general prediction error detection” section to “Experiment B (iii): Atypical Artifact Detection” section for more details):
-
(i)
Errors caused solely by the ill-posedness of (1), which is mostly determined by the strength of measurement noise and the amount of undersampling,
-
(ii)
Errors caused by adversarial perturbations to the network inputs,
-
(iii)
Errors caused by atypical artifacts that have not been seen during the training.
Our main contributions can be summarized as follows: We present a deterministic, modular and fast UQ-method for deep neural networks (DNNs), called Interval Neural Networks (INN). We evaluate INNs for the detection of the three different image reconstruction failure modes and demonstrate that they provide improved results compared to two existing UQ methods.
Related work
Whereas a number of methods from classical statistical learning theory, such as Gaussian processes and approximations thereof [6, 19], come with built-in uncertainty estimates, DNNs have been limited in this regard. A surge of efforts to treat neural networks from a variational perspective [3, 16] started to change that. In addition, there exist strands of research in deep learning explicitly occupied with the detection of failure modes caused by adversarial and out of distribution (OoD) inputs. These include Maximum Mean Discrepancy, Kernel Density Estimation and other tools, see [5] or the Minimum Covariance Determinant method [26], Support Vector Data Description [28], among others. We refer to [27] for a comprehensive overview. The detection of adversarial and OoD inputs in these works is typically done in the classification setting. We emphasize that image-to-image regression is a fundamentally different task: While classification is inherently discontinuous, image reconstruction addresses a problem that allows for stable solution methods in many cases, e.g., by sparse regularization. Furthermore, we are not interested in a crude, outright rejection of data points in the input space but rather seek to obtain fine-grained information about erroneous artifacts in the output space. More closely related to our goal is Monte Carlo dropout (MCDrop) [10] and direct variance estimation (ProbOut) [12], where epistemic and aleatoric uncertainty quantification was considered for segmentation and depth-estimation tasks. Hence, we include their approaches as baseline comparison methods, see “Baseline UQ methods” section.
Methods
Popular existing UQ frameworks for DNNs place parametric densities, most commonly Gaussian densities, over the DNN parameters or predictions. Instead of using specific parametrized densities, our INN method relies on bounding distributions using intervals. This results in a flexible and modular method that can be applied post hoc to a given DNN \(\varvec{\varPhi }\) that has already been trained. A schematic illustration is provided in Fig. 1: The INN is formed by wrap** additional weight and bias intervals around the weights and biases of the underlying prediction DNN. This allows us to equip the DNN \(\varvec{\varPhi }\) with uncertainty capabilities without the need to modify \(\varvec{\varPhi }\) itself. After training the INN we obtain prediction intervals that are guaranteed to contain the original prediction of the underlying network and are easy to interpret. They provide exact upper and lower bounds for the range of possible values that the DNN prediction may take when slightly modifying the network parameters within the prescribed weight and bias intervals.
Previously, the capacity of neural networks with interval weights and biases was evaluated for fitting interval-valued functions [11]. In contrast to [11], our targets \(\varvec{x}_i\) are neither interval-valued nor univariate, leading to a different loss function which allows us to equip trained neural networks with uncertainty capabilities post hoc. For a direct comparison, see 3 in 3.2 and Equation (18) in [11]. Further, [17, 30] explored neural networks implementing interval arithmetic for robust classifications. However, in their setting, the focus is purely on representing the inputs or outputs as intervals but not the weights and biases. In contrast, our proposed INNs determine interval bounds for all network parameters with the goal of providing uncertainty scores for the predictions of an underlying DNN.
Arithmetic of Interval Neural Networks
We will now give a description of those INN mechanisms that deviate from standard DNNs. The forward propagation of a single input \(\varvec{z}\) through a DNN is replaced by the forward propagation of a component-wise interval-valued input \([\underline{\varvec{z}}, \overline{\varvec{z}}]\) through the INN. This can be expressed similarly to standard feed-forward neural networks but using interval arithmetic instead. For interval-valued weight matrices \([\underline{\varvec{W}}, \overline{\varvec{W}}]\) and bias vectors \([\underline{\varvec{b}}, \overline{\varvec{b}}]\), the propagation through the \(\ell \)-th network layer can be expressed as
For nonnegative \([\underline{\varvec{z}}, \overline{\varvec{z}}]^{(\ell )}\), for example when using a nonnegative activation function \(\varrho \) such as the ReLU in the previous layer, we can explicitly rewrite (2) as
where the maximum and minimum are computed component-wise. Similarly, for point intervals \(\underline{\varvec{z}}^{(\ell )}=\overline{\varvec{z}}^{(\ell )}=:\varvec{z}^{(\ell )}\), for example, as inputs to the first network layer, we can rewrite (2) as
regardless of whether \(\varvec{z}^{(\ell )}\) is nonnegative or not. Optimizing the INN parameters requires obtaining the gradients of these operations. This can be achieved using automatic differentiation (backpropagation) in the same way as for standard neural networks.
Training Interval Neural Networks
Let \(\varvec{W}^{(\ell )}\) and \(\varvec{b}^{(\ell )}\) be the weights and biases of the underlying prediction network \(\varvec{\varPhi }\) and let \(\overline{\varvec{\varPhi }}:\mathbb {R}^n \rightarrow \mathbb {R}^n\) and \(\underline{\varvec{\varPhi }}:\mathbb {R}^n \rightarrow \mathbb {R}^n\) denote the functions map** a point interval input \(\varvec{z}\) to the upper and the lower interval bounds in the output layer of the INN respectively. Given data samples \(\left\{ \varvec{z}_i,\varvec{x}_i \right\} _{i=1}^m\) the INN parameters \([\underline{\varvec{W}}, \overline{\varvec{W}}]^{(\ell )}\) and \([\underline{\varvec{b}}, \overline{\varvec{b}}]^{(\ell )}\) are trained by minimizing the empirical loss
subject to the constraints \(\underline{\varvec{W}}^{(\ell )}\le \varvec{W}^{(\ell )}\le \overline{\varvec{W}}^{(\ell )}\) and \(\underline{\varvec{b}}^{(\ell )}\le \varvec{b}^{(\ell )}\le \overline{\varvec{b}}^{(\ell )}\) for each layer. This way \(\underline{\varvec{\varPhi }}(\varvec{z})\le \varvec{\varPhi }(\varvec{z})\le \overline{\varvec{\varPhi }}(\varvec{z})\) is always guaranteed. The first two terms in (3) encourage that the predicted interval \([\underline{\varvec{\varPhi }}(\varvec{z}_i),\overline{\varvec{\varPhi }}(\varvec{z}_i)]\) should contain the target signal \(\varvec{x}_i\), while penalizing each component that lies outside with the squared distance to the nearest interval bound. The second term penalizes the interval size, so that the predicted intervals cannot grow arbitrarily large. While a quadratic penalty of the interval size is also possible and leads to similar theoretical bounds as in (4), we choose to minimize the \(\ell _1\)-norm to make the intervals more outlier inclusive. In addition, the tightness parameter \(\beta > 0\) can further tune the outlier-sensitivity of the intervals. This allows for a calibration of the INN uncertainty scores according to an application specific risk-budget. In practice, we found that choosing \(\beta \) similar to the mean absolute error of the underlying prediction network yields a good trade-off between coverage [9] and tightness.
Properties of Interval Neural Networks
The uncertainty estimate of an INN is given by the width of the prediction interval, i.e., \(\varvec{u}(\varvec{z}) = \overline{\varvec{\varPhi }}(\varvec{z}) - \underline{\varvec{\varPhi }}(\varvec{z})\). In terms of computational overhead, INNs scale linearly in the cost of evaluating the underlying prediction DNN with a constant factor 2. In contrast, the popular MCDrop [10] scales linearly with a factor T which is proportional to the number of stochastic forward passes and at least \(T=10\) is recommended by the authors, see “Baseline UQ methods” section.
Further, INNs come with theoretical coverage guarantees that can be derived from the Markov inequality: Assuming that the loss (3) is optimized during training to yield an INN with vanishing expected gradient with respect to the data distribution, we obtain
for any \(\lambda > 0\). In other words, for input and target pair \((\varvec{z},\varvec{x})\) the probability of any component of the target lying inside the predicted interval enlarged by \(\lambda \beta \) is at least \(1-\frac{1}{\lambda }\). As \(\beta \) is usually very small, this ensures a fast decay of the probability of the components of \(\varvec{x}\) lying outside the predicted interval bounds. Consequently, a component with a small uncertainty score was correctly reconstructed up to small error with a high probability. Of course, the training distribution needs to be well representative of the true data distribution to extrapolate this property to unseen data.
Finally, the optimization of the loss (3) yields additional information: If the prediction \(\varvec{\varPhi }(\varvec{z})\) lies closer to one boundary of the predicted interval, the true target \(\varvec{x}\) has a higher probability of lying on the other side of the interval. Consequently, INNs can provide directional uncertainty scores. A quantitative assessment of this capability is given in Fig. 3c+d. We note that it is also possible to explore asymmetric uncertainty estimates in the probabilistic setting, e.g., via exponential family distributions [29] or quantile regression [24]. In contrast to INNs, these methods cannot be applied post hoc as they require substantial modifications to the underlying prediction network.
Baseline UQ methods
In addition to our INN approach, we consider two other related and popular UQ baseline methods for comparison. First, Monte Carlo dropout (MCDrop) [10] obtains uncertainty scores as the sample variance of multiple stochastic forward passes of the same input signal. In other words, if \(\varvec{\varPhi }_1,\dots ,\varvec{\varPhi }_T\) are realizations of independent draws of random dropout masks for the same underlying network \(\varvec{\varPhi }\), the component-wise uncertainty estimate is \(\varvec{u}_{{\textsc {MCDrop}}}(\varvec{z}) = (\tfrac{1}{T-1} ( \sum _{t=1}^T \varvec{\varPhi }_t(\varvec{z})^2- \tfrac{1}{T} (\sum _{t=1}^T \varvec{\varPhi }_t(\varvec{z}))^2 ))^{1/2}\). Second, a direct variance estimation (ProbOut) was proposed in [22] and later expanded in [12]. Here, the number of output components of the prediction network is doubled and trained to approximate the mean and variance of a Gaussian distribution. The resulting network \(\varvec{\varPhi }_{{\textsc {ProbOut}}}:\mathbb {R}^n\rightarrow \mathbb {R}^n\times \mathbb {R}^n, \varvec{z}\mapsto (\varvec{\varPhi }_{\text {mean}}(\varvec{z}), \varvec{\varPhi }_{\text {var}}(\varvec{z}))\) is trained by minimizing the empirical loss \(\sum _i \Vert (\varvec{y}_i-\varvec{\varPhi }_{\text {mean}}(\varvec{z}_i)) / \sqrt{\varvec{\varPhi }_{\text {var}}(\varvec{z}_i)}\Vert _2^2 + \Vert \log \varvec{\varPhi }_{\text {var}}(\varvec{z}_i)\Vert _1\). The component-wise uncertainty score of ProbOut is \(\varvec{u}_{{\textsc {ProbOut}}}(\varvec{z}) = (\varvec{\varPhi }_{\text {var}}(\varvec{z}))^{1/2}\). Note that, in contrast to INN and MCDrop, the ProbOut approach requires the incorporation of UQ already during training. Thus, it cannot be employed as a post hoc evaluation of an already trained, underlying network \(\varvec{\varPhi }\). The role of the actual prediction network is taken by \(\varvec{\varPhi }_\text {mean}\).
Experiments
We present experiments for two different inverse problems. First, a deconvolution task with 1D signals, and second a tomography task on real-world 2D image signals. Both setups are described in more detail below. The description of all hyperparameters for the experiments is kept brief and we refer to our publicly available code at https://github.com/luisoala/inn for full details.
Case study A: deconvolution of 1D signals
We start with a synthetic, didactic experiment, inspired by a one-dimensional deconvolution task, to demonstrate the properties of INNs discussed in “Properties of Interval Neural Networks” section. For this purpose, we choose \(n=m=512\) and \(\varvec{A}= \varvec{D}^\top \varvec{S}\varvec{D}\), where \(\varvec{D}\) is a discrete cosine transform (Type I DCT) and \(\varvec{S}\) is a diagonal matrix with entries \(s_j = \left( \frac{n-j}{n-1}\right) ^\nu \in [0,1]\), that decay with a fixed exponent \(\nu =8\). We draw synthetically generated signals \(\varvec{x}\) from a distribution of piecewise constant functions with random jump positions and heights, see Fig. 2. The corresponding measurements \(\varvec{y}\) are computed according to (1). We generate a data set consisting of 2000 sample pairs \((\varvec{y}_i, \varvec{x}_i)\), 1600 of which were used for training, 200 for validation and 200 for testing. The underlying prediction network \(\varvec{\varPhi }\) is a convolutional neural network (consisting of ten convolutional layers and three dropout layers in between) trained to directly map \(\varvec{y}\) to \(\varvec{x}\), i.e., we use \(\varvec{A}^\dagger =\mathrm {Id}\) and thus \(\varvec{z}=\varvec{A}^\dagger \varvec{y}=\varvec{y}\) in this experiment. We trained the underlying network \(\varvec{\varPhi }\) for 100 epochs using Adam [15]. The interval parameters of the INN were subsequently trained for another 100 epochs with \(\beta =2\cdot 10^{-3}\). For the MCDrop comparison, we use \(T=64\) samples. The ProbOut model was trained in the same way as \(\varvec{\varPhi }\) using 100 Adam epochs. Note that all subsequent evaluations, as well as the plots in Fig. 2 are computed using test samples.
In order to evaluate the UQ methods’ abilities to capture uncertainty due to noisy data, we consider additive Gaussian noise \(\varvec{\eta }\sim \mathcal {N}(\varvec{0},\sigma ^2\cdot \mathrm {Id})\) on the measurements over a range of noise levels \(\sigma \) (Fig. 3a) as well as \(\varvec{\eta }_1,\varvec{\eta }_2\sim \mathcal {N}(\varvec{0},\sigma ^2\cdot \mathrm {Id})\) on the measurements and targets, where (1) is adjusted to \(\varvec{y}=\varvec{A}(\varvec{x}+\varvec{\eta }_1)+\varvec{\eta }_2\) (Fig. 3b and right column of Fig. 2). In this case, INNs are able to capture the additional uncertainty of \(\varvec{\eta }_1\) using the bias parameters of the final network layer. In Fig. 3, it can be observed how in contrast to MCDrop, our method and ProbOut are able to capture independent noise in the data with ProbOut reacting to a lesser degree than the INN. Note also that in Fig. 3 some of the ProbOut evaluations are shifted to the right, indicating a reduced reconstruction performance compared to the other methods.
Finally, we determine the directional information of the INN uncertainty scores as discussed in “Properties of Interval Neural Networks” section. For this, we define the component-wise directionality ratio by \(\mathrm {DR}(\varvec{z}) = \max \{\overline{\varvec{\varPhi }}(\varvec{z}) -\varvec{\varPhi }(\varvec{z}),\varvec{\varPhi }(\varvec{z})-\underline{\varvec{\varPhi }}(\varvec{z})\} / {\min \{\overline{\varvec{\varPhi }}(\varvec{z})-\varvec{\varPhi }(\varvec{z}),\varvec{\varPhi }(\varvec{z})-\underline{\varvec{\varPhi }}(\varvec{z})\}}\), i.e., as the ratio between the larger and smaller part of the interval \([\underline{\varvec{\varPhi }}(\varvec{z}), \overline{\varvec{\varPhi }}(\varvec{z})]\) when divided by the prediction \(\varvec{\varPhi }(\varvec{z})\). The directionality accuracy (DA) is the relative frequency of target components corresponding to a given DR that are contained in the larger interval part. As displayed in Fig. 3c, d, INNs achieve a DA consistently above 0.5 (chance), indicating that the interval uncertainty scores contain directional information.
Case study B: limited angle computed tomography
Next, we consider a 2D computed tomography (CT) task on real-world data in order to evaluate the detection capabilities of the UQ methods with respect to the three failure modes (i)–(iii). More precisely, we consider limited angle CT, which has applications in dental tomography, breast tomosynthesis or electron tomography. For this, \(\varvec{A}\) is a subsampled discrete Radon transform with subsampling corresponding to a moderate missing wedge of \(30^\circ \). Limited angle measurements are simulated according to (1) and the non-learned inversion \(\varvec{A}^\dagger \) is based on the filtered backprojection algorithm (FBP) [21]. The underlying prediction network is a U-Net [25] variant. Our experiments are based on a data set consisting of \(512\times 512\) human CT scans from the AAPM Low Dose CT Grand Challenge data [20].Footnote 1 In total, it contains 2580 full-dose images with a slice thickness of 3mm from 10 patients. Eight of these ten patients were used for training (2036 samples), one for validation (214 samples) and one for testing (330 samples). We trained the underlying network \(\varvec{\varPhi }\) for 400 epochs using Adam [15]. The interval parameters of the INN were subsequently trained for another 15 epochs with \(\beta = 10^{-4}\). We limited the interval training to the last twelve layers. For the MCDrop comparison, we use \(T=128\) samples. The ProbOut model was trained in the same way as \(\varvec{\varPhi }\) using 400 Adam epochs.
Experiment B (i): general prediction error detection
First, we evaluate how helpful UQ scores are for estimating the prediction error caused by the ill-posedness of the challenging CT task, see Fig. 4. The wedge of missing angles in the measurements results in reconstruction artifacts especially at vertical edges in the images. In order to best visualize these geometric effects of the very structured null-space of the limited angle CT forward operator, we do not add noise in this experiment. INNs are clearly able to reveal the reconstruction uncertainty along the “missing edges.” For a more quantitative comparison of the UQ methods, we use the performance weighted correlation coefficient \( \mathrm {PWCC}(\varvec{z}, \varvec{x}) = \mathrm {corr}(|\varvec{\varPhi }(\varvec{z})-\varvec{x}|, \varvec{u}(\varvec{z})) / \Vert \varvec{\varPhi }(\varvec{z})-\varvec{x}\Vert _2^2 \) between the uncertainty score \(\varvec{u}\) and the absolute prediction error. Performance weighting (normalizing by the mean squared error of the prediction) is necessary to discourage rewards for poor prediction models with high uncertainties everywhere. The average results over the test set for three independent complete experimental runs are summarized in Table 1. Both INNs and MCDrop are able to detect prediction errors, with INNs achieving slightly higher correlations. In Fig. 3d, the directional accuracy of the INN is illustrated analogously to the corresponding experiment in “Case study A: deconvolution of 1D signals” section. Again it is consistently above 0.5 (chance).
Experiment B (ii): Adversarial Artifact Detection
Second, we assess the capacity of UQ methods to capture artifacts in the output that were caused by adversarial perturbations. To that end, we create perturbed inputs for each input sample \(\varvec{z}\) in the test set by employing the box-constrained L-BFGS algorithm [4] to minimize \(\Vert \varvec{\varPhi }(\varvec{z}_{\text {adv}})-\varvec{x}_{\text {adv.~tar.}}\Vert _2^2\) subject to \(\varvec{z}_{\text {adv}}\in [0,1]^n\). The adversarial targets \(\varvec{x}_{\text {adv.~tar.}}\) are created by subtracting 1.5 times its mean value from \(\varvec{x}_{\text {rec}}\) within a random \(50\times 50\) square, leading to clearly visible artifacts in the corresponding reconstructions; see Fig. 5. It is arguable, whether the technical aspects of such an adversarial perturbation (i.e., attacking subsequently to a model-based inversion) is a realistic scenario in the context of inverse problems. However, for our purposes, such a simple setup (see also [13]) is sufficient. We refer to [1], where adversarial noise is mapped to the measurement domain. In order to assess the detection capacity for this failure mode, the different UQ schemes are then used to produce uncertainty heatmaps for the generated adversarial inputs. A quantitative evaluation is carried out by computing the mean Pearson correlation coefficient between the pixel-wise change in the uncertainty heatmaps \(|\varvec{u}(\varvec{z})-\varvec{u}(\varvec{z}_{\text {adv}})|\) and the change of reconstructions \(|\varvec{x}_{\text {rec}}-\varvec{\varPhi }(\varvec{z}_{\text {adv}})|\). The results are summarized in Table 1 and illustrated in Fig. 5. We observe that both INN and ProbOut are able to detect the image region of adversarial perturbations, with INN achieving the highest correlation. This shows that both methods are able to visually highlight the effect that visually almost imperceptible input perturbations can have on the reconstructions.
Experiment B (iii): Atypical Artifact Detection
The third experiment is designed analogous to the setup described by [1], i.e., an atypical artifact, which was not present in the training data, is randomly placed in the input to produce \(\varvec{z}_{\text {OoD}}\). More precisely, the silhouette of a peace dove is inserted in each image of the test set; see Fig. 5. The simulation of the measurements and model-based inversions is carried out as before. A quantitative evaluation is carried out by computing the mean Pearson correlation coefficient between the change in the uncertainty heatmaps \(|\varvec{u}(\varvec{z})-\varvec{u}(\varvec{z}_{\text {OoD}})|\) and a binary mask marking the region of change in the inputs. This evaluation isolates the uncertainty caused by atypical artifacts and allows us to verify in a controlled manner how the uncertainty scores of each UQ method react to the artifacts. During deployment, such controlled isolation is not possible. Instead, the joint uncertainty heatmaps \(\varvec{u}(\varvec{z}_{\text {OoD}})\) will also capture other sources of uncertainty, thus providing a more comprehensive alarm system. The results are summarized in Table 1 and illustrated in Fig. 5. All three UQ methods are correlated with the input change; however, INN again achieves the highest correlation. This shows that UQ in general, and INNs in particular, can serve as a warning system for inputs containing atypical features that might otherwise lead to unnoticed and possibly erroneous reconstruction artifacts.
Conclusion
We introduced INNs as a deterministic, post hoc and fast approach for computing upper and lower bounds and subsequently uncertainty maps for pre-trained neural networks. We demonstrated that UQ in general and INNs in particular can be used to provide a fine-grained detection of failure modes of image reconstruction DNNs. INNs are able to capture uncertainty due to noise and can be used to obtain directional information. They perform well as an alarms system for errors due ill-posedness, adversarial noise and atypical artifacts and thus offer a promising tool to expose the weaknesses of deep image reconstruction models.
Notes
See: https://www.aapm.org/GrandChallenge/LowDoseCT/. We would like to thank Dr. Cynthia McCollough, the Mayo Clinic, and the American Association of Physicists in Medicine as well as the grants EB017095 and EB017185 from the National Institute of Biomedical Imaging and Bioengineering for providing the AAPM data.
References
Antun V, Renna F, Poon C, Adcock B, Hansen AC (2020) On instabilities of deep learning in image reconstruction and the potential costs of AI. Proc Natl Acad Sci 117(48):30088–30095. https://doi.org/10.1073/pnas.1907377117
Arridge S, Maass P, Öktem O, Schönlieb CB (2019) Solving inverse problems using data-driven models. Acta Numer 28:1–174
Barber D, Bishop C (1998) Ensemble learning in Bayesian neural networks. In: Generalization in neural networks and machine learning. Springer, pp 215–237
Byrd RH, Lu P, Nocedal J, Zhu C (1995) A limited memory algorithm for bound constrained optimization. SIAM J Sci Comput 16(5):1190–1208
Carlini N, Wagner D (2017) Adversarial examples are not easily detected: bypassing ten detection methods. In: Proceedings of the 10th ACM workshop on artificial intelligence and security, pp 3–14
Denker JS, Schwartz DB, Wittner BS, Solla SA, Howard RE, Jackel LD, Hopfield JJ (1987) Large automatic learning, rule extraction, and generalization. Complex Syst 1:877–922
Dietterich TG (2019) Robust artificial intelligence and robust human organizations. Front Comput Sci 13(1):1–3
Foucart S, Rauhut H (2013) A mathematical introduction to compressive sensing. Applied and Numerical Harmonic Analysis, Birkhäuser
Foygel Barber R, Candès EJ, Ramdas A, Tibshirani RJ (2020) The limits of distribution-free conditional predictive inference. Inf Inference J IMA 10(2):455–482. https://doi.org/10.1093/imaiai/iaaa017
Gal Y, Ghahramani Z (2016) Dropout as a Bayesian approximation: representing model uncertainty in deep learning. In: Balcan MF, Weinberger KQ (eds) Proceedings of The 33rd international conference on machine learning, proceedings of machine learning research, vol 48. PMLR, New York, pp 1050–1059
Garczarczyk Z (2000) Interval neural networks. In: 2000 IEEE international symposium on circuits and systems. Emerging technologies for the 21st Century. Proceedings (IEEE Cat No.00CH36353), vol 3. Presses Polytech. Univ. Romandes, Geneva, pp 567–570
Gast J, Roth S (2018) Lightweight probabilistic deep networks. In: 2018 IEEE/CVF conference on computer vision and pattern recognition, pp 3369–3378
Huang Y, Würfl T, Breininger K, Liu L, Lauritsch G, Maier A (2018) Some investigations on robustness of deep learning in limited angle tomography. In: Frangi AF, Schnabel JA, Davatzikos C, Alberola-López C, Fichtinger G (eds) Medical image computing and computer assisted intervention—MICCAI 2018. Springer, Cham, pp 145–153
** KH, McCann MT, Froustey E, Unser M (2017) Deep convolutional neural network for inverse problems in imaging. IEEE Trans Image Process 26:4509–4522
Kingma DP, Ba J (2015) Adam: a method for stochastic optimization. In: Bengio Y, LeCun Y (eds) 3rd international conference on learning representations, ICLR 2015, San Diego, CA, USA, May 7–9, 2015, conference track proceedings
Kingma DP, Salimans T, Welling M (2015) Variational dropout and the local reparameterization trick. In: Proceedings of the 28th international conference on neural information processing systems—Volume 2, NIPS’15. MIT Press, Cambridge, pp 2575–2583
Kowalski PA, Kulczycki P (2017) Interval probabilistic neural network. Neural Comput Appl 28(4):817–834
Lakshminarayanan B, Pritzel A, Blundell C (2017) Simple and scalable predictive uncertainty estimation using deep ensembles. In: Guyon I, Luxburg UV, Bengio S, Wallach H, Fergus R, Vishwanathan S, Garnett R (eds) Advances in neural information processing systems, vol 30. Curran Associates Inc, New York, pp 6402–6413
MacKay DJC (1992) Bayesian methods for adaptive models. PhD thesis, California Institute of Technology
McCollough C (2016) Tu-fg-207a-04: overview of the low dose CT grand challenge. Med Phys 43(6 Part 35):3759–3760
Natterer F (2001) The mathematics of computerized tomography. SIAM, Philadelphia
Nix DA, Weigend AS (1994) Estimating the mean and variance of the target probability distribution. In: Proceedings of 1994 IEEE international conference on neural networks (ICNN’94), vol 1, pp 55–60. https://doi.org/10.1109/ICNN.1994.374138
Oala L, Fehr J, Gilli L, Balachandran P, Leite AW, Calderon-Ramirez S, Li DX, Nobis G, Alvarado EAM, Jaramillo-Gutierrez G, Matek C, Shroff A, Kherif F, Sanguinetti B, Wiegand T (2020) Ml4h auditing: from paper to practice. In: Proceedings of the machine learning for health NeurIPS workshop, proceedings of machine learning research, vol 136. PMLR, pp 280–317. http://proceedings.mlr.press/v136/oala20a.html
Rodrigues F, Pereira FC (2020) Beyond expectation: deep joint mean and quantile regression for spatiotemporal problems. IEEE Trans Neural Netw Learn Syst 31(12):5377–5389
Ronneberger O, Fischer P, Brox T (2015) U-Net: convolutional networks for biomedical image segmentation. In: Navab N, Hornegger J, Wells WM, Frangi AF (eds) Medical image computing and computer-assisted intervention–MICCAI 2015, Lecture Notes in Computer Science. Springer, Berlin, pp 234–241
Rousseeuw PJ (1984) Least median of squares regression. J Am Stat Assoc 79(388):871–880
Ruff L, Kauffmann JR, Vandermeulen RA, Montavon G, Samek W, Kloft M, Dietterich TG, Müller KR (2021) A unifying review of deep and shallow anomaly detection. Proc IEEE 109(5):756–795. https://doi.org/10.1109/JPROC.2021.3052449
Tax DMJ, Duin RPW (2004) Support vector data description. Mach Learn 54(1):45–66
Wang H, **ngjian S, Yeung DY (2016) Natural-parameter networks: a class of probabilistic neural networks. In: Advances in neural information processing systems, pp 118–126
Yang D, Wu W (2012) A smoothing interval neural network. Discrete Dyn Nat Soc. https://doi.org/10.1155/2012/456919
Funding
Open Access funding enabled and organized by Projekt DEAL. J.M. acknowledges support by DFG-RTG 2260 BIOQIC. M.M. acknowledges support by DFG-SPP 1798 Grants KU 1446/21 and KU 1446/23. G.K. is grateful to MATH+-BMRC Project EF1x1 for financial support. W.S. acknowledges support by BMBF/BIFOLD (ref. 01IS18025A and ref 01IS18037I).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
L.O. co-chairs the DAISAM working group at the ITU/WHO Focus Group AI4H.
Ethical approval
This article does not contain any studies with human participants or animals performed by any of the authors.
Informed consent
No new patient data were acquired as part of this work; public data were used from https://www.aapm.org/GrandChallenge/LowDoseCT/.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Oala, L., Heiß, C., Macdonald, J. et al. Detecting failure modes in image reconstructions with interval neural network uncertainty. Int J CARS 16, 2089–2097 (2021). https://doi.org/10.1007/s11548-021-02482-2
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11548-021-02482-2