This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Uncertainty Propagation within Chained Models for Machine Learning Reconstruction of Neutrino-LAr Interactions

D. Douglas, 11footnotetext: Corresponding author.    A. Mishra    F. Petersen    D. Ratner    K. Terao
Abstract

Sequential or chained models are increasingly prevalent in machine learning for scientific applications, due to their flexibility and ease of development. Chained models are particularly useful when a task is separable into distinct steps with a hierarchy of meaningful intermediate representations. In reliability-critical tasks, it is important to quantify the confidence of model inferences. However, chained models pose an additional challenge for uncertainty quantification, especially when input uncertainties need to be propagated. In such cases, a fully uncertainty-aware chain of models is required, where each step accepts a probability distribution over the input space, and produces a probability distribution over the output space. In this work, we present a case study for adapting a single model within an existing chain, designed for reconstruction within neutrino-Argon interactions, developed for neutrino oscillation experiments such as MicroBooNE, ICARUS, and the future DUNE experiment. We test the performance of an input uncertainty-enabled model against an uncertainty-blinded model using a method for generating synthetic noise. By comparing these two, we assess the increase in inference quality achieved by exposing models to upstream uncertainty estimates.

1 Introduction

Deep learning has seen wide success and growing adoption in many fields of study and in various industries. Increasingly, these techniques are employed in reliability-critical tasks, where a model’s poor prediction can lead to grave outcomes, as in medical diagnostics [1], autonomous driving [2], and other applications where human health may be implicated. Similarly, deep learning based modeling is also finding adoption in scientific applications, such as neutrino physics reconstruction [3], control of particle accelerators [4], etc. In such fields, erroneous models can lead to deleterious effects on scientific knowledge and progress.

For such applications, in addition to the model’s predictions, we require a measure of the model’s confidence in its predictions, e.g., in the form of prediction intervals, calibrated probabilities, etc. One method for increasing the reliability of predictions is to produce a measure of confidence associated with a given model prediction. This additional information, along with the most probable or mean value, can enable a user to selectively ignore anomalous inferences, synthesize confident and uncertain inferences in a holistic way, and help diagnose mis-modeling for certain inputs.

Uncertainty in a model’s prediction can arise due to several sources [5, 6], typically categorized into epistemic and aleatoric. The choice of model’s architecture and complexity defines the space of possible functions which can be expressed, and the disparity between this space and the true process being modeled represents the model’s epistemic uncertainty [7]. Once the structure of a model is decided, it is then trained using a sample drawn from a distribution of possible inputs for the task. In this training process, a second type of uncertainty – aleatoric uncertainty – is introduced, through the disagreement between the training set and the real input space, as well as approximation error accumulated by fitting model weights numerically. In some cases, the input itself may have uncertainty, as it may be the product of a physical measurement, or the output of another uncertainty-enabled model. In this case, a model should propagate the input as a distribution, predicting an output distribution whose shape is a derived from both the input and model uncertainties.

With the advent of end-to-end deep learning approaches, most applications use a single model accepting data, carrying out representation learning, and outputting the final predictions [8]. However, in certain cases, it is desirable to approach the problem as a set of multiple discrete tasks in a sequence. This can be particularly useful when there are meaningful representations of the data between the input and final output spaces. These intermediate representations may be physically interesting in and of themselves. They may also correspond to an independently measurable quantity that can be used to form a specialized training or validation set for a separable task in the chain. In such a chain of models, errors and uncertainties from an upstream model in the sequence can affect the performance of the downstream models. Testing the performance of any of the individual models in the chain in isolation may not be an adequate measure of their final performance when deployed in the chain. Thus, it is essential to ascertain the effects of the propagation of uncertainties from the upstream models to the downstream models at each stage in the chain. Additionally, it is essential to understand the dependence of the final uncertainty upon the uncertainty of inputs, for various kinds of inputs with varying degrees of measurement error, including which steps in the chain have an inflationary or reductionary effect.

Chains of data driven models have been applied to complex tasks in material science [9], fluid mechanics [10], etc. Furthermore, they are often used in autonomous driving applications. Le et al. [11] compared the performance of chained neural networks to an end-to-end deep learning model for the task of locating nearest objects in geographic images. They found that while both approaches had similar accuracy, chained composite models exhibited an order of magnitude reduction in computational expense for training.

In this paper, we will focus on a sequential model which was developed for the reconstruction of neutrino-argon scattering events in a Liquid Argon Time Projection Chamber (LArTPC), a detector design widely used in neutrino oscillation experiments, such as the future Deep Underground Neutrino Experiment (DUNE). The initial input for this model, the detector output, is represented by a sparse point cloud – measurements of ionization signals left in the wake of energetic charged particles interacting in the active liquid argon (LAr) medium. The reconstruction model first performs image-based tasks, including semantic classification of voxels and location of key points like potential vertices, start and end points of track-like fragments. Further downstream models represent detector outputs as an abstracted particle flow description, reconstructing the physical type and kinematic properties of each particle as energy is deposited in the sensitive volume over the course of the interaction’s timeline.

A more complete description of this model can be found in Section 2.2. In the context of this reconstruction chain, we study the effects of uncertainty propagation in chained machine learning models. We compare the accuracy and uncertainty predictions of both uncertainty-aware (UA) and uncertainty-blinded models (referred to as “blind”), by producing parallel models trained with and without per-feature uncertainty estimates. We demonstrate improvements in these metrics in the context of the node and edge classification tasks performed by a Graph Neural Network (GNN) (the first model labelled as ”GrapPA“ in Figure 2).

2 Background: Liquid Argon Time Projection Chambers and ML Reconstruction

2.1 The Liquid Argon Time Projection Chamber

A LArTPC is a 2D/3D imaging detector technology used prominently in neutrino physics experiments because of their high quality spatial and energy resolution. A large volume of LAr acts as both the target for incoming neutrinos and the detection medium for imaging. When a neutrino interacts on an Argon nucleus within the detector, energetic charged particles are created at the interaction point, or vertex. These secondary particles continue through the material, leaving trails of ionization as they propagate through the material. Figure 1 shows an example of a neutrino interaction, with points representing the location of ionization signals within the detector.

In this study, we simulated the current design of the Deep Underground Neutrino Experiment (DUNE) near detector. Here, the imaging volume is 5 m ×\times 7 m ×\times 3 m recorded in the 3D image of a dimension 1 250 ×\times 1 750 ×\times 1 000 voxels.

A typical neutrino interaction may produce several particles with varying size including a straight proton trajectory of a few centimeters, a meters-long muon and/or charged pions as well as electromagnetic particles such as electrons each of which produces a fragmented trajectory (i.e. multiple disconnected short trajectories). Each voxel represents a region in which charge ionized by these particles may be measured. When charge meets a pre-configured threshold, it is digitized and saved. Thus, a typical interaction captured by the LArTPC design is densely sampled in the signal region (i.e. particle trajectories) without gaps for high precision physics inference, but is otherwise globally sparse.

Refer to caption
Figure 1: An example of a simulated neutrino interaction in the DUNE ND geometry, displaying typical degree of sparsity. This detector features a highly-segmented design, consisting of 35 dual-TPC modules packed in a 5 ×\times 7 grid.

2.2 Machine Learning-based Reconstruction of LArTPC Events (SPINE Reconstruction Model)

Scalable Particle Imaging with Neural Embedding (SPINE) [12] is the current state-of-the-art technique for analyzing LArTPC detector data. At its core, SPINE is a chain of multiple convolutional (CNNs) and graph neural networks which infer hierarchical physics features from the low to high level in sequential manner. The chain shown in Figure 2 can be broken into two parts: the first stage utilizes CNN-based algorithms to infer pixel-level information. The output of the first stage is particle fragments which are essentially clusters of pixels. Each fragment contains pixels that are densely connected (i.e. no gap) and several fragment-level information are also inferred using CNNs. For physics analysis, we are interested in individual particle instance which consists of one or more of these fragments.

The second stage of SPINE utilizes GNNs, noted as GrapPA in Figure 2, with the goal of identifying the right set of fragments that represent individual particle instance. Our study targets the GrapPA GNN model. The input graph is constructed by interpreting indvidual particle fragment from the output of the first stage as a node. The edges are constructed between all possible pairs, thus the input is a fully-connected graph. Then the edge/node regression/classification is applied to infer features. For instance, the edge classification is used to infer the right set of nodes (fragments) to infer a particle instance, and the node classification for classifying different types of particle fragments.

The input to the GrapPA stage is the fixed number of node and edge features which introduce a bottleneck of information propagation as they must represent features associated with an arbitrary number of pixels in the corresponding cluster. Therefore the uncertainties of input become critical to inform how accurately key features are captured. Furthermore, unlike the pixel-level features extracted by prior CNNs, GrapPA is the first stage in the chain to produce the outputs that are used by a wide range of physics analyses that follow after SPINE. Predicting the output uncertainty at the GrapPA stage will hence provide more informed outputs and can make an impact in physics measurements.

Refer to caption
Figure 2: Machine learning-based reconstruction chain for neutrino interaction analysis in a LArTPC [12]. In this paper, we examine the first model labelled as “GrapPA”, in which disconnected shower fragments are partitioned and initial fragments are identified.

3 Computational Approach

3.1 Literature Review

Numerous papers have studied the application of deep learning models for LArTPCs, including the reconstruction chain discussed above. Several of these have included studies of uncertainties for single-stage deep learning reconstruction-chain models; example tasks were single-particle classification, multi-particle classification, and semantic segmentation. [13, 14] compared different deep learning UQ approaches for epistemic uncertainties in the initial positions and directions of electrons scattered in Compton interactions in a low-energy LArTPC gamma ray telescope. [15] investigated the influence of uncertainties and errors in the detector response on the MicroBooNE physics analyses. This was carried out using an approach which is agnostic toward any specific Machine Learning model used for the reconstruction task. [16, 17] evaluated deep learning UQ methods on the task of particle classification using the PiLArNet Monte Carlo 3D LArTPC point cloud dataset [18]. However, while prior investigations have analyzed the epistemic and aleatoric uncertainties for single stages in the LArTPC reconstruction chain, there is a need to understand the propagation of input uncertainties through different stages in the reconstruction chain.

Outside of neutrino physics, various papers have studied the propagation of uncertainties through neural networks. Examples include variants of Moment Matching techniques [19] and Stable Distribution Propagation [20]. [21] showed a sampling-independent approach to propagating epistemic uncertainties in in autonomous driving based semantic segmentation on the CamVid dataset. For chained models, [22] formulated motion planning for autonomous driving as a continuous space, discrete-time model, where the first model stage predicts the future motion of other traffic participants and the ensuing model stage uses this input to determine the vehicle’s trajectory. Their results underscored the advantages of integrating uncertainties from prior stages in the motion planning task. Similarly, in autonomous vehicle motion planning, [23] have analyzed the need for calibrated uncertainties from upstream models for motion planning. In their analysis, the authors observed that better calibrated upstream predictions, led to improved metrics for the uncertainty-aware planner stage.

3.2 Method

In this study, we investigate how an uncertainty-enabled model within a large chain of such models makes well-calibrated inferences. We focus on Probabilistic Neural Networks (PNNs), which in this case means a model which predicts the parameters of a distribution over the prediction space. For a regression problem with Gaussian errors, this distribution may follow a normal distribution, while a 2-class classification problem may be expressed with a Bernoulli probability. A model should also be able to ingest the distributional description of an input and learn the connection between the shape of the input distribution and the confidence of the input.

To assess the quality of the uncertainty estimates, we construct a blinding test (Figure 3): we train two models identical in structure on the same input set, with one version receiving the full distributional information (mean and width parameters) as input features, while the other model receives only the mean information features. This second, “blind” model represents what a model with noisy deterministic inputs may hope to achieve, while the “uncertainty-aware” model can in principle make better predictions in terms of both accuracy and calibration of output uncertainties, without changing model complexity. As the two models share the same architecture, the blind model is given the same number of input features, but the input uncertainty features are masked with a static value.

fBlind\displaystyle f_{\mathrm{Blind}} :x^y,σy\displaystyle:\hat{x}\rightarrow y,\sigma_{y}
fUncertaintyAware\displaystyle f_{\mathrm{Uncertainty-Aware}} :x^,σx^y,σy\displaystyle:\hat{x},\sigma_{\hat{x}}\rightarrow y,\sigma_{y}
Refer to caption
Figure 3: Testing scheme for the performance of a chained model with and without access to upstream uncertainty information. A model under consideration (Here, “Probabilistic GNN”) recieves a noisy inference from an upstream model in the sequence. In the uncertainty-aware scheme, this model also sees a well-calibrated estimate of the uncertainty associated with the noise of the upstream model (dashed arrow). In the case of a blinded model, the upstream uncertainty estimate is replaced with a static value, masking the information, but retaining the shape of the input, so as to leave the model’s structure unaltered. Our hypothesis is that the uncertainty-aware model will always out-perform the blinded model either in terms of accuracy or in terms of improved confidence estimates.

In our case study, we do not have access to true upstream uncertainties, but can instead construct some noisy inputs by adding synthetic noise. We first choose a scale of uncertainty – 10% - 30% of the ground truth value in our example – based on our expected uncertainty output from the upstream step. Then, synthetic noise is added to the deterministic features as follows:

σi\displaystyle\sigma_{i} =αixi\displaystyle=\alpha_{i}x_{i}
αi\displaystyle\alpha_{i} U[0.1,0.3]\displaystyle\sim U[0.1,0.3]
x^i\displaystyle\hat{x}_{i} =xi+βi\displaystyle=x_{i}+\beta_{i}
β\displaystyle\beta N[0,σi]\displaystyle\sim N[0,\sigma_{i}]

Using this method for per-feature smearing, we obtain an f^i\hat{f}_{i} and σi\sigma_{i} per deterministic feature where the noise is normally distributed and σi\sigma_{i} gives an estimate of the error distribution, but not the true error |f^ifi||\hat{f}_{i}-f_{i}|.

The two approaches are trained on simulated data using a binary cross-entropy loss function.

In discussions of model performance, we will focus on the metrics of accuracy, binary cross-entropy, mis-calibration area, and binary score entropy. Accuracy here is the observed frequency of correctly-labelled classification items. Mis-calibration area (MCA) is an assessment of the correlation between the true error rate and the expected error rate given the model’s predicted uncertainty. Deviation of this metric from unity represents an inability of the predicted uncertainties to adequately map to the true error, and so the degree of this disagreement is represented by the area between the model’s uncertainty and the ideal predictor [24]. Binary entropy here is the entropy of the model’s output treated as a Bernoulli process [25]. This metric gives an estimate of the sharpness of the model’s predictions, as it indicates the likelihood of scores to be preferentially in the high-confidence regions – near 0 or 1, producing a low entropy value – or in the low-confidence regions – near the decision boundary, producing a high entropy value.

4 Results & Discussion

As an example for adding uncertainty quantification to the existing machine learning-based reconstruction chain, we examine the first GNN stage outlined in Figure 2, labelled “GrapPA”. At this stage in the reconstruction of a LArTPC image, it is common to find many independent interactions which are coincident in time. Each of the showers displays a high level of fragmentation, due to the prevalence of neutral particles which may be produced in one region of the detector and travel for some distance before re-interacting to again produce visible, charged particles [26].

The tasks at this stage are two-fold: to associate each fragment to the other fragments belonging to the same shower, and to identify the most upstream “parent” fragment within each shower. For this problem, the inputs are presented as a fully-connected graph, where each fragment is represented as a node with a feature vector whose values are derived from the fragment’s point cloud representation:

  • Fragment point cloud centroid (3 parameters)

  • Fragment point cloud covariance (upper diagonal) (6 parameters)

  • Fragment size (number of points) (1 parameter)

Likewise, each edge in the fully-connected graph is represented by a feature vector which is defined by the two point cloud representations of the fragments it connects:

  • Points of closest approach (6 parameters)

As described in Section 3.2, blinded and uncertainty-aware models are trained on simulated data of electromagnetic showers using a generic event generator [26]. Independent shower events are simulated and superimposed within the detector geometry to produce a data-like sample with a high degree of pileup.

This model is designed to infer both the node states and the edge states of the graph, and simultaneously predict a label for each of these tasks. In the edge task, the label indicates whether an edge connects two causally-related showers, so that all pairs of fragments within a shower will share an edge with label 1. Conversely, edges conneting pairs of fragments belonging to different showers are labelled 0. Second, the node task is a per-node label inference, where a true label of 1 corresponds to the fragment which initiated the larger shower structure, while downstream, “child” nodes have a true label of 0.

There are some significant qualitative differences in these tasks. The classes represented in the node task are inherently unbalanced, with one parent node corresponding to many child nodes (give an actual estimate of this imbalance). Likewise, the prevalence of truly “on” edges is much smaller than that of truly “off” edges in the edge classification task. Lastly, properly associating all of the fragments within a shower without including is more complicated than just classifying edges as “on” or “off”, as partitioning of the graph into its proper sub-graphs involves ensuring that all edges connecting unrelated showers are predicted as “off”. In the full ML-reco scheme, this is done with an adjustable cut, using an Adjusted Rand Index (ARI) metric in constructing each sub-graph [26].

The node classification task can be seen as somewhat simpler, in that the shape of each parent node will often present a “cone” which will point towards child nodes. It is observed in both the blinded and uncertainty-aware models that the node task is able to achieve a higher baseline accuracy than the edge task for each trained model. This logic can be applied to the edge classification task, to the extent that this effect will make classification of edges between parent nodes and their child nodes simpler, but adds no benefit for predicting the true label between child nodes within a shower and across unrelated showers.

Refer to caption

Figure 4: An example input for the shower fragment association task. Here, 2 independent shower objects are represented by 5 fragments. The point-cloud representation of each fragment is used to form a feature vector of uniform length, and each pair of fragment point clouds are likewise aggregated into a vector of fixed length. In this way, the fragmented point cloud images are represented as a fully-connected graph.

For a set of models trained with an injected noise between 10% and 30%, the resulting metrics are summarized in Table 1.

We see a marked separation in the accuracy of blinded vs. unblinded models for the edge classification task, shown in Figure 5. In the node classification task, shown in Figure 6 there is much less separation in performance, though there is a better overall accuracy when compared to the edge task.

Metric Accuracy Loss MCA Entropy
Blind (edge task)   0.846 ±\pm 0.001   0.333 ±\pm 0.001   0.0135 ±\pm 0.0074   0.482 ±\pm 0.006
UA (edge task) 0.864 ±\pm 0.001 0.301 ±\pm 0.001 0.0161 ±\pm 0.0091 0.437 ±\pm 0.006
Blind (node task) 0.923 ±\pm 0.000 0.188 ±\pm 0.000 0.0109 ±\pm 0.0029 0.275 ±\pm 0.005
UA (node task) 0.928 ±\pm 0.000 0.176 ±\pm 0.000 0.0117 ±\pm 0.0038 0.256 ±\pm 0.005
Table 1: Metrics evaluated on an ensemble of 50 models each of the Uncertainty-Aware (UA) and Uncertainty-Blinded (blind) models. See Figures 5 through 11
Refer to caption
Figure 5: Edge classification performance of the blinded GNN model compared to the uncertainty-aware model. For this task, we see a statistically significant improvement by exposing the model to input uncertainty information.
Refer to caption
Figure 6: Node classification performance of the blinded GNN model compared to the uncertainty-aware model. For this task, we see a smaller improvement when adding uncertainty information than is observed in the edge task.

There is also a significant difference in the final loss achieved by the two kinds of models. We see an improvement in loss for both the edge task and the node task in Figures 7 and 8.

Refer to caption
Figure 7: Edge classification loss distribution for the ensemble. The loss here is binary cross-entropy, which corresponds to the negative log-likelihood for a Bernoulli process.
Refer to caption
Figure 8: Node classification loss distribution for the ensemble.

Interestingly, both models produce equally well-calibrated scores with no additional methods applied to enforce this. Figure 9 shows broadly similar distributions for both kinds of models, with some individual uncertainty-aware models significantly out-performing the blind models. Figure 10 shows a nearly identical result for the two kinds of model when performing the node task.

Refer to caption
Figure 9: Edge mis-calibration area for the ensemble.
Refer to caption
Figure 10: Node mis-calibration area for the ensemble. Both kinds of models produce nearly identically well-calibrated inferences.

Lastly, as a measure of sharpness, or a model’s ability to make more confident predictions, we use the binary entropy of each model’s predictions. This metric yields a lower value for score distributions which are peaked near the true label values – corresponding to high-confidence inferences – while it yields high values when inferences are clustered near the decision boundary – indicating low-confidence. This metric is constrained to the [0, 1] interval. What is observed is a significantly higher clustering of scores in the regions of high-confidence for the uncertainty-aware models in both tasks, with the edge task exhibiting a more pronounced separation.

Refer to caption
Figure 11: Score binary entropy distribution for the edge task. Uncertainty-Aware models show a systematically lower value, indicating higher clustering towards the true label values, and more confident predictions.
Refer to caption
Figure 12: Score binary entropy distribution for the edge task. Uncertainty-Aware models show a systematically lower value, indicating higher clustering towards the true label values, and more confident predictions.

As the gain in improvement is expected to depend strongly upon the level of noise in the input features, we also examined models trained on noise at different scales. Figures 13 and 14 show the comparative accuracy of blind and uncertainty-aware models with input uncertainties ranging from 5% (fixed) to 5% - 50% per-feature. From this scan, we start to see a separation in performance between the blind and uncertainty-aware models around 5%-15%, while the node classification task shows a comparable level of performance throughout the range we probed.

Refer to caption
Figure 13: Improvement in accuracy of edge classification from a blinded model to an uncertainty-aware model for a range of injected feature noise.
Refer to caption
Figure 14: Node classification performance of the blinded GNN model compared to the uncertainty-aware model for a range of injected feature noise. Note that this task does not see the improvements observed in the edge classification task.

5 Conclusions

The objective of this study is to investigate if in a sequence of chained models, having access to predictive uncertainties from upstream models led to improvements in the performance of the downstream models. This improvement could be quantified via improved accuracy, better calibration, or increased confidence of predictions.

We have seen that for models embedded within a larger chain, as are an increasingly common approach in complex tasks, there is a need for uncertainty quantification in terms of model noise, and in terms of propagation of per-input uncertainties to output uncertainty.

In this paper, we have examined a GNN-based problem within a neutrino-argon scattering reconstruction algorithm, and observed that edge inference is greatly improved for noisy labels when a model is trained with an estimate of the input uncertainty. In this same case study, we have also seen that the node classification task is less sensitive to noise added in this way, as both the blinded and uncertainty-aware models perform approximately equally well for a wide range of noise parameters.

Acknowledgments

This work was supported by the U.S. Department of Energy, Office of Science, Office of High Energy Physics under Contract DE-AC02-76SF00515. Computational resources were provided by the SLAC Shared Scientific Data Facility. This work was also supported in part by Zoox, Inc.

References

  • [1] T. Liu, E. Siegel and D. Shen, Deep learning and medical image analysis for covid-19 diagnosis and prediction, Annual review of biomedical engineering 24 (2022) 179.
  • [2] S. Grigorescu, B. Trasnea, T. Cocias and G. Macesanu, A survey of deep learning techniques for autonomous driving, Journal of field robotics 37 (2020) 362.
  • [3] L. Dominé, K. Terao and D. Collaboration), Scalable deep convolutional neural networks for sparse, locally dense liquid argon time projection chamber data, Physical Review D 102 (2020) 012005.
  • [4] L. Gupta, A. Edelen, N. Neveu, A. Mishra, C. Mayes and Y.-K. Kim, Improving surrogate model accuracy for the lcls-ii injector frontend using convolutional neural networks and transfer learning, Machine Learning: Science and Technology 2 (2021) 045025.
  • [5] A. Kendall and Y. Gal, What uncertainties do we need in bayesian deep learning for computer vision?, Advances in neural information processing systems 30 (2017) .
  • [6] N. Tagasovska and D. Lopez-Paz, Single-model uncertainties for deep learning, Advances in neural information processing systems 32 (2019) .
  • [7] J. Antorán, J. Allingham and J.M. Hernández-Lobato, Depth uncertainty in neural networks, Advances in neural information processing systems 33 (2020) 10620.
  • [8] D. Amodei, S. Ananthanarayanan, R. Anubhai, J. Bai, E. Battenberg, C. Case et al., Deep speech 2: End-to-end speech recognition in english and mandarin, in International conference on machine learning, pp. 173–182, PMLR, 2016.
  • [9] T. Shafighfard, F. Kazemi, F. Bagherzadeh, M. Mieloszyk and D.-Y. Yoo, Chained machine learning model for predicting load capacity and ductility of steel fiber–reinforced concrete beams, Computer-Aided Civil and Infrastructure Engineering (2024) .
  • [10] M.F. Wahid, R. Tafreshi, Z. Khan and A. Retnanto, Multiphase flow rate prediction using chained multi-output regression models, Geoenergy Science and Engineering 231 (2023) 212403.
  • [11] A. Le Borgne, X. Marjou, F. Parzysz and T. Lemlouma, Comparing a composite model versus chained models to locate a nearest visual object, in 2023 IEEE/ACIS 8th International Conference on Big Data, Cloud Computing, and Data Science (BCD), pp. 231–236, IEEE, 2023.
  • [12] F. Drielsma, K. Terao, L. Dominé and D.H. Koh, Scalable, end-to-end, deep-learning-based data reconstruction chain for particle imaging detectors, 2021.
  • [13] M. Buuck, A. Mishra, E. Charles, N. Di Lalla, O. Hitchcock, M. Monzani et al., Low-energy electron-track imaging for a liquid argon time-projection-chamber telescope concept using probabilistic deep learning, The Astrophysical Journal 942 (2023) 77.
  • [14] B. Khek, A. Mishra, M. Buuck and T. Shutt, Gamma ray source localization for time projection chamber telescopes using convolutional neural networks, AI 3 (2022) 975.
  • [15] P. Abratenko, R. An, J. Anthony, L. Arellano, J. Asaadi, A. Ashkenazi et al., Novel approach for evaluating detector-related uncertainties in a lartpc using microboone data, The European Physical Journal C 82 (2022) .
  • [16] D. Koh, A. Mishra and K. Terao, Deep neural network uncertainty quantification for lartpc reconstruction, Journal of Instrumentation 18 (2023) P12013.
  • [17] K.D.H. Koh, A. Mishra and K. Terao, Evaluating deep learning uncertainty quantification methods for neutrino physics applications, in 35th Conference on Neural Information Processing Systems, 2021.
  • [18] C. Adams, K. Terao and T. Wongjirad, Pilarnet: Public dataset for particle imaging liquid argon detectors in high energy physics, arXiv preprint arXiv:2006.01993 (2020) .
  • [19] S. Wang and C. Manning, Fast dropout training, in international conference on machine learning, pp. 118–126, PMLR, 2013.
  • [20] F. Petersen, A. Mishra, H. Kuehne, C. Borgelt, O. Deussen and M. Yurochkin, Uncertainty quantification via stable distribution propagation, International Conference on Learning Representations (ICLR) (2024) .
  • [21] J. Postels, F. Ferroni, H. Coskun, N. Navab and F. Tombari, Sampling-free epistemic uncertainty estimation using approximated variance propagation, in Proceedings of the IEEE/CVF international conference on computer vision, pp. 2931–2940, 2019.
  • [22] W. Shao, J. Xu, Z. Cao, H. Wang and J. Li, Uncertainty-aware prediction and application in planning for autonomous driving: Definitions, methods, and comparison, arXiv preprint arXiv:2403.02297 (2024) .
  • [23] A. Ivanovic, M. Itkina, R. McAllister, I. Gilitschenski and F. Shkurti, On the importance of uncertainty calibration in perception-based motion planning, IEEE International Conference on Robotics and Automation (ICRA) (2024) .
  • [24] T. Gneiting, F. Balabdaoui and A.E. Raftery, Probabilistic forecasts, calibration and sharpness, Journal of the Royal Statistical Society: Series B (Statistical Methodology) 69 (2007) 243 [https://rss.onlinelibrary.wiley.com/doi/pdf/10.1111/j.1467-9868.2007.00587.x].
  • [25] D.J. MacKay, Information theory, inference and learning algorithms, Cambridge university press (2003).
  • [26] DeepLearnPhysics Collaboration collaboration, Clustering of electromagnetic showers and particle interactions with graph neural networks in liquid argon time projection chambers, Phys. Rev. D 104 (2021) 072004.