| Literature DB >> 36247971 |
Blair Jamieson1, Matt Stubbs2, Sheela Ramanna2, John Walker1,3, Nick Prouse3, Ryosuke Akutsu3, Patrick de Perio3,4, Wojciech Fedorko3.
Abstract
Water Cherenkov detectors like Super-Kamiokande, and the next generation Hyper-Kamiokande are adding gadolinium to their water to improve the detection of neutrons. By detecting neutrons in addition to the leptons in neutrino interactions, an improved separation between neutrino and anti-neutrinos, and reduced backgrounds for proton decay searches can be expected. The neutron signal itself is still small and can be confused with muon spallation and other background sources. In this paper, machine learning techniques are employed to optimize the neutron capture detection capability in the new intermediate water Cherenkov detector (IWCD) for Hyper-K. In particular, boosted decision tree (XGBoost), graph convolutional network (GCN), and dynamic graph convolutional neural network (DGCNN) models are developed and benchmarked against a statistical likelihood-based approach, achieving up to a 10% increase in classification accuracy. Characteristic features are also engineered from the datasets and analyzed using SHAP (SHapley Additive exPlanations) to provide insight into the pivotal factors influencing event type outcomes. The dataset used in this research consisted of roughly 1.6 million simulated particle gun events, divided nearly evenly between neutron capture and a background electron source. The current samples used for training are representative only, and more realistic samples will need to be made for the analyses of real data. The current class split is 50/50, but there is expected to be a difference between the classes in the real experiment, and one might consider using resampling techniques to address the issue of serious imbalances in the class distribution in real data if necessary.Entities:
Keywords: graph neural networks; machine learning; neutrino physics; particle physics; water Cherenkov detector
Year: 2022 PMID: 36247971 PMCID: PMC9561466 DOI: 10.3389/fdata.2022.978857
Source DB: PubMed Journal: Front Big Data ISSN: 2624-909X
Figure 1Unrolled cylinder event displays showing the charged deposit in units of photoelectrons as the colored points for a sample electron background event. Multi PMT modules without any charge are shown in yellow.
Figure 2(A–L) Comparison of 12 engineered features separated by neutron capture and spallation electron background events. The data consists of nearly 1.6 million events, generated by WCSim for the IWCD detector geometry.
Figure 3Confusion matrix for the XGBoost model trained on the dataset of neutron capture and spallation background electron events.
Figure 4Beeswarm plot of SHAP values for the neutron capture and spallation background dataset, simulated using WCSim for the IWCD tank geometry. The SHAP value for each feature in every event is plotted as a dot in the plot, where the x-axis position corresponds to the SHAP value and the colorbar shows the feature value (blue is low, red is high). High SHAP values influence the model output toward 1 (electron-like event) and low SHAP values (negative) influence the model outputs toward 0 (neutron-like event).
GCN model applied to padded and non-padded, fully connected, uniformly edge weighted, and inverse square distance (1/d2) weighted graphs for the simulated IWCD neutron capture and background datasets.
|
|
|
|
|
|
|---|---|---|---|---|
| Padded | 61.3 | 63.1 | 63.1 | 0.667 |
| Non-padded | 58.5 | 59.8 | 59.9 | 0.628 |
| Padded (1/ | 59.7 | 61.3 | 61.4 | 0.632 |
Figure 5Applied DGCNN architecture for neutron capture and electron background event discrimination. Two dynamic edge convolutional blocks were applied, followed by a fully connected layer, global max pooling, and a final multi-layer perceptron layer.
DGCNN model classification accuracies for variations of the number of nearest neighbors k in the DGCNN dynamic edge convolution blocks from 10 to 30 in increments of 5.
|
|
|
|
|
|
|---|---|---|---|---|
| 10 | 70.9 | 72.0 | 71.9 | 0.792 |
| 15 | 71.8 | 72.2 | 72.3 | 0.796 |
| 20 | 71.7 | 72.3 | 72.3 | 0.796 |
| 25 | 71.8 | 72.4 | 72.4 | 0.797 |
| 30 | 71.4 | 72.4 | 72.4 | 0.796 |
Comparison of training times for the different models applied in this study, sorted in ascending order by training time per epoch.
|
|
|
|
|
|---|---|---|---|
| XGBoost | 1,450 | 50 | 0.036 |
| DGCNN (k = 10) | 25 | 1,980 | 1.32 |
| DGCNN (k = 15) | 25 | 2,100 | 1.4 |
| DGCNN (k = 20) | 25 | 2,700 | 1.8 |
| DGCNN (k = 25) | 25 | 3,420 | 2.28 |
| DGCNN (k = 30) | 25 | 3,960 | 2.64 |
| GCN (non-padded) | 75 | 1,020 | 13.6 |
| Likelihood Ratio | 1 | 80 | 80 |
| GCN (padded) | 5 | 1,800 | 360 |
Overall accuracies for neutron capture vs. electron background classification for the likelihood analysis (Likelihood), XGBoost, GCN, and DGCNN methods.
|
|
|
|
|
|
|---|---|---|---|---|
| Spallation | 62.5 | 71.4 | 63.1 |
|
Figure 6Comparison of the ROC curves for the XGBoost, GCN, and DGCNN results presented in this paper.