Abstract
This report presents a series of recommendations for data to train and evaluate radiation detection algorithms and performance metrics to evaluate these algorithms. These recommendations were formed through a community consensus approach through the Detection Radiation Algorithms Group (DRAG), a multi-institution collaboration spanning eight Department of Energy laboratories and John Hopkins Applied Physics Laboratory. This report includes recommendations on background data variability, and metrics to quantify variability, sources and shielding configurations to include in data collection campaigns and detector response variability. In addition, this report describes several anomaly detection and identification algorithms and recommends metrics to report their performance. Finally, this report ends with a discussion on machine learning algorithms.