Since 18 of December 2019 conferences.iaea.org uses Nucleus credentials. Visit our help pages for information on how to Register and Sign-in using Nucleus.

5–8 Nov 2018
Vienna International Centre
CET timezone

Using deep machine learning to conduct object-based identification and motion detection on safeguards video surveillance

Not scheduled
7m
M Building (Vienna International Centre)

M Building

Vienna International Centre

Panelist (Panel Session) Leveraging technological advancements for safeguards applications (TEC) [TEC] Collection, Processing and Analysis of Surveillance Data

Speaker

Dr Yonggang Cui (Brookhaven National Laboratory)

Description

Video surveillance is one of the core monitoring technologies used by the IAEA Department of Safeguards at safeguarded nuclear facilities worldwide. Current IAEA image-review software has functions of scene-change detection, black image detection and missing scene analysis, but their capabilities are not optimum. Because of this limitation, detection of safeguards relevant events heavily depends on inspectors’ visual examination of surveillance videos, which is a time-consuming process prone to errors. To improve the process, we are developing a deep machine learning technique to detect objects of interest in video streams and conduct object-based motion detection. We hypothesize that deep machine learning is effective in solving problems with multi-dimensional features/parameters such as different geometries of casks and fuel assemblies and different configuration settings from one facility to another. A trained deep machine learning algorithm can process video streams automatically to flag, locate, and identify objects of interest in the images. The initial focus of our research is for application at nuclear reactors, such as pressurized heavy-water reactors, where video surveillance is broadly deployed. In a proof-of-concept experiment, we trained and tested a computationally efficient convolutional neural network (CNN) – You Only Look Once (YOLO) – with data collected at a test bed at Brookhaven National Laboratory (BNL). This quick study showed promised for high precision and real time identification of target objects and sequences for image resolutions comparable to those of IAEA’s surveillance system. Currently we are tuning our model with representative training data sets, which are being collected at simulated nuclear facilities at Sandia National Laboratories (SNL) and BNL. In this paper, we will discuss the development of the CNN model and report the results of this study in detail. Our tuned model and algorithm could eventually be integrated with IAEA image review software to significantly reduce the inspector image review burden.

Topics TEC3
Which "Key Question" does your Abstract address? TEC3.1

Primary author

Dr Yonggang Cui (Brookhaven National Laboratory)

Co-authors

Dr Heidi Smartt (Sandia National Laboratories) Dr Maikael Thomas (Sandia National Laboratories) Dr Michael Reed Smith (Sandia National Laboratories) Dr Shinjae Yoo (Brookhaven National Laboratory) Mr Warren Stern (Brookhaven National Laboratory) Dr Yihui Ren (Brookhaven National Laboratory) Dr Yuewei Lin (Brookhaven National Laboratory) Ms Zoe Gastelum (Sandia National Laboratories)

Presentation materials