UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

Dual-Mode Learning for Multi-Dataset X-Ray Security Image Detection

Yang, Fenghong; Jiang, Runqing; Yan, Yan; Xue, Jing-Hao; Wang, Biao; Wang, Hanzi; (2024) Dual-Mode Learning for Multi-Dataset X-Ray Security Image Detection. IEEE Transactions on Information Forensics and Security 10.1109/tifs.2024.3364368. (In press). Green open access

[thumbnail of FenghongYang-TIFS-2024.pdf]
Preview
PDF
FenghongYang-TIFS-2024.pdf - Accepted Version

Download (4MB) | Preview

Abstract

With the recent advance of deep learning, a large number of methods have been developed for prohibited item detection in X-ray security images. Generally, these methods train models on a single X-ray image dataset that may contain only limited categories of prohibited items. To detect more prohibited items, it is desirable to train a model on the multi-dataset that is constructed by combining multiple datasets. However, directly applying existing methods to the multi-dataset cannot guarantee good performance because of the large domain discrepancy between datasets and the occlusion in images. To address the above problems, we propose a novel Dual-Mode Learning Network (DML-Net) to effectively detect all the prohibited items in the multi-dataset. In particular, we develop an enhanced RetinaNet as the architecture of DML-Net, where we introduce a lattice appearance enhanced sub-net to enhance appearance representations. Such a way benefits the detection of occluded prohibited items. Based on the enhanced RetinaNet, the learning process of DML-Net involves both common mode learning (detecting the common prohibited items across datasets) and unique mode learning (detecting the unique prohibited items in each dataset). For common mode learning, we introduce an adversarial prototype alignment module to align the feature prototypes from different datasets in the domain-invariant feature space. For unique mode learning, we take advantage of feature distillation to enforce the student model to mimic the features extracted by multiple pre-trained teacher models. By tightly combining and jointly training the dual modes, our DML-Net method successfully eliminates the domain discrepancy and exhibits superior model capacity on the multi-dataset. Extensive experimental results on several combined X-ray image datasets demonstrate the effectiveness of our method against several state-of-the-art methods. Our code is available at https://github.com/vampirename/dmlnet.

Type: Article
Title: Dual-Mode Learning for Multi-Dataset X-Ray Security Image Detection
Open access status: An open access version is available from UCL Discovery
DOI: 10.1109/tifs.2024.3364368
Publisher version: http://dx.doi.org/10.1109/tifs.2024.3364368
Language: English
Additional information: This version is the author accepted manuscript. For information on re-use, please refer to the publisher's terms and conditions.
UCL classification: UCL
UCL > Provost and Vice Provost Offices > UCL BEAMS
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Maths and Physical Sciences
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Maths and Physical Sciences > Dept of Statistical Science
URI: https://discovery.ucl.ac.uk/id/eprint/10187012
Downloads since deposit
10Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item