Electronics, Vol. 13, Pages 1440: Improved Transformer-Based Deblurring of Commodity Videos in Dynamic Visual Cabinets

2 weeks ago 16

Electronics, Vol. 13, Pages 1440: Improved Transformer-Based Deblurring of Commodity Videos in Dynamic Visual Cabinets

Electronics doi: 10.3390/electronics13081440

Authors: Shuangyi Huang Qianjie Liang Kai Xie Zhengfang He Chang Wen Jianbiao He Wei Zhang

In the dynamic visual cabinet, the occurrence of motion blur when consumers take out commodities will reduce the accuracy of commodity detection. Recently, although Transformer-based video deblurring networks have achieved results compared to Convolutional Neural Networks in some blurring scenarios, they are still challenging for the non-uniform blurring problem that occurs when consumers pick up the commodities, such as the problem of difficult alignment of blurred video frames of small commodities and the problem of underutilizing the effective information between the video frames of commodities. Therefore, an improved Transformer video deblurring network is proposed. Firstly, a multi-scale Transformer feature extraction method is utilized for non-uniform blurring. Secondly, for the problem of difficult alignment of small-item-blurred video frames, a temporal interactive attention mechanism is designed for video frame alignment. Finally, a feature recurrent fusion mechanism is introduced to supplement the effective information of commodity features. The experimental results show that the proposed method has practical significance in improving the accuracy of commodity detection. Moreover, compared with the recent Transformer deblurring algorithm Video Restoration Transformer, the Peak Signal-to-Noise Ratio of this paper’s algorithm is higher than that of the Deep Video Deblurring dataset and the Fuzzy Commodity Dataset by 0.23 dB and 0.81 dB, respectively.

Read Entire Article