scispace - formally typeset
Search or ask a question
Author

Yiyang Zhang

Bio: Yiyang Zhang is an academic researcher from Beihang University. The author has contributed to research in topics: Engineering & Sketch recognition. The author has an hindex of 1, co-authored 1 publications receiving 9 citations.

Papers
More filters
Journal ArticleDOI
TL;DR: This paper proposes an adaptive template matching-based single object tracking algorithm framework to achieve template update online, based on the Faster-RCNN model, and presents a parallel strategy to accelerate the process of template matching.

15 citations

Proceedings ArticleDOI
18 Nov 2022
TL;DR: In this paper , a basic gesture recognition system built on the existing Google deep learning framework TensorFlow and gesture recognition components in MediaPipe and OpenCv machine vision open-source library is presented.
Abstract: With the development of society, gestures are used in many aspects, but the computer's functionality for gesture recognition is still to be improved. This article is mainly a preliminary idea of a basic gesture recognition system built based on the existing Google deep learning framework TensorFlow and gesture recognition components in MediaPipe and OpenCv machine vision open-source library. The training dataset is first subjected to skeleton key point coordinate extraction, then the pre-processed dataset is used to train the neural network and constitute the preliminary model, and finally the model is corrected and changed in the end.

Cited by
More filters
Journal ArticleDOI
TL;DR: Considering the relationship between the current frame and the previous frame of a moving object target in a time series, a temporal regularization strategy to improve the BACF tracker (denoted as TRBACF), a typical representative of the aforementioned trackers is proposed.

28 citations

Proceedings ArticleDOI
01 Jan 2020
TL;DR: This work presents the experimentation and the performance comparison between the Jetson Nano and Jetson TX2 development kits, when implementing the Template Matching method, in order to get an evaluation criterion to select one of them in image processing projects.
Abstract: Template Matching is a widely used method for object detection in digital images, it requires great processing power since it is an exhaustive method that compares the intensity levels of a source image pixel-to-pixel with a template image that contains the object to identify. Nowadays there are dedicated embedded systems that provide high processing capabilities, such as the NVIDIA Jetson family. This work presents the experimentation and the performance comparison between the Jetson Nano and Jetson TX2 development kits, when implementing the Template Matching method, in order to get an evaluation criterion to select one of them in image processing projects. It was carried out to six images with different sizes and two variants in terms of the size of the template image. The processing times for the sequential implementation using the CPUs and the parallel implementation with the GPUs were obtained quantitatively. It was observed that the processing times using the parallel versions on average doubled those of the sequential versions and that the Jetson TX2 exceeded the Jetson Nano in execution speeds.

27 citations

Journal ArticleDOI
TL;DR: A novel adaptive spatial-temporal regularized correlation filter model to learn an appropriate regularization for achieving robust tracking and a relative peak discriminative method for model updating to avoid the model degradation is proposed.

11 citations

Journal ArticleDOI
TL;DR: A water target dataset containing 9936 images is created from offshore USV experiments in which the human-in-the-loop annotation and mosaic data augmentation methods are used and the integrated USV-based system for water target recognition achieves high recognition capability while maintaining a high degree of robustness.
Abstract: Water target recognition is a critical challenge for the perception technology of unmanned surface vessels (USVs). In the application of USV, detection accuracy and the inference time both matter, while it is tough to strike a balance and single-frame water target detection behaves unstable in the video detection. To solve these problems, many strategies are applied to increase YOLOv4’s performance, including network pruning, the focal loss function, blank label training, and preprocessing with histogram normalization. The optimized detection method achieves a mean average precision (mAP) of 81.74% and a prediction speed of 26.77 frames per second (FPS), which meets the USV navigation requirements. To build the integrated USV-based system for water target recognition, a water target dataset containing 9936 images is created from offshore USV experiments in which the human-in-the-loop annotation and mosaic data augmentation methods are used. The issues of miss detection and false alarm can be considerably mitigated by cascading the Siamese-RPN tracking network, and the major color of a water target can be retrieved using a local contrast saliency color detection scheme. The system being tested is called “ME120” includes an embedded edge computing platform (Nvidia Jetson AGX Xavier). Finally, online dataset learning demonstrates the improved YOLOv4 achieves an increase of 66.98% in FPS at the cost of a decrease of 0.79% in mAP when compared with the original YOLOv4 and offline navigation experiments validate that our system achieves high recognition capability while maintaining a high degree of robustness.

5 citations

Journal ArticleDOI
TL;DR: In this paper , a water target recognition system for underwater USVs is presented, which includes an embedded edge computing platform (Nvidia Jetson AGX Xavier) and an integrated USV-based system.
Abstract: Water target recognition is a critical challenge for the perception technology of unmanned surface vessels (USVs). In the application of USV, detection accuracy and the inference time both matter, while it is tough to strike a balance and single-frame water target detection behaves unstable in the video detection. To solve these problems, many strategies are applied to increase YOLOv4’s performance, including network pruning, the focal loss function, blank label training, and preprocessing with histogram normalization. The optimized detection method achieves a mean average precision (mAP) of 81.74% and a prediction speed of 26.77 frames per second (FPS), which meets the USV navigation requirements. To build the integrated USV-based system for water target recognition, a water target dataset containing 9936 images is created from offshore USV experiments in which the human-in-the-loop annotation and mosaic data augmentation methods are used. The issues of miss detection and false alarm can be considerably mitigated by cascading the Siamese-RPN tracking network, and the major color of a water target can be retrieved using a local contrast saliency color detection scheme. The system being tested is called “ME120” includes an embedded edge computing platform (Nvidia Jetson AGX Xavier). Finally, online dataset learning demonstrates the improved YOLOv4 achieves an increase of 66.98% in FPS at the cost of a decrease of 0.79% in mAP when compared with the original YOLOv4 and offline navigation experiments validate that our system achieves high recognition capability while maintaining a high degree of robustness.

5 citations