L
Lijun Gong
Researcher at Tencent
Publications - 17
Citations - 1578
Lijun Gong is an academic researcher from Tencent. The author has contributed to research in topics: Convolutional neural network & Feature extraction. The author has an hindex of 8, co-authored 17 publications receiving 1195 citations. Previous affiliations of Lijun Gong include City University of Hong Kong.
Papers
More filters
Proceedings ArticleDOI
VITAL: VIsual Tracking via Adversarial Learning
Yibing Song,Chao Ma,Xiaohe Wu,Lijun Gong,Linchao Bao,Wangmeng Zuo,Chunhua Shen,Rynson W. H. Lau,Ming-Hsuan Yang +8 more
TL;DR: Zhang et al. as mentioned in this paper used a generative network to randomly generate masks, which are applied to adaptively dropout input features to capture a variety of appearance changes, and the network identifies the mask that maintains the most robust features of the target objects over a long temporal span.
Proceedings ArticleDOI
CREST: Convolutional Residual Learning for Visual Tracking
TL;DR: This paper proposes the CREST algorithm to reformulate DCFs as a one-layer convolutional neural network, and applies residual learning to take appearance changes into account to reduce model degradation during online update.
Posted Content
CREST: Convolutional Residual Learning for Visual Tracking
TL;DR: In this paper, the discriminative correlation filters (DCFs) are reformulated as a one-layer convolutional neural network for an end-to-end training.
Journal ArticleDOI
Joint Face Hallucination and Deblurring via Structure Generation and Detail Enhancement
Yibing Song,Jiawei Zhang,Lijun Gong,Shengfeng He,Linchao Bao,Jinshan Pan,Qingxiong Yang,Ming-Hsuan Yang +7 more
TL;DR: This paper proposes a facial component guided deep Convolutional Neural Network to restore a coarse face image, which is denoted as the base image where the facial component is automatically generated from the input face image.
Posted Content
VITAL: VIsual Tracking via Adversarial Learning
Yibing Song,Chao Ma,Xiaohe Wu,Lijun Gong,Linchao Bao,Wangmeng Zuo,Chunhua Shen,Rynson W. H. Lau,Ming-Hsuan Yang +8 more
TL;DR: Zhang et al. as discussed by the authors used a generative network to randomly generate masks, which are applied to adaptively dropout input features to capture a variety of appearance changes, and the network identifies the mask that maintains the most robust features of the target objects over a long temporal span.