J
Jiayang Chen
Researcher at The Chinese University of Hong Kong
Publications - 7
Citations - 46
Jiayang Chen is an academic researcher from The Chinese University of Hong Kong. The author has contributed to research in topics: Computer science & Biology. The author has an hindex of 1, co-authored 1 publications receiving 2 citations.
Papers
More filters
Posted ContentDOI
Interpretable RNA Foundation Model from Unannotated Data for Highly Accurate RNA Structure and Function Predictions
Jiayang Chen,Zhihang Hu,Siqi Sun,Qingxiong Tan,Yixuan Wang,Qinze Yu,Licheng Zong,Liang Hong,Jin Xiao,Irwin King,Yu Li +10 more
TL;DR: This work proposes a novel RNA foundation model (RNA-FM) to take advantage of all the 23 million non-coding RNA sequences through self-supervised learning and discovers that the pre-trained RNA-FM could infer sequential and evolutionary information of non-Coding RNAs without using any labels.
Journal ArticleDOI
E2Efold-3D: End-to-End Deep Learning Method for accurate de novo RNA 3D Structure Prediction
Tao Shen,Zhihang Hu,Zhangzhi Peng,Jiayang Chen,Peng Xiong,Liang Hong,Liangzhen Zheng,Yixuan Wang,Irwin King,Sheng Wang,Siqi Sun,Yu Li +11 more
TL;DR: The first end-to-end deep learning approach, E2Efold-3D, is developed to accurately perform the de novo RNA structure prediction and achieves promising results when predicting RNA complex structures, a feat that none of the previous systems could accomplish.
Posted ContentDOI
Self-supervised contrastive learning for integrative single cell RNA-seq data analysis
Wenkai Han,Wenkai Han,Yuqi Cheng,Yuqi Cheng,Jiayang Chen,Huawen Zhong,Zhihang Hu,Siyuan Chen,Licheng Zong,Irwin King,Xin Gao,Yu Li +11 more
TL;DR: Wang et al. as mentioned in this paper presented a self-supervised Contrastive LEArning framework for scRNA-seq (CLEAR) profile representation and downstream analysis, which overcomes the heterogeneity of the experimental data with a specifically designed representation learning task.
Journal ArticleDOI
Contact-Distil: Boosting Low Homologous Protein Contact Map Prediction by Self-Supervised Distillation
TL;DR: Extensive experiments show Contact-Distil outperforms previous state-of-the-arts by large margins on CAMEO-L dataset for low homologous PCMP, i.e., around 13.3% and 9.5% improvements against Alphafold2 and MSA Transformer when MSA count less than 10.
Journal ArticleDOI
Weakly supervised training for eye fundus lesion segmentation in patients with diabetic retinopathy.
TL;DR: It is demonstrated that combining unlabeled medical images with coarse segmentation results can effectively improve the robustness of the lesion segmentation model and a practical framework for improving the performance of medical image segmentation given limited labeled data samples is proposed.