D
Daya Guo
Researcher at Sun Yat-sen University
Publications - 37
Citations - 2510
Daya Guo is an academic researcher from Sun Yat-sen University. The author has contributed to research in topics: Computer science & Parsing. The author has an hindex of 13, co-authored 28 publications receiving 661 citations. Previous affiliations of Daya Guo include Harbin Institute of Technology & Microsoft.
Papers
More filters
Posted Content
CodeBERT: A Pre-Trained Model for Programming and Natural Languages
Zhangyin Feng,Daya Guo,Duyu Tang,Nan Duan,Xiaocheng Feng,Ming Gong,Linjun Shou,Bing Qin,Ting Liu,Daxin Jiang,Ming Zhou +10 more
TL;DR: This work develops CodeBERT with Transformer-based neural architecture, and trains it with a hybrid objective function that incorporates the pre-training task of replaced token detection, which is to detect plausible alternatives sampled from generators.
Posted Content
GraphCodeBERT: Pre-training Code Representations with Data Flow
Daya Guo,Shuo Ren,Shuai Lu,Zhangyin Feng,Duyu Tang,Shujie Liu,Long Zhou,Nan Duan,Alexey Svyatkovskiy,Fu Shengyu,Michele Tufano,Shao Kun Deng,Colin B. Clement,Dawn Drain,Neel Sundaresan,Jian Yin,Daxin Jiang,Ming Zhou +17 more
TL;DR: Results show that code structure and newly introduced pre-training tasks can improve GraphCodeBERT and achieves state-of-the-art performance on the four downstream tasks and it is shown that the model prefers structure-level attentions over token- level attentions in the task of code search.
Proceedings ArticleDOI
CodeBERT: A Pre-Trained Model for Programming and Natural Languages
Zhangyin Feng,Daya Guo,Duyu Tang,Nan Duan,Xiaocheng Feng,Ming Gong,Linjun Shou,Bing Qin,Ting Liu,Daxin Jiang,Ming Zhou +10 more
TL;DR: CodeBERT as mentioned in this paper is a pre-trained model for natural language code search and code documentation generation with a hybrid objective function that incorporates the pre-training task of replaced token detection, which is to detect plausible alternatives sampled from generators.
Journal ArticleDOI
Graph-Based Reasoning over Heterogeneous External Knowledge for Commonsense Question Answering
Shangwen Lv,Daya Guo,Jingjing Xu,Duyu Tang,Nan Duan,Ming Gong,Linjun Shou,Daxin Jiang,Guihong Cao,Songlin Hu +9 more
TL;DR: This work proposes to automatically extract evidence from heterogeneous knowledge sources, and answer questions based on the extracted evidence, and achieves the state-of-the-art accuracy on the CommonsenseQA dataset.
Posted Content
CodeBLEU: a Method for Automatic Evaluation of Code Synthesis
Shuo Ren,Daya Guo,Shuai Lu,Long Zhou,Shujie Liu,Duyu Tang,Neel Sundaresan,Ming Zhou,Ambrosio Blanco,Shuai Ma +9 more
TL;DR: This work introduces a new automatic evaluation metric, dubbed CodeBLEU, which absorbs the strength of BLEU in the n-gram match and further injects code syntax via abstract syntax trees (AST) and code semantics via data-flow and can achieve a better correlation with programmer assigned scores compared with BLEu and accuracy.