J
Joohong Lee
Researcher at Hanyang University
Publications - 9
Citations - 161
Joohong Lee is an academic researcher from Hanyang University. The author has contributed to research in topics: Social anxiety & Tokenization (data security). The author has an hindex of 3, co-authored 9 publications receiving 83 citations.
Papers
More filters
Journal ArticleDOI
Semantic Relation Classification via Bidirectional LSTM Networks with Entity-Aware Attention Using Latent Entity Typing
TL;DR: This article proposed an end-to-end recurrent neural model that incorporates an entity-aware attention mechanism with a latent entity typing (LET) method, which not only effectively utilizes entities and their latent types as features, but also builds word representations by applying self-attention based on symmetrical similarity of a sentence itself.
Posted Content
Semantic Relation Classification via Bidirectional LSTM Networks with Entity-aware Attention using Latent Entity Typing.
TL;DR: This work proposes a novel end-to-end recurrent neural model which incorporates an entity-aware attention mechanism with a latent entity typing (LET) method and demonstrates that the model outperforms existing state-of-the-art models without any high-level features.
Posted Content
An Empirical Study of Tokenization Strategies for Various Korean NLP Tasks
TL;DR: Experimental results demonstrate that a hybrid approach of morphological segmentation followed by BPE works best in Korean to/from English machine translation and natural language understanding tasks such as KorNLI, KorSTS, NSMC, and PAWS-X.
Posted Content
KLUE: Korean Language Understanding Evaluation.
Sungjoon Park,Jihyung Moon,Sungdong Kim,Won Ik Cho,Jiyoon Han,Jang-Won Park,Chisung Song,Junseong Kim,Yongsook Song,Tae-Hwan Oh,Joohong Lee,Juhyun Oh,Sungwon Lyu,Younghoon Jeong,Inkwon Lee,Sangwoo Seo,Dongjun Lee,Hyunwoo Kim,Myeonghwa Lee,Seongbo Jang,Seungwon Do,Sunkyoung Kim,KyungTae Lim,Jongwon Lee,Kyumin Park,Jamin Shin,Seonghyun Kim,Lucy Park,Alice Oh,Jung-Woo Ha,Kyunghyun Cho +30 more
TL;DR: The Korean Language Understanding Evaluation (KLUE) benchmark as mentioned in this paper is a collection of 8 Korean NLP tasks, including Topic Classification, SemanticTextual Similarity, Natural Language Inference, Named Entity Recognition, Relation Extraction, Dependency Parsing, Machine Reading Comprehension, and Dialogue State Tracking.
Proceedings Article
An Empirical Study of Tokenization Strategies for Various Korean NLP Tasks.
TL;DR: In this article, a hybrid approach of morphological segmentation followed by Byte Pair Encoding (BPE) is proposed for Korean NLP tasks, and the results show that BPE segmentation is the most effective.