scispace - formally typeset
Search or ask a question
Topic

Word embedding

About: Word embedding is a research topic. Over the lifetime, 4683 publications have been published within this topic receiving 153378 citations. The topic is also known as: word embeddings.


Papers
More filters
Journal ArticleDOI
TL;DR: The experimental results show that the model can achieve state-ofthe-art accuracy on student feedback dataset and will be upgrade by increasing the data samples of neutral comments in dataset.
Abstract: Objectives: Teacher’s evaluation in education system is quite important to improve the learning experience ininstitutions. For this purpose, sentiment analysis model is developedto identify the student sentiments from the piece of text. Methods/ Statistical Analysis: Long Short-Term Memory Model (LSTM) is used for analyzing the sentiments expressed by students through textual feedback. For this purpose, dataset has been built through student’s feedback and then divided into 70% and 30% for training and testing. The proposed model has been trained using softmax and adam along with drop out values 0.1 and 0.2. Obtained results showed that our model provides 99%, and 90% accuracy over training and validation with 0.2 and 0.5 losses respectively. Findings: It was found that proposed model provides an efficient way for sentiment analysis for teacher’s evaluation. Model used input as word embedding over the LSTM for mapping the words. Andmoreover, the model is collected significant semantic and syntactic information by implementing pre-trained word vector model. Hence, this model has the prospective to overcome several flaws in traditional methods e.g., bag-of-words, n-gram, Naive Bayes and SVM models where order and information about word is vanished. The experimental results show that the model can achieve state-ofthe-art accuracy on student feedback dataset. Application/Improvements: The study helps for improving the quality of teaching in education system. And moreover,it will be upgrade by increasing the data samples of neutral comments in dataset. Keywords: Course Evaluation, Opinion Mining, Sentiment Analysis, Student’s Feedback, LSTM, RNN

19 citations

Proceedings Article
01 Aug 2018
TL;DR: The knowledge-enriched word embedding (KEWE) is provided, which encodes the knowledge on reading difficulty into the representation of words in the form of a knowledge graph.
Abstract: In this paper, we present a method which learns the word embedding for readability assessment. For the existing word embedding models, they typically focus on the syntactic or semantic relations of words, while ignoring the reading difficulty, thus they may not be suitable for readability assessment. Hence, we provide the knowledge-enriched word embedding (KEWE), which encodes the knowledge on reading difficulty into the representation of words. Specifically, we extract the knowledge on word-level difficulty from three perspectives to construct a knowledge graph, and develop two word embedding models to incorporate the difficulty context derived from the knowledge graph to define the loss functions. Experiments are designed to apply KEWE for readability assessment on both English and Chinese datasets, and the results demonstrate both effectiveness and potential of KEWE.

19 citations

Journal ArticleDOI
TL;DR: High accuracy in ATS was achieved, enabling seamless curation of research evidence from the biomedical literature to use for clinical decision-making, and a highly positive correlation among the evaluations performed by three domain experts concerning different metrics is found, suggesting that the automated summarization is satisfactory.
Abstract: Background: Automatic text summarization (ATS) enables users to retrieve meaningful evidence from big data of biomedical repositories to make complex clinical decisions. Deep neural and recurrent networks outperform traditional machine-learning techniques in areas of natural language processing and computer vision; however, they are yet to be explored in the ATS domain, particularly for medical text summarization. Objective: Traditional approaches in ATS for biomedical text suffer from fundamental issues such as an inability to capture clinical context, quality of evidence, and purpose-driven selection of passages for the summary. We aimed to circumvent these limitations through achieving precise, succinct, and coherent information extraction from credible published biomedical resources, and to construct a simplified summary containing the most informative content that can offer a review particular to clinical needs. Methods: In our proposed approach, we introduce a novel framework, termed Biomed-Summarizer, that provides quality-aware Patient/Problem, Intervention, Comparison, and Outcome (PICO)-based intelligent and context-enabled summarization of biomedical text. Biomed-Summarizer integrates the prognosis quality recognition model with a clinical context–aware model to locate text sequences in the body of a biomedical article for use in the final summary. First, we developed a deep neural network binary classifier for quality recognition to acquire scientifically sound studies and filter out others. Second, we developed a bidirectional long-short term memory recurrent neural network as a clinical context–aware classifier, which was trained on semantically enriched features generated using a word-embedding tokenizer for identification of meaningful sentences representing PICO text sequences. Third, we calculated the similarity between query and PICO text sequences using Jaccard similarity with semantic enrichments, where the semantic enrichments are obtained using medical ontologies. Last, we generated a representative summary from the high-scoring PICO sequences aggregated by study type, publication credibility, and freshness score. Results: Evaluation of the prognosis quality recognition model using a large dataset of biomedical literature related to intracranial aneurysm showed an accuracy of 95.41% (2562/2686) in terms of recognizing quality articles. The clinical context–aware multiclass classifier outperformed the traditional machine-learning algorithms, including support vector machine, gradient boosted tree, linear regression, K-nearest neighbor, and naive Bayes, by achieving 93% (16127/17341) accuracy for classifying five categories: aim, population, intervention, results, and outcome. The semantic similarity algorithm achieved a significant Pearson correlation coefficient of 0.61 (0-1 scale) on a well-known BIOSSES dataset (with 100 pair sentences) after semantic enrichment, representing an improvement of 8.9% over baseline Jaccard similarity. Finally, we found a highly positive correlation among the evaluations performed by three domain experts concerning different metrics, suggesting that the automated summarization is satisfactory. Conclusions: By employing the proposed method Biomed-Summarizer, high accuracy in ATS was achieved, enabling seamless curation of research evidence from the biomedical literature to use for clinical decision-making.

19 citations

Journal ArticleDOI
TL;DR: A novel prototype system that analyzes the emergency-related tweets to classify them as need or available tweets, and a novel two-word sliding window approach is proposed to generate the combine embedding of two adjacent words.
Abstract: Social media has evolved itself as a significant tool used by people for information spread during emergencies like natural or man-made disasters. Real-time analysis of this huge collected data can play a vital role in crisis estimation, response and assistance exercises. We propose a novel prototype system that analyzes the emergency-related tweets to classify them as need or available tweets. The presented system also takes care of non-English tweets as there is no boundary of language for social media users. Several classifiers along with different learning methodologies are used to show their usefulness for an efficient solution. Here, a new supervised learning technique based on word embedding is incorporated in the novel hybrid model that comprises of LSTM and CNN. The system will further give a ranked list of tweets, along with a relevance score for each tweet with respect to the topic. Finally for each of the identified need tweets, its corresponding availability tweets are mapped. For the mapping task, a novel two-word sliding window approach is proposed to generate the combine embedding of two adjacent words. The experimental results show significant improvement in the performance. We evaluate our proposed system with FIRE-2016 and CrisisLex datasets to illustrate its effectiveness during mobilization of needful resources.

19 citations

Journal ArticleDOI
TL;DR: SentiVec is a kernel optimization function system for sentiment word embedding, which is based on two phases, and the optimal sentiment vectors successfully extract the features in terms of semantic and sentiment information, which makes it outperform the baseline methods on word similarity, word analogy, and SA tasks.
Abstract: Deep learning-based sentiment analysis (SA) methods have drawn more attention in recent years, which calls for more precise word embedding methods. This article proposes SentiVec, a kernel optimization function system for sentiment word embedding, which is based on two phases. The first phase is a supervised learning method, and the second phase consists of two unsupervised updating models, object-word-to-surrounding-words reward model (O2SR) and context-to-object-word reward model (C2OR). SentiVec is aimed at: 1) integrating the statistical information and sentiment orientation into sentiment word vectors and 2) propagating and updating the semantic information to all the word representations in a corpus. Extensive experimental results show that the optimal sentiment vectors successfully extract the features in terms of semantic and sentiment information, which makes it outperform the baseline methods on word similarity, word analogy, and SA tasks.

19 citations


Network Information
Related Topics (5)
Recurrent neural network
29.2K papers, 890K citations
87% related
Unsupervised learning
22.7K papers, 1M citations
86% related
Deep learning
79.8K papers, 2.1M citations
85% related
Reinforcement learning
46K papers, 1M citations
84% related
Graph (abstract data type)
69.9K papers, 1.2M citations
84% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
2023317
2022716
2021736
20201,025
20191,078
2018788