scispace - formally typeset
Search or ask a question

Showing papers on "Malayalam published in 2020"


Proceedings ArticleDOI
16 Dec 2020
TL;DR: The HASOC track as mentioned in this paper is dedicated to evaluate technology for finding offensive language and hate speech, which has attracted much interest and over 40 research groups have participated as well as described their approaches in papers.
Abstract: This paper presents the HASOC track and its two parts. HASOC is dedicated to evaluate technology for finding Offensive Language and Hate Speech. HASOC is creating test collections for languages with few resources and English for comparison. The first track within HASOC has continued work from 2019 and provided a testbed of Twitter posts for Hindi, German and English. The second track within HASOC has created test resources for Tamil and Malayalam in native and Latin script. Posts were extracted mainly from Youtube and Twitter. Both tracks have attracted much interest and over 40 research groups have participated as well as described their approaches in papers. In this overview, we present the tasks, the data and the main results.

127 citations


Proceedings ArticleDOI
16 Dec 2020
TL;DR: The Dravidian-CodeMix-FIRE 2020 Track as discussed by the authors focused on sentiment analysis of code-mixed text in code mixed text for Tamil and Malayalam, and participants were given a dataset of YouTube comments and the goal of the shared task submissions was to recognise the sentiment of each comment by classifying them into positive, negative, neutral, mixed-feeling classes or by recognizing whether the comment is not in the intended language.
Abstract: Sentiment analysis of Dravidian languages has received attention in recent years However, most social media text is code-mixed and there is no research available on sentiment analysis of code-mixed Dravidian languages The Dravidian-CodeMix-FIRE 2020, a track on Sentiment Analysis for Dravidian Languages in Code-Mixed Text, focused on creating a platform for researchers to come together and investigate the problem There were two languages for this track: (i) Tamil, and (ii) Malayalam The participants were given a dataset of YouTube comments and the goal of the shared task submissions was to recognise the sentiment of each comment by classifying them into positive, negative, neutral, mixed-feeling classes or by recognising whether the comment is not in the intended language The performance of the systems was evaluated by weighted-F1 score

87 citations


Proceedings Article
01 May 2020
TL;DR: Free high quality multi-speaker speech corpora for Gujarati, Kannada, Malayalam, Marathi, Tamil and Telugu, which are six of the twenty two official languages of India spoken by 374 million native speakers are presented.
Abstract: We present free high quality multi-speaker speech corpora for Gujarati, Kannada, Malayalam, Marathi, Tamil and Telugu, which are six of the twenty two official languages of India spoken by 374 million native speakers. The datasets are primarily intended for use in text-to-speech (TTS) applications, such as constructing multilingual voices or being used for speaker or language adaptation. Most of the corpora (apart from Marathi, which is a female-only database) consist of at least 2,000 recorded lines from female and male native speakers of the language. We present the methodological details behind corpora acquisition, which can be scaled to acquiring data for other languages of interest. We describe the experiments in building a multilingual text-to-speech model that is constructed by combining our corpora. Our results indicate that using these corpora results in good quality voices, with Mean Opinion Scores (MOS) > 3.6, for all the languages tested. We believe that these resources, released with an open-source license, and the described methodology will help in the progress of speech applications for the languages described and aid corpora development for other, smaller, languages of India and beyond.

34 citations


Proceedings ArticleDOI
01 Jul 2020
TL;DR: This work proposes a technique called Unified Transliteration and Subword Segmentation to leverage language similarity while exploiting parallel data from related language pairs and proposes a Multilingual Transfer Learning technique to leverage parallelData from multiple related languages to assist translation for low resource language pair of interest.
Abstract: A large percentage of the world’s population speaks a language of the Indian subcontinent, comprising languages from both Indo-Aryan (eg Hindi, Punjabi, Gujarati, etc) and Dravidian (eg Tamil, Telugu, Malayalam, etc) families A universal characteristic of Indian languages is their complex morphology, which, when combined with the general lack of sufficient quantities of high-quality parallel data, can make developing machine translation (MT) systems for these languages difficult Neural Machine Translation (NMT) is a rapidly advancing MT paradigm and has shown promising results for many language pairs, especially in large training data scenarios Since the condition of large parallel corpora is not met for Indian-English language pairs, we present our efforts towards building efficient NMT systems between Indian languages (specifically Indo-Aryan languages) and English via efficiently exploiting parallel data from the related languages We propose a technique called Unified Transliteration and Subword Segmentation to leverage language similarity while exploiting parallel data from related language pairs We also propose a Multilingual Transfer Learning technique to leverage parallel data from multiple related languages to assist translation for low resource language pair of interest Our experiments demonstrate an overall average improvement of 5 BLEU points over the standard Transformer-based NMT baselines

28 citations


Proceedings Article
01 May 2020
TL;DR: This paper uses various existing approaches to create multiple word embeddings for 14 Indian languages using both contextual and non-contextual approaches, and releases a total of 436 models using 8 different approaches.
Abstract: Dense word vectors or ‘word embeddings’ which encode semantic properties of words, have now become integral to NLP tasks like Machine Translation (MT), Question Answering (QA), Word Sense Disambiguation (WSD), and Information Retrieval (IR). In this paper, we use various existing approaches to create multiple word embeddings for 14 Indian languages. We place these embeddings for all these languages, viz., Assamese, Bengali, Gujarati, Hindi, Kannada, Konkani, Malayalam, Marathi, Nepali, Odiya, Punjabi, Sanskrit, Tamil, and Telugu in a single repository. Relatively newer approaches that emphasize catering to context (BERT, ELMo, etc.) have shown significant improvements, but require a large amount of resources to generate usable models. We release pre-trained embeddings generated using both contextual and non-contextual approaches. We also use MUSE and XLM to train cross-lingual embeddings for all pairs of the aforementioned languages. To show the efficacy of our embeddings, we evaluate our embedding models on XPOS, UPOS and NER tasks for all these languages. We release a total of 436 models using 8 different approaches. We hope they are useful for the resource-constrained Indian language NLP. The title of this paper refers to the famous novel “A Passage to India” by E.M. Forster, published initially in 1924.

21 citations


Posted Content
TL;DR: The methods of constructing sentence aligned parallel corpora using tools enabled by recent advances in machine translation and cross-lingual retrieval using deep neural network based methods are reported on.
Abstract: We present sentence aligned parallel corpora across 10 Indian Languages - Hindi, Telugu, Tamil, Malayalam, Gujarati, Urdu, Bengali, Oriya, Marathi, Punjabi, and English - many of which are categorized as low resource. The corpora are compiled from online sources which have content shared across languages. The corpora presented significantly extends present resources that are either not large enough or are restricted to a specific domain (such as health). We also provide a separate test corpus compiled from an independent online source that can be independently used for validating the performance in 10 Indian languages. Alongside, we report on the methods of constructing such corpora using tools enabled by recent advances in machine translation and cross-lingual retrieval using deep neural network based methods.

17 citations


Book ChapterDOI
01 Jan 2020
TL;DR: The present article discusses the use of dynamic modes from dynamic mode decomposition (DMD) method with random mapping for sentiment classification, and observes that the proposed approach provides competing result.
Abstract: Sentiment analysis (SA) or polarity identification is a research topic which receives considerable number of attention. The work in this research attempts to explore the sentiments or opinions in text data related to any event, politics, movies, product reviews, sports, etc. The present article discusses the use of dynamic modes from dynamic mode decomposition (DMD) method with random mapping for sentiment classification. Random mapping is performed using random kitchen sink (RKS) method. The present work aims to explore the use of dynamic modes as the feature for sentiment classification task. In order to conduct the experiment and analysis, the dataset used consists of tweets from SAIL 2015 shared task (tweets in Tamil, Bengali, Hindi) and Malayalam languages. The dataset for Malayalam is prepared by us for the work. The evaluations are performed using accuracy, F1-score, recall, and precision. It is observed from the evaluations that the proposed approach provides competing result.

17 citations


Proceedings Article
31 May 2020
TL;DR: This article presented sentence aligned parallel corpora across 10 Indian Languages - Hindi, Telugu, Tamil, Malayalam, Gujarati, Urdu, Bengali, Oriya, Marathi, Punjabi, and English - many of which are categorized as low resource.
Abstract: We present sentence aligned parallel corpora across 10 Indian Languages - Hindi, Telugu, Tamil, Malayalam, Gujarati, Urdu, Bengali, Oriya, Marathi, Punjabi, and English - many of which are categorized as low resource. The corpora are compiled from online sources which have content shared across languages. The corpora presented significantly extends present resources that are either not large enough or are restricted to a specific domain (such as health). We also provide a separate test corpus compiled from an independent online source that can be independently used for validating the performance in 10 Indian languages. Alongside, we report on the methods of constructing such corpora using tools enabled by recent advances in machine translation and cross-lingual retrieval using deep neural network based methods.

15 citations


Journal ArticleDOI
TL;DR: Experiments conducted on real datasets show that the proposed deep learning-based approach for parts-of-speech tagging for the Malayalam language outperforms some of the already available methods in terms of precision and accuracy.
Abstract: Parts-of-speech tagging is a process in linguistics which deals with tagging each word in a sentence with their corresponding parts-of-speech This process is considered to be one of the pre-processing steps for many natural language processing tasks Earlier approaches were based on simple heuristics and later several methods were reported in the literature that incorporated machine learning techniques such as artificial neural networks Very recently, with the advancement of deep learning-based approaches, parts-of-speech tagging process became more accurate and a reasonable number of taggers are now available for high resource languages such as English But the low resource languages such as Malayalam is still lacking computationally efficient and accurate methods and techniques for parts-of-speech tagging In this direction, this work proposes a deep learning-based approach for parts-of-speech tagging for the Malayalam language Experiments conducted on real datasets show that the proposed method outperforms some of the already available methods in terms of precision and accuracy

12 citations


01 Jan 2020
TL;DR: In this paper, the WLV-RIT entry to the Hate Speech and Offensive Content Identification in Indo-European Languages (HASOC) shared task 2020 was described. And they achieved 0.89 weighted average F1 score for the test set and ranked 5th place out of 12 participants.
Abstract: This paper describes the WLV-RIT entry to the Hate Speech and Offensive Content Identification in Indo-European Languages (HASOC) shared task 2020. The HASOC 2020 organizers provided participants with annotated datasets containing social media posts of code-mixed in Dravidian languages (Malayalam-English and Tamil-English). We participated in task 1: Offensive comment identification in Code-mixed Malayalam Youtube comments. In our methodology, we take advantage of available English data by applying cross-lingual contextual word embeddings and transfer learning to make predictions to Malayalam data. We further improve the results using various fine tuning strategies. Our system achieved 0.89 weighted average F1 score for the test set and it ranked 5th place out of 12 participants.

10 citations


Journal ArticleDOI
TL;DR: The authors examines the relations between trade, faith, and textual traditions in early modern Indian Ocean region and the birth of Arabi-Malayalam, a new system of writing which has facilitated the growth of a vernacular Islamic textual tradition in Malabar since the seventeenth century.
Abstract: This article examines the relations between trade, faith, and textual traditions in early modern Indian Ocean region and the birth of Arabi-Malayalam, a new system of writing which has facilitated the growth of a vernacular Islamic textual tradition in Malabar since the seventeenth century. As a transliterated scriptorial-literary tradition, Arabi-Malayalam emerged out of the polyglossic lingual sphere of the Malabar Coast, and remains as one of the important legacies of social and religious interactions in precolonial south Asia. The first part of this article examines the social, epistemic and normative reasons that led to the scriptorial birth of Arabi-Malayalam, moving beyond a handful of Malayalam writings that locate its origin in the social and economic necessities of Arab traders in the early centuries of Islam. The second part looks at the complex relationship between Muslim scribes and their vernacular audience in the aftermath of Portuguese violence and destruction of Calicut—one of the largest Indian Ocean ports before the sixteenth century. This part focuses on Qadi Muhammed bin Abdul Aziz and his Muhiyuddinmala, the first identifiable text in Arabi-Malayalam, examining how the Muhiyuddinmala represents a transition from classical Arabic theological episteme to the vernacular-popular poetic discourse which changed the pietistic behaviour of the Mappila Muslims of Malabar.

01 Jan 2020
TL;DR: This paper describes the system submitted by the team, KBCNMUJAL, for Task 2 of the shared task “Hate Speech and Offensive Content Identification in Indo-European Languages (HASOC)” at Forum for Information Retrieval Evaluation, December 16-20, 2020, Hyderabad, India.
Abstract: This paper describes the system submitted by our team, KBCNMUJAL, for Task 2 of the shared task Hate Speech and Offensive Content Identification in Indo-European Languages (HASOC), at Forum for Information Retrieval Evaluation, December 16-20, 2020, Hyderabad, India. The datasets of two Dravidian languages Viz. Malayalam and Tamil of size 4000 observations, each were shared by the HASOC organizers. These datasets are used to train the machine using different machine learning algorithms, based on classification and regression models. The datasets consist of tweets or YouTube comments with two class labels offensive and not offensive. The machine is trained to classify such social media messages in these two categories. Appropriate n-gram feature sets are extracted to learn the specific characteristics of the Hate Speech text messages. These feature models are based on TFIDF weights of n-gram. The referred work and respective experiments show that the features such as word, character and combined model of word and character n-grams could be used to identify the term patterns of offensive text contents. As a part of the HASOC shared task, the test data sets are made available by the HASOC track organizers. The best performing classification models developed for both languages are applied on test datasets. The model which gives the highest accuracy result on training dataset for Malayalam language was experimented to predict the categories of respective test data. This system has obtained an F1 score of 0.77. Similarly the best performing model for Tamil language has obtained an F1 score of 0.87. This work has received 2nd and 3rd rank in this shared Task 2 for Malayalam and Tamil language respectively. The proposed system is named HASOC_kbcnmujal.

Journal ArticleDOI
TL;DR: This paper proposes image-based features for speech signal classification because it is possible to identify different patterns by visualizing their speech patterns and the highest accuracy of 94.51% was obtained.
Abstract: Like other applications, under the purview of pattern classification, analyzing speech signals is crucial. People often mix different languages while talking which makes this task complicated. This happens mostly in India, since different languages are used from one state to another. Among many, Southern part of India suffers a lot from this situation, where distinguishing their languages is important. In this paper, we propose image-based features for speech signal classification because it is possible to identify different patterns by visualizing their speech patterns. Modified Mel frequency cepstral coefficient (MFCC) features namely MFCC- Statistics Grade (MFCC-SG) were extracted which were visualized by plotting techniques and thereafter fed to a convolutional neural network. In this study, we used the top 4 languages namely Telugu, Tamil, Malayalam, and Kannada. Experiments were performed on more than 900 hours of data collected from YouTube leading to over 150000 images and the highest accuracy of 94.51% was obtained.

Proceedings ArticleDOI
25 Oct 2020
TL;DR: A G2P system which can handle code-switching developed based onMalayalam-English code-switch speech and text corpora is presented and the overlapping phonemes for English – Malayalam are identified and analysed.
Abstract: Grapheme to phoneme conversion is an integral aspect of speech processing. Conversational speech in Malayalam – a low resource Indic language has inter-sentential, intrasentential code-switching as well as frequent intra-word codeswitching with English. Monolingual G2P systems cannot process such special intra-word code-switching scenarios. A G2P system which can handle code-switching developed based on Malayalam-English code-switch speech and text corpora is presented. Since neither Malayalam nor English are phonetic subset of each other, the overlapping phonemes for English – Malayalam are identified and analysed. Additional rules used to handle special cases of Malayalam phonemes and intra-word code-switching in the G2P system is also presented specifically.

Proceedings ArticleDOI
01 Sep 2020
TL;DR: In this work, 6 different Indian languages such as Hindi, Bengali, Gujarati, Malayalam, Tamil and Telugu is worked on, and how BLEU varies with the usage of Word embedding technique have been clearly shown.
Abstract: Machine translation involves the conversion of text from one language to the other language. In the world of web, a huge number of resources are made available in English. Many of the people are not familiar with this global language. Manually transmuting them into native languages such as Hindi (Indian National language) is a tedious task. In such scenarios, automatic machine translation is an efficient approach. In our work, 8 advanced architectures have been experimented and contrasted their efficiencies. Six different Indian languages such as Hindi, Bengali, Gujarati, Malayalam, Tamil and Telugu is worked on. How BLEU varies with the usage of Word embedding technique have been clearly shown.. Encoder to decoder networks are found fine for short sentences. But if the length of the sentence exceeds 20, then attention architecture is suitable. The 4 Layer Bi-directional LSTM is a great choice in these networks to achieve higher BLEU is also observed. In our work, CFILT, UFAL, ILCC datasets have been considered and achieved a BLEU score of 21.97.

Book ChapterDOI
08 Sep 2020
TL;DR: In this work, morphological complexity of Malayalam is quantitatively analyzed on a text corpus containing 8 million words based on the parameters type-token growth rate (TTGR), type-Token ratio (TTR) and moving average type- token ratio (MATTR).
Abstract: This paper presents a quantitative analysis on the morphological complexity of Malayalam language. Malayalam is a Dravidian language spoken in India, predominantly in the state of Kerala with about 38 million native speakers. Malayalam words undergo inflections, derivations and compounding leading to an infinitely extending lexicon. In this work, morphological complexity of Malayalam is quantitatively analyzed on a text corpus containing 8 million words. The analysis is based on the parameters type-token growth rate (TTGR), type-token ratio (TTR) and moving average type-token ratio (MATTR). The values of the parameters obtained in the current study is compared to that of the values of other morphologically complex languages.

Book ChapterDOI
01 Jan 2020
TL;DR: In this article, the authors compared reading difficulties that arise when studying English and Malayalam and found that reading difficulties manifest differently depending on the characteristics of the language being studied, and the errors were classified into multiple categories.
Abstract: Despite the high prevalence of reading disabilities among Indian children, many school teachers are not adept at identifying and assessing these difficulties. Screening tools for reading disabilities are available in English but are unavailable in many Indian languages. Reading disabilities manifest differently depending on the characteristics of the language being studied. This paper compares reading difficulties that arise when studying English and Malayalam. In a previous study, we designed a bilingual screening test in English and Malayalam and tested it with 135 school children in Kerala. In the current study, the screening test was modified in light of the findings from our previous study. We administered our updated bilingual screening test to 25 second grade children, ages 7–8, who were studying at two other schools in Kerala. Student errors were classified into multiple categories. Similarities and differences between errors in English and Malayalam were identified, and the errors that were specific to Malayalam were analyzed in further detail.

Proceedings Article
01 May 2020
TL;DR: A 24 hour text-to-speech corpus for 3 major Indian languages namely Hindi, Malayalam and Bengali is released and a state-of-the-art TTS system for each of these languages is trained.
Abstract: India is a country where several tens of languages are spoken by over a billion strong population. Text-to-speech systems for such languages will thus be extremely beneficial for wide-spread content creation and accessibility. Despite this, the current TTS systems for even the most popular Indian languages fall short of the contemporary state-of-the-art systems for English, Chinese, etc. We believe that one of the major reasons for this is the lack of large, publicly available text-to-speech corpora in these languages that are suitable for training neural text-to-speech systems. To mitigate this, we release a 24 hour text-to-speech corpus for 3 major Indian languages namely Hindi, Malayalam and Bengali. In this work, we also train a state-of-the-art TTS system for each of these languages and report their performances. The collected corpus, code, and trained models are made publicly available.

Book ChapterDOI
01 Jan 2020
TL;DR: It is observed in this work that NMT requires very less amount of data size for training and thus exhibits satisfactory translation for few thousands of training sentences as well.
Abstract: Natural language translation is one of the most difficult tasks being handled by computer scientist community. This is certainly one task in which machine is definitely lagging behind the cognitive powers of human beings. From the initial ages of computer science a lot of approaches have been proposed for solving this task. Statistical machine translation (SMT) is one of the conventional and matured ways of solving the problem of MT. This approach is based on the Bayes rule and requires huge datasets to train the system. SMT performs well on similar grammar structured language pairs. In recent years, neural MT (NMT) has emerged as an alternate way of addressing the same issue. In this approach, we train a neural network on the source and target language pairs and train the system to develop the translation rules. In this work, we explore different configurations for setting up an NMT system for six different Indian languages. We have focused more on Hindi, which is the most widely spoken language, and more datasets are available for this language. We have also deployed the same system on Bangla, Tamil, Telugu, Urdu, and Malayalam as well. We have experimented with eight different architecture combinations of NMT for English to Indian languages and compared our results with conventional MT techniques. We have also observed in this work that NMT requires very less amount of data size for training and thus exhibits satisfactory translation for few thousands of training sentences as well.

Posted Content
TL;DR: This paper describes the WLV-RIT entry to the Hate Speech and Offensive Content Identification in IndoEuropean Languages (HASOC) shared task 2020, and takes advantage of available English data by applying cross-lingual contextual word embeddings and transfer learning to make predictions to Malayalam data.
Abstract: This paper describes the WLV-RIT entry to the Hate Speech and Offensive Content Identification in Indo-European Languages (HASOC) shared task 2020. The HASOC 2020 organizers provided participants with annotated datasets containing social media posts of code-mixed in Dravidian languages (Malayalam-English and Tamil-English). We participated in task 1: Offensive comment identification in Code-mixed Malayalam Youtube comments. In our methodology, we take advantage of available English data by applying cross-lingual contextual word embeddings and transfer learning to make predictions to Malayalam data. We further improve the results using various fine tuning strategies. Our system achieved 0.89 weighted average F1 score for the test set and it ranked 5th place out of 12 participants.

Proceedings Article
13 Dec 2020
TL;DR: In this article, the authors focus on the Dravidian languages, namely Tamil, Telugu, Kannada, and Malayalam, and bring the related languages into a single script.
Abstract: Bilingual lexicons are a vital tool for under-resourced languages and recent state-of-the-art approaches to this leverage pretrained monolingual word embeddings using supervised or semi- supervised approaches. However, these approaches require cross-lingual information such as seed dictionaries to train the model and find a linear transformation between the word embedding spaces. Especially in the case of low-resourced languages, seed dictionaries are not readily available, and as such, these methods produce extremely weak results on these languages. In this work, we focus on the Dravidian languages, namely Tamil, Telugu, Kannada, and Malayalam, which are even more challenging as they are written in unique scripts. To take advantage of orthographic information and cognates in these languages, we bring the related languages into a single script. Previous approaches have used linguistically sub-optimal measures such as the Levenshtein edit distance to detect cognates, whereby we demonstrate that the longest common sub-sequence is linguistically more sound and improves the performance of bilingual lexicon induction. We show that our approach can increase the accuracy of bilingual lexicon induction methods on these languages many times, making bilingual lexicon induction approaches feasible for such under-resourced languages.

Journal ArticleDOI
TL;DR: The challenges in building an efficient NER for one of the south Indian language namely Malayalam are highlighted and the different issues that the authors need to address are presented.

Posted Content
TL;DR: In this paper, the authors investigated the cross-lingual dynamics by clustering visually similar images together and exploring how they move across language barriers, and found that Telugu, Malayalam, Tamil and Kannada languages tend to be dominant in soliciting political images.
Abstract: Social media has been on the vanguard of political information diffusion in the 21st century. Most studies that look into disinformation, political influence and fake-news focus on mainstream social media platforms. This has inevitably made English an important factor in our current understanding of political activity on social media. As a result, there has only been a limited number of studies into a large portion of the world, including the largest, multilingual and multi-cultural democracy: India. In this paper we present our characterisation of a multilingual social network in India called ShareChat. We collect an exhaustive dataset across 72 weeks before and during the Indian general elections of 2019, across 14 languages. We investigate the cross lingual dynamics by clustering visually similar images together, and exploring how they move across language barriers. We find that Telugu, Malayalam, Tamil and Kannada languages tend to be dominant in soliciting political images (often referred to as memes), and posts from Hindi have the largest cross-lingual diffusion across ShareChat (as well as images containing text in English). In the case of images containing text that cross language barriers, we see that language translation is used to widen the accessibility. That said, we find cases where the same image is associated with very different text (and therefore meanings). This initial characterisation paves the way for more advanced pipelines to understand the dynamics of fake and political content in a multi-lingual and non-textual setting.

Journal ArticleDOI
TL;DR: The article discusses the results and issues encountered in developing this word-splitting tool for Malayalam, mainly in the context of improving the alignments between parallel texts that form a core resource in the Machine Translation task.
Abstract: This article explores in depth various sandhi (joining) rules in Kerala’s Malayalam language, which play a vital role in framing of the inflected and agglutinated forms of words and their compounds...

Journal ArticleDOI
TL;DR: This article investigated the home and school literacy practices of the children enrolled in the Malayalam classes at Organization X with a view to understand the impact of these lessons on the community in Malayalee.
Abstract: The Malayalee community, a minority language group in Singapore, lacks institutional support for learning their mother tongue in schools. In school, most Malayalee children opt to study Tamil or Hindi, a Non-Tamil Indian Language (NTIL) in place of Malayalam. Over the years, ground-up initiatives by volunteers have resulted in ad-hoc community-run classes which are conducted by volunteers. In 2010, a community-run initiative, Organization X, was set up to formalize the learning and teaching of Malayalam in Singapore. This paper aims to investigate the home and school literacy practices of the children enrolled in the Malayalam classes at Organization X with a view to understand the impact of these lessons on the community. The study found that when both home and school literacy practices were viewed as social activities, it contributed to the maintenance of the language in the community.

Journal ArticleDOI
TL;DR: The authors examined how the film industry's apprenticeship and unpaid labour arrangements affect below-the-line labour and less influential job profiles on a film set, and explored how labour and bargaining rights are conceptualized differently by film organizations based on their ideological positions.
Abstract: Labour discourses in the film industry are often couched in the language of ‘welfare’ and an effort to maintain harmony among different filmmaking sectors. But such arrangements do not proffer equal participation or bargaining rights to everyone in the industry. Focusing on the Malayalam language film industry based in Kerala, this article examines how the film industry’s apprenticeship and unpaid labour arrangements affect below-the-line labour and less influential job profiles on a film set. In corollary, I also explore how labour and bargaining rights are conceptualized differently by film organizations based on their ideological positions. Using a mixed-methods approach, including media ethnography and interviews with members of different trade guilds who form part of Malayalam cinema’s professional, technical and service sectors, I demonstrate how structural inequalities in the film industry are overlooked while the cine-worker’s agency is co-opted by a neoliberal system that masquerades as welfare.

Journal ArticleDOI
TL;DR: In this paper, the formations of masculinity and its links to the perceived vulnerability of young women's sexualized bodies by focusing on popular Malayalam cinema from Kerala are explored. But, their focus was on women.
Abstract: This paper explores the formations of masculinity and its links to the perceived vulnerability of young women’s sexualized bodies by focusing on popular Malayalam cinema from Kerala. I will specifi...

Journal ArticleDOI
22 Jul 2020-Glossa
TL;DR: This paper used a dichotic listening task with fluent Malayalam-English bilinguals, in which they were presented with synchronized nonce words, one in each language in separate ears, with competing onsets of a labial stop (Malayalam) and an English, both voiced or both voiceless.
Abstract: Listeners often experience cocktail-party situations, encountering multiple ongoing conversations while tracking just one. Capturing the words spoken under such conditions requires selective attention and processing, which involves using phonetic details to discern phonological structure. How do bilinguals accomplish this in L1-L2 competition? We addressed that question using a dichotic listening task with fluent Malayalam-English bilinguals, in which they were presented with synchronized nonce words, one in each language in separate ears, with competing onsets of a labial stop (Malayalam) and a labial fricative (English), both voiced or both voiceless. They were required to attend to the Malayalam or the English item, in separate blocks, and report the initial consonant they heard. We found that perceptual intrusions from the unattended to the attended language were influenced by voicing, with more intrusions on voiced than voiceless trials. This result supports our proposal for the feature specification of consonants in Malayalam-English bilinguals, which makes use of privative features, underspecification and the “standard approach” to laryngeal features, as against “laryngeal realism”. Given this representational account, we observe that intrusions result from phonetic properties in the unattended signal being assimilated to the closest matching phonological category in the attended language, and are more likely for segments with a greater number of phonological feature specifications.

Proceedings ArticleDOI
01 Dec 2020
TL;DR: The use of cross-lingual word embeddings for detecting cognates among fourteen Indian Languages is demonstrated and the use of context from a knowledge graph to generate improved feature representations for cognate detection is introduced.
Abstract: Cognates are variants of the same lexical form across different languages; for example “fonema” in Spanish and “phoneme” in English are cognates, both of which mean “a unit of sound”. The task of automatic detection of cognates among any two languages can help downstream NLP tasks such as Cross-lingual Information Retrieval, Computational Phylogenetics, and Machine Translation. In this paper, we demonstrate the use of cross-lingual word embeddings for detecting cognates among fourteen Indian Languages. Our approach introduces the use of context from a knowledge graph to generate improved feature representations for cognate detection. We, then, evaluate the impact of our cognate detection mechanism on neural machine translation (NMT), as a downstream task. We evaluate our methods to detect cognates on a challenging dataset of twelve Indian languages, namely, Sanskrit, Hindi, Assamese, Oriya, Kannada, Gujarati, Tamil, Telugu, Punjabi, Bengali, Marathi, and Malayalam. Additionally, we create evaluation datasets for two more Indian languages, Konkani and Nepali. We observe an improvement of up to 18% points, in terms of F-score, for cognate detection. Furthermore, we observe that cognates extracted using our method help improve NMT quality by up to 2.76 BLEU. We also release our code, newly constructed datasets and cross-lingual models publicly.

Journal ArticleDOI
TL;DR: A methodology for deep level tagging of Malayalam text is demonstrated, which is the process of assigning deeper level information to every noun and verb in the text along with normal POS tags.
Abstract: Abstract In recent years, there has been tremendous growth in the amount of natural language text through various sources. Computational analysis of this text has got considerable attention among the NLP researchers. Automatic analysis and representation of natural language text is a step by step procedure. Deep level tagging is one of such steps applied over the text. In this paper, we demonstrate a methodology for deep level tagging of Malayalam text. Deep level tagging is the process of assigning deeper level information to every noun and verb in the text along with normal POS tags. In this study, we move towards a direction that is not much explored in the case of Malayalam language. Malayalam is a morphologically rich and agglutinative language. The morphological features of the language are effectively utilized for the computational analysis of Malayalam text. The language level details required for the study are provided by Thunjath Ezhuthachan Malayalam University, Tirur.