Cross-Lingual Word Embeddings

Cross-Lingual Word Embeddings
Author: Anders Søgaard
Publisher: Springer Nature
Total Pages: 120
Release: 2022-05-31
Genre: Computers
ISBN: 3031021711

Download Cross-Lingual Word Embeddings Book in PDF, Epub and Kindle

The majority of natural language processing (NLP) is English language processing, and while there is good language technology support for (standard varieties of) English, support for Albanian, Burmese, or Cebuano--and most other languages--remains limited. Being able to bridge this digital divide is important for scientific and democratic reasons but also represents an enormous growth potential. A key challenge for this to happen is learning to align basic meaning-bearing units of different languages. In this book, the authors survey and discuss recent and historical work on supervised and unsupervised learning of such alignments. Specifically, the book focuses on so-called cross-lingual word embeddings. The survey is intended to be systematic, using consistent notation and putting the available methods on comparable form, making it easy to compare wildly different approaches. In so doing, the authors establish previously unreported relations between these methods and are able to present a fast-growing literature in a very compact way. Furthermore, the authors discuss how best to evaluate cross-lingual word embedding methods and survey the resources available for students and researchers interested in this topic.

Cross-Lingual Word Embeddings with Universal Concepts and Their Applications

Cross-Lingual Word Embeddings with Universal Concepts and Their Applications
Author: Pezhman Sheinidashtegol
Publisher:
Total Pages:
Release: 2020
Genre: Electronic dissertations
ISBN:

Download Cross-Lingual Word Embeddings with Universal Concepts and Their Applications Book in PDF, Epub and Kindle

Enormous amounts of data are generated in many languages every day due to our increasing global connectivity. This increases the demand for the ability to read and classify data regardless of language. Word embedding is a popular Natural Language Processing (NLP) strategy that uses language modeling and feature learning to map words to vectors of real numbers. However, these models need a significant amount of data annotated for the training. While gradually, the availability of labeled data is increasing, most of these data are only available in high resource languages, such as English. Researchers with different sets of proficient languages seek to address new problems with multilingual NLP applications. In this dissertation, I present multiple approaches to generate cross-lingual word embedding (CWE) using universal concepts (UC) amongst languages to address the limitations of existing methods. My work consists of three approaches to build multilingual/bilingual word embeddings. The first approach includes two steps: pre-processing and processing. In the pre-processing step, we build a bilingual corpus containing both languages' knowledge in the form of sentences for the most frequent words in English and their translated pair in the target language. In this step, knowledge of the source language is shared with the target language and vice versa by swapping one word per sentence with its corresponding translation. In the second step, we use a monolingual embeddings estimator to generate the CWE. The second approach generates multilingual word embeddings using UCs. This approach consists of three parts. For part I, we introduce and build UCs using bilingual dictionaries and graph theory by defining words as nodes and translation pairs as edges. In part II, we explain the configuration used for word2vec to generate encoded-word embeddings. Finally, part III includes decoding the generated embeddings using UCs. The final approach utilizes the supervised method of the MUSE project, but, the model trained on our UCs. Finally, we applied our last two proposed methods to some practical NLP applications; document classification, cross-lingual sentiment analysis, and code-switching sentiment analysis. Our proposed methods outperform the state of the art MUSE method on the majority of applications.

Cross-lingual Word Embeddings for Low-resource and Morphologically-rich Languages

Cross-lingual Word Embeddings for Low-resource and Morphologically-rich Languages
Author: Ali Hakimi Parizi
Publisher:
Total Pages: 0
Release: 2021
Genre:
ISBN:

Download Cross-lingual Word Embeddings for Low-resource and Morphologically-rich Languages Book in PDF, Epub and Kindle

Despite recent advances in natural language processing, there is still a gap in state-of-the-art methods to address problems related to low-resource and morphologically-rich languages. These methods are data-hungry, and due to the scarcity of training data for low-resource and morphologically-rich languages, developing NLP tools for them is a challenging task. Approaches for forming cross-lingual embeddings and transferring knowledge from a rich- to a low-resource language have emerged to overcome the lack of training data. Although in recent years we have seen major improvements in cross-lingual methods, these methods still have some limitations that have not been addressed properly. An important problem is the out-of-vocabulary word (OOV) problem, i.e., words that occur in a document being processed, but that the model did not observe during training. The OOV problem is more significant in the case of low-resource languages, since there is relatively little training data available for them, and also in the case of morphologically-rich languages, since it is very likely that we do not observe a considerable number of their word forms in the training data. Approaches to learning sub-word embeddings have been proposed to address the OOV problem in monolingual models, but most prior work has not considered sub-word embeddings in cross-lingual models. The hypothesis of this thesis is that it is possible to leverage sub-word information to overcome the OOV problem in low-resource and morphologically-rich languages. This thesis presents a novel bilingual lexicon induction task to demonstrate the effectiveness of sub-word information in the cross-lingual space and how it can be employed to overcome the OOV problem. Moreover, this thesis presents a novel cross-lingual word representation method that incorporates sub-word information during the training process to learn a better cross-lingual shared space and also better represent OOVs in the shared space. This method is particularly suitable for low-resource scenarios and this claim is proven through a series of experiments on bilingual lexicon induction, monolingual word similarity, and a downstream task, document classification. More specifically, it is shown that this method is suitable for low-resource languages by conducting bilingual lexicon induction on twelve low-resource and morphologically-rich languages.

Embeddings in Natural Language Processing

Embeddings in Natural Language Processing
Author: Mohammad Taher Pilehvar
Publisher: Morgan & Claypool Publishers
Total Pages: 177
Release: 2020-11-13
Genre: Computers
ISBN: 1636390226

Download Embeddings in Natural Language Processing Book in PDF, Epub and Kindle

Embeddings have undoubtedly been one of the most influential research areas in Natural Language Processing (NLP). Encoding information into a low-dimensional vector representation, which is easily integrable in modern machine learning models, has played a central role in the development of NLP. Embedding techniques initially focused on words, but the attention soon started to shift to other forms: from graph structures, such as knowledge bases, to other types of textual content, such as sentences and documents. This book provides a high-level synthesis of the main embedding techniques in NLP, in the broad sense. The book starts by explaining conventional word vector space models and word embeddings (e.g., Word2Vec and GloVe) and then moves to other types of embeddings, such as word sense, sentence and document, and graph embeddings. The book also provides an overview of recent developments in contextualized representations (e.g., ELMo and BERT) and explains their potential in NLP. Throughout the book, the reader can find both essential information for understanding a certain topic from scratch and a broad overview of the most successful techniques developed in the literature.

EMBEDDIA.

EMBEDDIA.
Author: Marko Robnik Šikonja
Publisher:
Total Pages: 0
Release: 2019
Genre:
ISBN:

Download EMBEDDIA. Book in PDF, Epub and Kindle

Embeddings in Natural Language Processing

Embeddings in Natural Language Processing
Author: Mohammad Taher Pilehvar
Publisher: Springer Nature
Total Pages: 157
Release: 2022-05-31
Genre: Computers
ISBN: 3031021770

Download Embeddings in Natural Language Processing Book in PDF, Epub and Kindle

Embeddings have undoubtedly been one of the most influential research areas in Natural Language Processing (NLP). Encoding information into a low-dimensional vector representation, which is easily integrable in modern machine learning models, has played a central role in the development of NLP. Embedding techniques initially focused on words, but the attention soon started to shift to other forms: from graph structures, such as knowledge bases, to other types of textual content, such as sentences and documents. This book provides a high-level synthesis of the main embedding techniques in NLP, in the broad sense. The book starts by explaining conventional word vector space models and word embeddings (e.g., Word2Vec and GloVe) and then moves to other types of embeddings, such as word sense, sentence and document, and graph embeddings. The book also provides an overview of recent developments in contextualized representations (e.g., ELMo and BERT) and explains their potential in NLP. Throughout the book, the reader can find both essential information for understanding a certain topic from scratch and a broad overview of the most successful techniques developed in the literature.

EuroWordNet: A multilingual database with lexical semantic networks

EuroWordNet: A multilingual database with lexical semantic networks
Author: Piek Vossen
Publisher: Springer Science & Business Media
Total Pages: 180
Release: 2013-11-11
Genre: Computers
ISBN: 9401714916

Download EuroWordNet: A multilingual database with lexical semantic networks Book in PDF, Epub and Kindle

This book describes the main objective of EuroWordNet, which is the building of a multilingual database with lexical semantic networks or wordnets for several European languages. Each wordnet in the database represents a language-specific structure due to the unique lexicalization of concepts in languages. The concepts are inter-linked via a separate Inter-Lingual-Index, where equivalent concepts across languages should share the same index item. The flexible multilingual design of the database makes it possible to compare the lexicalizations and semantic structures, revealing answers to fundamental linguistic and philosophical questions which could never be answered before. How consistent are lexical semantic networks across languages, what are the language-specific differences of these networks, is there a language-universal ontology, how much information can be shared across languages? First attempts to answer these questions are given in the form of a set of shared or common Base Concepts that has been derived from the separate wordnets and their classification by a language-neutral top-ontology. These Base Concepts play a fundamental role in several wordnets. Nevertheless, the database may also serve many practical needs with respect to (cross-language) information retrieval, machine translation tools, language generation tools and language learning tools, which are discussed in the final chapter. The book offers an excellent introduction to the EuroWordNet project for scholars in the field and raises many issues that set the directions for further research in semantics and knowledge engineering.

Explorations in Word Embeddings

Explorations in Word Embeddings
Author: Zheng Zhang
Publisher:
Total Pages: 0
Release: 2019
Genre:
ISBN:

Download Explorations in Word Embeddings Book in PDF, Epub and Kindle

Word embeddings are a standard component of modern natural language processing architectures. Every time there is a breakthrough in word embedding learning, the vast majority of natural language processing tasks, such as POS-tagging, named entity recognition (NER), question answering, natural language inference, can benefit from it. This work addresses the question of how to improve the quality of monolingual word embeddings learned by prediction-based models and how to map contextual word embeddings generated by pretrained language representation models like ELMo or BERT across different languages.For monolingual word embedding learning, I take into account global, corpus-level information and generate a different noise distribution for negative sampling in word2vec. In this purpose I pre-compute word co-occurrence statistics with corpus2graph, an open-source NLP-application-oriented Python package that I developed: it efficiently generates a word co-occurrence network from a large corpus, and applies to it network algorithms such as random walks. For cross-lingual contextual word embedding mapping, I link contextual word embeddings to word sense embeddings. The improved anchor generation algorithm that I propose also expands the scope of word embedding mapping algorithms from context independent to contextual word embeddings.