A transportable distributional semantics architecture
Barzegar, Siamak
Barzegar, Siamak
Loading...
Publication Date
2019-02-13
Type
Thesis
Downloads
Citation
Abstract
Distributional semantics is built upon the assumption that the context surrounding a given word in text provides important information about its meaning (Distributional hypothesis). A rephrasing of the distributional hypothesis states that words that occur in similar contexts tend to have a similar meaning. Distributional semantics focuses on the construction of a semantic representation of a word based on the statistical distribution of word co-occurrence in texts. Distributional Semantic Models (DSMs) represent co-occurrence patterns under a vector space representation. In recent years, word embedding/distributional semantic models have evolved to become a fundamental component in many natural language processing (NLP) architectures due to their ability of capturing and quantifying semantic associations at scale. Distributional Semantics have been applied for different tasks in NLP area such as finding similar or related phrase/words, the computation of semantic relatedness measures, semantic relation classification and so forth. Distributional semantic models are strongly dependent on the size and the quality of the reference corpora, which embeds the common-sense knowledge necessary to build comprehensive models. While high-quality texts containing large-scale common-sense information and domain-specific information are present in English, such as Wikipedia, other languages may lack sufficient textual support to build comprehensive distributional models. Distributional Semantic Models are also often limited to semantic similarity/relatedness between two entities/terms with no explicit relation type. Often, it is not possible to assign a direct semantic relation between entities. This thesis seeks to analyse transportability aspects (Language and Domain) as and explores both coarse & fine-grained semantics for direct and indirect relation classification using a unified architecture (Indra) for developing language and domain independent DSM models with advanced (compositional) relation classification capabilities.
Publisher
NUI Galway
Publisher DOI
Rights
Attribution-NonCommercial-NoDerivs 3.0 Ireland