Instead, this study focuses on a specific machine learning task, namely text classification, exploring the effect of semantic augmentation on deep neural models to the classification performance. Our worked is focused on the feature level, applying semantic enrichment on the input space of the classification process. We separate the embedding generation from the semantic enrichment phase, as in Faruqui et al. (Reference Faruqui, Dodge, Jauhar, Dyer, Hovy and Smith2015), where the semantic augmentation can be applied as a post-processing step.
A deep semantic matching approach for identifying relevant messages for social media analysis Scientific Reports.
Posted: Tue, 25 Jul 2023 07:00:00 GMT [source]
We can note that the most common approach deals with latent semantics through Latent Semantic Indexing (LSI) [2, 120], a method that can be used for data dimension reduction and that is also known as latent semantic analysis. In this semantic space, alternative forms expressing the semantic text analysis same concept are projected to a common representation. It reduces the noise caused by synonymy and polysemy; thus, it latently deals with text semantics. Another technique in this direction that is commonly used for topic modeling is latent Dirichlet allocation (LDA) [121].
The assignment of meaning to terms is based on what other words usually occur in their close vicinity. To create such representations, you need many texts as training data, usually Wikipedia articles, books and websites. Natural language processing (NLP) is a field of artificial intelligence that focuses on creating interactions between computers and human language. It aims to facilitate communication between humans and machines by teaching computers to read, process, understand and perform actions based on natural language.
The retrofitting system is specialized in Glavaš and Vulić (Reference Glavaš and Vulić2018) via a feed-forward DNN that explicitly maps semantic/relational constraints into modified training instances to produce specialized embeddings. The authors report significant gains in a series of tasks that range from word similarity to lexical simplification and dialog state tracking. Additionally, they explore a joint model where the objective function considers a weighted linear combination of both corpus co-occurrence statistics and relatedness based on the knowledge resource. Thus, this paper reports a systematic mapping study to overview the development of semantics-concerned studies and fill a literature review gap in this broad research field through a well-defined review process. Semantics can be related to a vast number of subjects, and most of them are studied in the natural language processing field.
The formal semantics defined by Sheth et al. [28] is commonly represented by description logics, a formalism for knowledge representation. The application of description logics in natural language processing is the theme of the brief review presented by Cheng et al. [29]. Specifically for the task of irony detection, Wallace [23] presents both philosophical formalisms and machine learning approaches. The author argues that a model of the speaker is necessary to improve current machine learning methods and enable their application in a general problem, independently of domain. He discusses the gaps of current methods and proposes a pragmatic context model for irony detection. Methods that deal with latent semantics are reviewed in the study of Daud et al. [16].
Creemos en la capacidad y en el poder evolutivo de la mente para superar las barreras que impone el movimiento
© Copyright 2023 Design by malzate
Comentarios recientes