1,434 results on '"distributional semantics"'
Search Results
2. Decomposing unaccusativity: a statistical modelling approach.
- Author
-
Kim, Songhee, Binder, Jeffrey R., Humphries, Colin, and Conant, Lisa L.
- Subjects
- *
COMPARATIVE grammar , *STATISTICAL models , *RESEARCH funding , *NEUROBIOLOGY , *HYPOTHESIS , *SEMANTICS , *COGNITION , *LANGUAGE acquisition - Abstract
While the two types of intransitive verbs, i.e. unergative and unaccusative, are hypothesised to be syntactically represented, many have proposed a semantic account where abstract properties related to agentivity and telicity, often conceptualised as binary properties, determine the classification. Here we explore the extent to which graded, embodied features rooted in neurobiological systems contribute to the distinction, representing verb meanings as continuous human ratings over various experiential dimensions. Unlike prior studies that classified verbs based on categorical intuition, we assessed the degree of unaccusativity by acceptability of the prenominal past participle construction, one of the unaccusativity diagnostics. Five models were constructed to explain these data: categorical syntactic/semantic, feature-based event-semantic, experiential, and distributional models. The experiential model best explained the diagnostic test data, suggesting that the unaccusative/unergative distinction may be an emergent phenomenon related to differences in underlying experiential content. The experiential model's advantages, including interpretability and scalability, are also discussed. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Instances of bias: the gendered semantics of generic masculines in German revealed by instance vectors.
- Author
-
Schmitz, Dominic
- Subjects
GERMAN language ,VECTOR valued functions ,BEHAVIORAL research ,SEMANTICS ,MALES - Abstract
While research using behavioural methods has repeatedly shown that generic masculines in German come with a male bias, computational methods only entered this area of research very recently. The present paper shows that some assumptions made by these recent computational studies – treating genericity as an inflectional function and computing a vector for generic usage strongly correlated with the grammatical masculine – are not without issue, and offers the use of semantic instance vectors as a possible solution to these issues. Beyond this methodological improvement, the present paper finds that generic masculines are indeed semantically more similar to specific masculines than to specific feminines – results that are in line with findings by the recent computational studies and the majority of previous behavioural studies. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Valence without meaning: Investigating form and semantic components in pseudowords valence.
- Author
-
Gatti, Daniele, Raveling, Laura, Petrenco, Aliona, and Günther, Fritz
- Subjects
- *
LEGAL judgments , *SEMANTICS , *VOCABULARY , *HUMAN beings , *LANGUAGE & languages - Abstract
Valence is a dominant semantic dimension, and it is fundamentally linked to basic approach-avoidance behavior within a broad range of contexts. Previous studies have shown that it is possible to approximate the valence of existing words based on several surface-level and semantic components of the stimuli. Parallelly, recent studies have shown that even completely novel and (apparently) meaningless stimuli, like pseudowords, can be informative of meaning based on the information that they carry at the subword level. Here, we aimed to further extend this evidence by investigating whether humans can reliably assign valence to pseudowords and, additionally, to identify the factors explaining such valence judgments. In Experiment 1, we trained several models to predict valence judgments for existing words from their combined form and meaning information. Then, in Experiment 2 and Experiment 3, we extended the results by predicting participants' valence judgments for pseudowords, using a set of models indexing different (possible) sources of valence and selected the best performing model in a completely data-driven procedure. Results showed that the model including basic surface-level (i.e., letters composing the pseudoword) and orthographic neighbors information performed best, thus tracing back pseudoword valence to these components. These findings support perspectives on the nonarbitrariness of language and provide insights regarding how humans process the valence of novel stimuli. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. The good, the bad, and the ambivalent: Extrapolating affective values for 38,000+ Chinese words via a computational model.
- Author
-
Wang, Tianqi and Xu, Xu
- Subjects
- *
NATURAL language processing , *PROBABILITY density function , *VALUES (Ethics) , *CHINESE language , *DATABASES - Abstract
Word affective ratings are important tools in psycholinguistic research, natural language processing, and many other fields. However, even for well-studied languages, such norms are usually limited in scale. To extrapolate affective (i.e., valence and arousal) values for words in the SUBTLEX-CH database (Cai & Brysbaert, 2010, PLoS ONE, 5(6):e10729), we implemented a computational neural network which captured how words' vector-based semantic representations corresponded to the probability densities of their valence and arousal. Based on these probability density functions, we predicted not only a word's affective values, but also their respective degrees of variability that could characterize individual differences in human affective ratings. The resulting estimates of affective values largely converged with human ratings for both valence and arousal, and the estimated degrees of variability also captured important features of the variability in human ratings. We released the extrapolated affective values, together with their corresponding degrees of variability, for over 38,000 Chinese words in the Open Science Framework (https://osf.io/s9zmd/). We also discussed how the view of embodied cognition could be illuminated by this computational model. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. AI Language Models: An Opportunity to Enhance Language Learning.
- Author
-
Cong, Yan
- Subjects
NATURAL language processing ,LANGUAGE models ,SECOND language acquisition ,LANGUAGE research ,LANGUAGE acquisition - Abstract
AI language models are increasingly transforming language research in various ways. How can language educators and researchers respond to the challenge posed by these AI models? Specifically, how can we embrace this technology to inform and enhance second language learning and teaching? In order to quantitatively characterize and index second language writing, the current work proposes the use of similarities derived from contextualized meaning representations in AI language models. The computational analysis in this work is hypothesis-driven. The current work predicts how similarities should be distributed in a second language learning setting. The results suggest that similarity metrics are informative of writing proficiency assessment and interlanguage development. Statistically significant effects were found across multiple AI models. Most of the metrics could distinguish language learners' proficiency levels. Significant correlations were also found between similarity metrics and learners' writing test scores provided by human experts in the domain. However, not all such effects were strong or interpretable. Several results could not be consistently explained under the proposed second language learning hypotheses. Overall, the current investigation indicates that with careful configuration and systematic metrics design, AI language models can be promising tools in advancing language education. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Medical device similarity analysis: a promising approach to medical device equivalence regulation.
- Author
-
Sündermann, Jan, Delgado Fernandez, Joaquin, Kellner, Rupert, Doll, Theodor, Froriep, Ulrich P., and Bitsch, Annette
- Subjects
LATENT semantic analysis ,MEDICAL equipment ,CHEMICAL fingerprinting ,MEDICAL laws ,VALUES (Ethics) - Abstract
Background: This study aims to facilitate the identification of similar devices for both, the European Medical Device Regulation (MDR) and the US 510(k) equivalence pathway by leveraging existing data. Both are related to the regulatory pathway of read across for chemicals, where toxicological data from a known substance is transferred to one under investigation, as they aim to streamline the accreditation process for new devices and chemicals. Research design and methods: This study employs latent semantic analysis to generate similarity values, harnessing the US Food and Drug Administration 510k-database, utilizing their 'Device Descriptions' and 'Intended Use' statements. Results: For the representative inhaler cluster, similarity values up to 0.999 were generated for devices within a 510(k)-predicate tree, whereas values up to 0.124 were gathered for devices outside this group. Conclusion: Traditionally, MDR equivalence involves manual review of many devices, which is laborious. However, our results suggest that the automated calculation of similarity coefficients streamlines this process, thus reducing regulatory effort, which can be beneficial for patients needing medical devices. Although this study is focused on the European perspective, it can find application within 510(k) equivalence regulation. The conceptual approach is reminiscent of chemical fingerprint similarity analysis employed in read-across. Plain Language Summary: This study addresses improvement of the registration process for medical devices by using automated methods to determine how similar they are to existing devices. Such a process is already used in chemistry for analysis of related substances. In the context of Medical Device Regulation (MDR), which sets standards for these devices, this process might be applicable in device equivalence evaluation. Traditionally, proving equivalence involves manually finding devices that are similar, but this is time-consuming, repetitive and labor-intensive. This study proposes a new approach, using advanced computer methods and a database from the US Food and Drug Administration (FDA) to automatically identify similar devices. This could make the process much quicker and more accurate and furthermore reduce bias. The study suggests that by applying these automated methods, the impact of recent regulatory changes could be reduced. This means that proving equivalence, a critical step to facilitate device accreditation, could be done more efficiently. The study shows potential for a significant transformation in compliance processes within the medical device industry, making them more streamlined and automated. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Training and evaluation of vector models for Galician.
- Author
-
Garcia, Marcos
- Subjects
- *
MACHINE learning , *LANGUAGE models , *VECTOR spaces , *SEMANTICS , *CORPORA - Abstract
This paper presents a large and systematic assessment of distributional models for Galician. To this end, we have first trained and evaluated static word embeddings (e.g., word2vec, GloVe), and then compared their performance with that of current contextualised representations generated by neural language models. First, we have compiled and processed a large corpus for Galician, and created four datasets for word analogies and concept categorisation based on standard resources for other languages. Using the aforementioned corpus, we have trained 760 static vector space models which vary in their input representations (e.g., adjacency-based versus dependency-based approaches), learning algorithms, size of the surrounding contexts, and in the number of vector dimensions. These models have been evaluated both intrinsically, using the newly created datasets, and on extrinsic tasks, namely on POS-tagging, dependency parsing, and named entity recognition. The results provide new insights into the performance of different vector models in Galician, and about the impact of several training parameters on each task. In general, fastText embeddings are the static representations with the best performance in the intrinsic evaluations and in named entity recognition, while syntax-based embeddings achieve the highest results in POS-tagging and dependency parsing, indicating that there is no significant correlation between the performance in the intrinsic and extrinsic tasks. Finally, we have compared the performance of static vector representations with that of BERT-based word embeddings, whose fine-tuning obtains the best performance on named entity recognition. This comparison provides a comprehensive state-of-the-art of current models in Galician, and releases new transformer-based models for NER. All the resources used in this research are freely available to the community, and the best models have been incorporated into SemantiGal, an online tool to explore vector representations for Galician. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. The pluralization palette: unveiling semantic clusters in English nominal pluralization through distributional semantics.
- Author
-
Shafaei-Bajestan, Elnaz, Moradipour-Tari, Masoumeh, Uhrig, Peter, and Baayen, R. Harald
- Abstract
Using distributional semantics, we show that English nominal pluralization exhibits semantic clusters. For instance, the change in semantic space from singulars to plurals differs depending on whether a word denotes, e.g., a fruit, or an animal. Languages with extensive noun classes such as Swahili and Kiowa distinguish between these kind of words in their morphology. In English, even though not marked morphologically, plural semantics actually also varies by semantic class. A semantically informed method, CosClassAvg, is introduced that is compared to two other methods, one implementing a fixed shift from singular to plural, and one creating plural vectors from singular vectors using a linear mapping (FRACSS). Compared to FRACSS, CosClassAvg predicted plural vectors that were more similar to the corpus-extracted plural vectors in terms of vector length, but somewhat less similar in terms of orientation. Both FRACSS and CosClassAvg outperform the method using a fixed shift vector to create plural vectors, which does not do justice to the intricacies of English plural semantics. A computational modeling study revealed that the observed difference between the plural semantics generated by these three methods carries over to how well a computational model of the listener can understand previously unencountered plural forms. Among all methods, CosClassAvg provides a good balance for the trade-off between productivity (being able to understand novel plural forms) and faithfulness to corpus-extracted plural vectors (i.e., understanding the particulars of the meaning of a given plural form). [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Large language models and linguistic intentionality.
- Author
-
Grindrod, Jumbly
- Abstract
Do large language models like Chat-GPT or Claude meaningfully use the words they produce? Or are they merely clever prediction machines, simulating language use by producing statistically plausible text? There have already been some initial attempts to answer this question by showing that these models meet the criteria for entering meaningful states according to metasemantic theories of mental content. In this paper, I will argue for a different approach—that we should instead consider whether language models meet the criteria given by our best metasemantic theories of linguistic content. In that vein, I will illustrate how this can be done by applying two such theories to the case of language models: Gareth Evans’ (1982) account of naming practices and Ruth Millikan’s (1984, 2004, 2005) teleosemantics. In doing so, I will argue that it is a mistake to think that the failure of LLMs to meet plausible conditions for mental intentionality thereby renders their outputs meaningless, and that a distinguishing feature of linguistic intentionality—dependency on a pre-existing linguistic system—allows for the plausible result that LLM outputs are meaningful. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. Words do not just label concepts: activating superordinate categories through labels, lists, and definitions.
- Author
-
Rissman, Lilia and Lupyan, Gary
- Subjects
- *
PROMPTS (Psychology) , *TASK performance , *PHONOLOGICAL awareness , *SEMANTICS , *CONCEPTS , *VOCABULARY , *ENGLISH language , *LANGUAGE acquisition - Abstract
We investigate the interface between concepts and word meanings by asking English speakers to list members of superordinate categories under one of three conditions: (1) when cued by a label (e.g. animals), (2) an exemplar list (e.g. dog, cat, mouse), or (3) a definition (e.g. "living creatures that roam the Earth"). We find that categories activated by labels lead to participants listing more category-typical responses, as quantified through typicality ratings, similarity in word embedding space, and accuracy in guessing category labels. This effect is stronger for some categories than others (e.g. stronger for appetizers than animals). These results support the view that a word is not merely a label for a concept, but rather a unique way of accessing and organizing conceptual space. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
12. Where Is Happily Ever After? A Study of Emotions and Locations in Russian Short Stories of 1900–1930
- Author
-
Moskvina, Anna, Kirina, Margarita, Brilly, Mitja, Advisory Editor, Hoalst-Pullen, Nancy, Advisory Editor, Leitner, Michael, Advisory Editor, Patterson, Mark W., Advisory Editor, Veress, Márton, Advisory Editor, Bakaev, Maxim, editor, Bolgov, Radomir, editor, Chugunov, Andrei V., editor, Pereira, Roberto, editor, R, Elakkiya, editor, and Zhang, Wei, editor
- Published
- 2024
- Full Text
- View/download PDF
13. Unsupervised Ultra-Fine Entity Typing with Distributionally Induced Word Senses
- Author
-
Sevgili, Özge, Remus, Steffen, Jana, Abhik, Panchenko, Alexander, Biemann, Chris, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Ignatov, Dmitry I., editor, Khachay, Michael, editor, Kutuzov, Andrey, editor, Madoyan, Habet, editor, Makarov, Ilya, editor, Nikishina, Irina, editor, Panchenko, Alexander, editor, Panov, Maxim, editor, Pardalos, Panos M., editor, Savchenko, Andrey V., editor, Tsymbalov, Evgenii, editor, Tutubalina, Elena, editor, and Zagoruyko, Sergey, editor
- Published
- 2024
- Full Text
- View/download PDF
14. Context Synthesis Accelerates Vocabulary Learning Through Reading: The Implication of Distributional Semantic Theory on Second Language Vocabulary Research
- Author
-
Wang-Kildegaard, Bowen and Ji, Feng
- Subjects
vocabulary learning ,reading ,distributional semantics ,second language acquisition ,psycholinguistics ,computer-assisted language learning ,corpus linguistics - Abstract
Besides explicit inference of word meanings, associating words with diverse contexts may be a key mechanism underlying vocabulary learning through reading. Drawing from distributional semantic theory, we developed a text modification method called reflash to facilitate both word-context association and explicit inference. Using a set of left and right arrows, learners can jump to a target word’s previous or subsequent occurrences in digital books to synthesize clues across contexts. Participants read stories with target words modified by reflash-only, gloss-only, gloss + reflash, or unmodified. Learning outcomes were measured via Vocabulary Knowledge Scale and a researcher-developed interview to probe word-context association. We modeled the learning trajectories of words across five weeks among three adolescent L2 English learners (113 word-learner pairings) using Bayesian multilevel models. We found that reflash-only words made more gains than words in other conditions on both outcomes, controlling for key covariates such as types of existing knowledge. Our analysis also revealed that context synthesis may be particularly useful for learning specific types of words like homonyms, which has significant pedagogical implications.
- Published
- 2023
15. Language as a cognitive and social tool at the time of large language models
- Author
-
Borghi, Anna M., De Livio, Chiara, Gervasi, Angelo Mattia, Mannella, Francesco, Nolfi, Stefano, and Tummolini, Luca
- Published
- 2024
- Full Text
- View/download PDF
16. Domain embeddings for generating complex descriptions of concepts in Italian language
- Author
-
Maisto, Alessandro
- Published
- 2024
- Full Text
- View/download PDF
17. From vector spaces to DRM lists: False Memory Generator, a software for automated generation of lists of stimuli inducing false memories.
- Author
-
Petilli, Marco A., Marelli, Marco, Mazzoni, Giuliana, Marchetti, Michela, Rinaldi, Luca, and Gatti, Daniele
- Subjects
- *
FALSE memory syndrome , *VECTOR spaces , *STIMULUS & response (Psychology) , *COGNITIVE psychology , *NEW words , *MEMORY testing - Abstract
The formation of false memories is one of the most widely studied topics in cognitive psychology. The Deese–Roediger–McDermott (DRM) paradigm is a powerful tool for investigating false memories and revealing the cognitive mechanisms subserving their formation. In this task, participants first memorize a list of words (encoding phase) and next have to indicate whether words presented in a new list were part of the initially memorized one (recognition phase). By employing DRM lists optimized to investigate semantic effects, previous studies highlighted a crucial role of semantic processes in false memory generation, showing that new words semantically related to the studied ones tend to be more erroneously recognized (compared to new words less semantically related). Despite the strengths of the DRM task, this paradigm faces a major limitation in list construction due to its reliance on human-based association norms, posing both practical and theoretical concerns. To address these issues, we developed the False Memory Generator (FMG), an automated and data-driven tool for generating DRM lists, which exploits similarity relationships between items populating a vector space. Here, we present FMG and demonstrate the validity of the lists generated in successfully replicating well-known semantic effects on false memory production. FMG potentially has broad applications by allowing for testing false memory production in domains that go well beyond the current possibilities, as it can be in principle applied to any vector space encoding properties related to word referents (e.g., lexical, orthographic, phonological, sensory, affective, etc.) or other type of stimuli (e.g., images, sounds, etc.). [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. Extending the Architecture of Language From a Multimodal Perspective.
- Author
-
Hagoort, Peter and Özyürek, Aslı
- Abstract
Language is inherently multimodal. In spoken languages, combined spoken and visual signals (e.g., co‐speech gestures) are an integral part of linguistic structure and language representation. This requires an extension of the parallel architecture, which needs to include the visual signals concomitant to speech. We present the evidence for the multimodality of language. In addition, we propose that distributional semantics might provide a format for integrating speech and co‐speech gestures in a common semantic representation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
19. Visual experience modulates the sensitivity to the distributional history of words in natural language
- Author
-
Anceresi, Giorgia, Gatti, Daniele, Vecchi, Tomaso, Marelli, Marco, and Rinaldi, Luca
- Published
- 2024
- Full Text
- View/download PDF
20. Towards a word similarity gold standard for Akkadian: creation and model optimization.
- Author
-
Sahala, Aleksi, Baragli, Beatrice, Lentini, Giulia, and Tushingham, Poppy
- Subjects
NATURAL language processing ,MACHINE learning ,SEMANTICS - Abstract
We present a word similarity gold standard for Akkadian, a language documented in ancient Mesopotamian sources from the 24th century BCE until the first century CE. The gold standard comprises 300 word pairs ranked by their paradigmatic similarity by five independently working Assyriologists. We use the gold standard to tune PMI + SVD and fastText models to improve their performance. We also present a hyper-parametrized PMI + SVD model for building count-based word embeddings, that aims to deal with the data sparsity and repetition issues encountered in Akkadian texts. Our model combines Dirichlet smoothing with context distribution smoothing, and uses context similarity weighting to down-sample distortion caused by formulaic litanies and partially or fully duplicated passages. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
21. Frequency effects in linear discriminative learning.
- Author
-
Heitmeier, Maria, Yu-Ying Chuang, Axen, Seth D., and Baayen, R. Harald
- Subjects
MACHINE learning ,WORD frequency ,AUTOMATIC speech recognition ,REVENUE accounting ,WORD recognition ,MANDARIN dialects ,REACTION time - Abstract
Word frequency is a strong predictor in most lexical processing tasks. Thus, any model of word recognition needs to account for howword frequency effects arise. The Discriminative Lexicon Model (DLM)models lexical processing withmappings between words' forms and their meanings. Comprehension and production are modeled via linear mappings between the two domains. So far, the mappings within the model can either be obtained incrementally via error-driven learning, a computationally expensive process able to capture frequency effects, or in an efficient, but frequency-agnostic solution modeling the theoretical endstate of learning (EL) where all words are learned optimally. In the present study we show how an efficient, yet frequency-informed mapping between form and meaning can be obtained (Frequency-informed learning; FIL). We find that FIL well approximates an incremental solution while being computationallymuch cheaper. FIL shows a relatively low type-and high token-accuracy, demonstrating that the model is able to process most word tokens encountered by speakers in daily life correctly. We use FIL to model reaction times in the Dutch Lexicon Project by means of a Gaussian Location Scale Model and find that FIL predicts well the S-shaped relationship between frequency and the mean of reaction times but underestimates the variance of reaction times for low frequency words. FIL is also better able to account for priming effects in an auditory lexical decision task in Mandarin Chinese, compared to EL. Finally, we used ordered data from CHILDES to compare mappings obtained with FIL and incremental learning. We show that themappings are highly correlated, but that with FIL some nuances based on word ordering effects are lost. Our results show how frequency effects in a learning model can be simulated efficiently, and raise questions about how to best account for low-frequency words in cognitive models. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. Transformer Networks of Human Conceptual Knowledge.
- Author
-
Bhatia, Sudeep and Richie, Russell
- Subjects
- *
SIMILARITY (Psychology) , *BIG data , *COGNITION research - Abstract
We present a computational model capable of simulating aspects of human knowledge for thousands of real-world concepts. Our approach involves a pretrained transformer network that is further fine-tuned on large data sets of participant-generated feature norms. We show that such a model can successfully extrapolate from its training data, and predict human knowledge for new concepts and features. We apply our model to stimuli from 25 previous experiments in semantic cognition research and show that it reproduces many findings on semantic verification, concept typicality, feature distribution, and semantic similarity. We also compare our model against several variants, and by doing so, establish the model properties that are necessary for good prediction. The success of our approach shows how a combination of language data and (laboratory-based) psychological data can be used to build models with rich world knowledge. Such models can be used in the service of new psychological applications, such as the modeling of naturalistic semantic verification and knowledge retrieval, as well as the modeling of real-world categorization, decision-making, and reasoning. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
23. Approaches to Cross-Language Retrieval of Similar Legal Documents Based on Machine Learning.
- Author
-
Zhebel, V. V., Devyatkin, D. A., Zubarev, D. V., and Sochenkov, I. V.
- Abstract
In order to study global experience for legislation changing and rule-making necessitates, tools for information retrieval of regulatory documents written in different languages become increasingly necessary. One of the aspects of information identification is retrieval of thematically similar documents for a given input document. In this context, an important task of cross-lingual search arises when the user of an information system specifies a reference document in one language, and the search results contain relevant documents in other languages. The article describes different approaches to solving this problem: from classic mediator-based methods to more modern solutions, based on distributional semantics. The test collection used in the study was taken from the United Nations Digital Library, which provides legal documents in both the original English and their Russian translations. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
24. The online hostility hypothesis: representations of Muslims in online media.
- Author
-
Sandberg, Linn, Dahlberg, Stefan, and Ivarsflaten, Elisabeth
- Subjects
HOSTILITY ,VIRTUAL communities ,MUSLIMS ,SOCIAL interaction - Abstract
Using a large data set of online media content in eight European countries, this paper broadens the empirical investigation of the online hostility hypothesis, which posits that interactions on social sites such as blogs and forums contain more hostile expressions toward minority groups than social interactions offline or in editorial news media. Overall, our results are consistent with the online hostility hypothesis when comparing news media content with social sites, but we find that negatively charged representations are common in both media types. It is instead the amount of attention to Muslims and Islam on social sites that most clearly differs and is the main driver of online hostility in the online media environment more broadly conceived. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
25. Predicting Patterns of Similarity Among Abstract Semantic Relations
- Author
-
Ichien, Nicholas, Lu, Hongjing, and Holyoak, Keith J
- Subjects
Behavioral and Social Science ,Clinical Research ,Basic Behavioral and Social Science ,Humans ,Individuality ,Judgment ,Semantics ,relations ,similarity ,analogy ,reasoning ,distributional semantics ,Psychology ,Cognitive Sciences ,Experimental Psychology - Abstract
Although models of word meanings based on distributional semantics have proved effective in predicting human judgments of similarity among individual concepts, it is less clear whether or how such models might be extended to account for judgments of similarity among relations between concepts. Here we combine an individual-differences approach with computational modeling to predict human judgments of similarity among word pairs instantiating a variety of abstract semantic relations (e.g., contrast, cause-effect, part-whole). A measure of cognitive capacity predicted individual differences in the ability to discriminate among distinct relations. The human pattern of relational similarity judgments, both at the group level and for individual participants, was best predicted by a model that takes representations of word meanings based on distributional semantics as its inputs and uses them to learn an explicit representation of relations. These findings indicate that although the meanings of abstract semantic relations are not directly coded in the meanings of individual words, important aspects of relational similarity can be derived from distributional semantics. (PsycInfo Database Record (c) 2022 APA, all rights reserved).
- Published
- 2022
26. AI Language Models: An Opportunity to Enhance Language Learning
- Author
-
Yan Cong
- Subjects
large language models ,natural language processing ,second language writing ,automatic writing assessment ,cosine similarity scores ,distributional semantics ,Information technology ,T58.5-58.64 - Abstract
AI language models are increasingly transforming language research in various ways. How can language educators and researchers respond to the challenge posed by these AI models? Specifically, how can we embrace this technology to inform and enhance second language learning and teaching? In order to quantitatively characterize and index second language writing, the current work proposes the use of similarities derived from contextualized meaning representations in AI language models. The computational analysis in this work is hypothesis-driven. The current work predicts how similarities should be distributed in a second language learning setting. The results suggest that similarity metrics are informative of writing proficiency assessment and interlanguage development. Statistically significant effects were found across multiple AI models. Most of the metrics could distinguish language learners’ proficiency levels. Significant correlations were also found between similarity metrics and learners’ writing test scores provided by human experts in the domain. However, not all such effects were strong or interpretable. Several results could not be consistently explained under the proposed second language learning hypotheses. Overall, the current investigation indicates that with careful configuration and systematic metrics design, AI language models can be promising tools in advancing language education.
- Published
- 2024
- Full Text
- View/download PDF
27. Assessment in Conversational Intelligent Tutoring Systems: Are Contextual Embeddings Really Better?
- Author
-
Carmon, Colin M., Hu, Xiangen, Graesser, Arthur C., Filipe, Joaquim, Editorial Board Member, Ghosh, Ashish, Editorial Board Member, Prates, Raquel Oliveira, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, Wang, Ning, editor, Rebolledo-Mendez, Genaro, editor, Dimitrova, Vania, editor, Matsuda, Noboru, editor, and Santos, Olga C., editor
- Published
- 2023
- Full Text
- View/download PDF
28. A Hybrid Approach of Distributional Semantics and Event Semantics for Telicity
- Author
-
Yanaka, Hitomi, Kacprzyk, Janusz, Series Editor, Loukanova, Roussanka, editor, Lumsdaine, Peter LeFanu, editor, and Muskens, Reinhard, editor
- Published
- 2023
- Full Text
- View/download PDF
29. The online hostility hypothesis: representations of Muslims in online media
- Author
-
Linn Sandberg, Stefan Dahlberg, and Elisabeth Ivarsflaten
- Subjects
Online hostility ,anti-muslims ,distributional semantics ,word embeddings ,Social Sciences - Abstract
ABSTRACTUsing a large data set of online media content in eight European countries, this paper broadens the empirical investigation of the online hostility hypothesis, which posits that interactions on social sites such as blogs and forums contain more hostile expressions toward minority groups than social interactions offline or in editorial news media. Overall, our results are consistent with the online hostility hypothesis when comparing news media content with social sites, but we find that negatively charged representations are common in both media types. It is instead the amount of attention to Muslims and Islam on social sites that most clearly differs and is the main driver of online hostility in the online media environment more broadly conceived.
- Published
- 2023
- Full Text
- View/download PDF
30. Probing Lexical Ambiguity in Chinese Characters via Their Word Formations: Convergence of Perceived and Computed Metrics.
- Author
-
Wang, Tianqi, Xu, Xu, Xie, Xurong, and Ng, Manwa Lawrence
- Subjects
- *
CHINESE characters , *AMBIGUITY , *PROSODIC analysis (Linguistics) , *POLYSEMY , *NATIVE language , *VOCABULARY , *SEMANTICS - Abstract
Lexical ambiguity is pervasive in language, and the nature of the representations of an ambiguous word's multiple meanings is yet to be fully understood. With a special focus on Chinese characters, the present study first established that native speaker's perception about a character's number of meanings was heavily influenced by the availability of its distinct word formations, while whether these meanings would be perceived to be closely related was driven by further conceptual analysis. These notions were operationalized as two computed metrics, which assessed the degree of dispersion across individual word formations and the degree of propinquity across clusters of word formations, respectively, in a distributional semantic space. The observed correlations between the computed and the perceived metrics indicated that the utility of word formations to tap into meaning representations of Chinese characters was indeed cognitively plausible. The results have demonstrated the extent to which distributional semantics could inform about meaning representations of Chinese characters, which has theoretical implications for the representation of ambiguous words more generally. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
31. Free Association in a Neural Network.
- Author
-
Richie, Russell, Aka, Ada, and Bhatia, Sudeep
- Subjects
- *
RECOLLECTION (Psychology) , *BIG data - Abstract
Free association among words is a fundamental and ubiquitous memory task. Although distributed semantics (DS) models can predict the association between pairs of words, and semantic network (SN) models can describe transition probabilities in free association data, there have been few attempts to apply established cognitive process models of memory search to free association data. Thus, researchers are currently unable to explain the dynamics of free association using memory mechanisms known to be at play in other retrieval tasks, such as free recall from lists. We address this issue using a popular neural network model of free recall, the context maintenance and retrieval (CMR) model, which we fit using stochastic gradient descent on a large data set of free association norms. Special cases of CMR mimic existing DS and SN models of free association, and we find that CMR outperforms these models on out-of-sample free association data. We also show that training CMR on free association data generates improved predictions for free recall from lists, demonstrating the value of free association for the study of many different types of memory phenomena. Overall, our analysis provides a new account of the dynamics of free association, predicts free association with increased accuracy, integrates theories of free association with established models of memory, and shows how large data sets and neural network training methods can be used to model complex cognitive processes that operate over thousands of representations. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
32. Predicción de valores léxico-afectivos para un conjunto de emociones complejas mediante análisis de semántica distribucional.
- Author
-
Yerro Avincetto, Matías M., Vivas, Jorge R., González, Mariela A., and Passoni, Lucía I.
- Subjects
- *
SELF-evaluation , *SEMANTIC network analysis , *COMPUTATIONAL intelligence , *EMOTIONS , *MODELS & modelmaking , *SEMANTICS , *ACQUISITION of data , *SPANISH language , *AGE groups , *WORD recognition - Abstract
The tradition of dimensional models in the study of emotions suggests that affective space is better defined by a small number of non-specific general dimensions. This dimensional perspective in the study of emotions postulates that the minimal entities of representation are dimensions such as valence (attraction vs. rejection) and arousal (level of activation). Data collection in this perspective is done through the laborious process of resorting to the estimates of hundreds of people who must decide on a continuum of two dimensions: how positive or negative the object to which the concept alludes is, and what level of arousal it generates. To do so, generally, an image-based questionnaire developed to measure an emotional response is used, called SAM (Self Assessment Manikin), which is a set of synthesized drawings that can be used to guide the participants' response. As can be noted, the volume and quality of the procedures used to study these affective variables associated with concepts involves an arduous process of data collection and processing. The complexity of this work, due to the enormous amount and type of data, makes computational intelligence a very useful and novel tool for its approach in order to obtain reliable and reproducible results. Processing by means of distributional semantics obtains the meaning of a word by locating the context in which it appears through an intelligent search in large volumes of electronically stored data. Lexical-affective information, represented by the valence and activation dimensions, is a type of information that seems to be represented by the social use of a term and, therefore, it is plausible to infer it by studying the contexts in which the term appears. In this paper we consider the plausibility of the application of the method of estimating lexical-affective values by means of distributional semantics for the Spanish language. In order to achieve this objective, the Spanish adaptation of the ANEW is used for this purpose and, two procedures were carried out. On the one hand, the most traditionally used computational method of estimation by means of linear regressions was compared with a model based on neural networks, which showed the better fit of the latter. On the other hand, lexical-affective values were estimated for a set of complex emotions and, in order to verify the strength of the results, such estimation was compared with empirical data taken and processed in a linguistic community (Argentine), different from the community that gave rise to the data with which the computational model was trained (Spanish). The results have been very encouraging since, between the computationally derived estimates and the empirically derived data with the Argentine population, correlations were found to be sufficiently strong and comparable to those that can be found when the same comparison is made between empirically derived results with different age groups, or between different genders, or when comparing different geographical regions. These results point to a striking common base of the language, a basic common core of concepts, which can be explored by means of the procedures and techniques of distributional semantics. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
33. Is More Always Better? Testing the Addition Bias for German Language Statistics.
- Author
-
Wolfer, Sascha
- Subjects
- *
GERMAN language , *WORD frequency , *LATENT semantic analysis , *ENGLISH language , *LANGUAGE ability testing - Abstract
This replication study aims to investigate a potential bias toward addition in the German language, building upon previous findings of Winter and colleagues who identified a similar bias in English. Our results confirm a bias in word frequencies and binomial expressions, aligning with these previous findings. However, the analysis of distributional semantics based on word vectors did not yield consistent results for German. Furthermore, our study emphasizes the crucial role of selecting appropriate translational equivalents, highlighting the significance of considering language‐specific factors when testing for such biases for languages other than English. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
34. What do firms say in reporting on impacts of climate change? An approach to monitoring ESG actions and environmental policy.
- Author
-
Chou, Christine, Clark, Robin, and Kimbrough, Steven O.
- Subjects
ENVIRONMENTAL policy ,CLIMATE change ,TEXT mining ,RESEARCH questions ,ENVIRONMENTAL responsibility ,ETHICAL investments - Abstract
This article focuses on two research questions arising from the 2010 U.S. Securities and Exchange Commission (SEC) Advisory on climate change reporting: (1) How does the discussion of climate change in SEC filings change after the Advisory? and (2) What are firms talking about when they talk about climate change? Findings were obtained from the 218,000 10‐K filings to the SEC during the 2000–2019 period. The study develops and applies text mining methodology based on extracting information from the "semantic associates" in the "neighborhoods" of indicative terms. On (1) it finds that climate change‐related reporting does increase substantially after the SEC guidance. On (2) a nuanced picture emerges. Firms with comparatively larger transition risks tend to discuss climate change comparatively more, focusing on regulation‐related topics. Firms exposed to the physical risks of climate change tend to discuss climate change somewhat less, focusing on meteorological topics. The results enrich our understanding regarding environmental policies and firms' behaviors regarding climate change. Theoretical and practical implications are provided. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
35. Unsupervised image translation with distributional semantics awareness.
- Author
-
Peng, Zhexi, Wang, He, Weng, Yanlin, Yang, Yin, and Shao, Tianjia
- Subjects
DEEP learning ,GENERATIVE adversarial networks - Abstract
Unsupervised image translation (UIT) studies the mapping between two image domains. Since such mappings are under-constrained, existing research has pursued various desirable properties such as distributional matching or two-way consistency. In this paper, we re-examine UIT from a new perspective: distributional semantics consistency, based on the observation that data variations contain semantics, e.g., shoes varying in colors. Further, the semantics can be multi-dimensional, e.g., shoes also varying in style, functionality, etc. Given two image domains, matching these semantic dimensions during UIT will produce mappings with explicable correspondences, which has not been investigated previously. We propose distributional semantics mapping (DSM), the first UIT method which explicitly matches semantics between two domains. We show that distributional semantics has been rarely considered within and beyond UIT, even though it is a common problem in deep learning. We evaluate DSM on several benchmark datasets, demonstrating its general ability to capture distributional semantics. Extensive comparisons show that DSM not only produces explicable mappings, but also improves image quality in general. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
36. Unsupervised image translation with distributional semantics awareness
- Author
-
Zhexi Peng, He Wang, Yanlin Weng, Yin Yang, and Tianjia Shao
- Subjects
generative adversarial networks (GANs) ,manifold alignment ,unsupervised learning ,image-to-image translation ,distributional semantics ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
Abstract Unsupervised image translation (UIT) studies the mapping between two image domains. Since such mappings are under-constrained, existing research has pursued various desirable properties such as distributional matching or two-way consistency. In this paper, we re-examine UIT from a new perspective: distributional semantics consistency, based on the observation that data variations contain semantics, e.g., shoes varying in colors. Further, the semantics can be multi-dimensional, e.g., shoes also varying in style, functionality, etc. Given two image domains, matching these semantic dimensions during UIT will produce mappings with explicable correspondences, which has not been investigated previously. We propose distributional semantics mapping (DSM), the first UIT method which explicitly matches semantics between two domains. We show that distributional semantics has been rarely considered within and beyond UIT, even though it is a common problem in deep learning. We evaluate DSM on several benchmark datasets, demonstrating its general ability to capture distributional semantics. Extensive comparisons show that DSM not only produces explicable mappings, but also improves image quality in general.
- Published
- 2023
- Full Text
- View/download PDF
37. Similar-Sounding Words Flesh Out Fuzzy Meanings.
- Author
-
Haslett, David A. and Cai, Zhenguang G.
- Abstract
Most words are low in frequency, yet a prevailing theory of word meaning (the distributional hypothesis: that words with similar meanings occur in similar contexts) and corresponding computational models struggle to represent low-frequency words. We conducted two preregistered experiments to test the hypothesis that similar-sounding words flesh out deficient semantic representations. In Experiment 1, native English speakers made semantic relatedness decisions about a cue (e.g., dodge) followed either by a target that overlaps in form and meaning with a higher frequency word (evade, which overlaps with avoid) or by a control (elude), matched on distributional and formal similarity to the cue. (Participants did not see higher frequency words like avoid.) As predicted, participants decided faster and more often that overlapping targets, compared to controls, were semantically related to cues. In Experiment 2, participants read sentences containing the same cues and targets (e.g., The kids dodged something and She tried to evade/elude the officer). We used MouseView.js to blur the sentences and create a fovea-like aperture directed by the participant's cursor, allowing us to approximate fixation duration. While we did not observe the predicted difference at the target region (e.g., evade/elude), we found a lag effect, with shorter fixations on words following overlapping targets, suggesting easier integration of those meanings. These experiments provide evidence that words with overlapping forms and meanings bolster representations of low-frequency words, which supports approaches to natural language processing that incorporate both formal and distributional information and which revises assumptions about how an optimal language will evolve. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
38. No frills: Simple regularities in language can go a long way in the development of word knowledge.
- Author
-
Unger, Layla, Yim, Hyungwook, Savic, Olivera, Dennis, Simon, and Sloutsky, Vladimir M.
- Subjects
- *
NATURAL language processing , *CHILDREN'S language - Abstract
Recent years have seen a flourishing of Natural Language Processing models that can mimic many aspects of human language fluency. These models harness a simple, decades‐old idea: It is possible to learn a lot about word meanings just from exposure to language, because words similar in meaning are used in language in similar ways. The successes of these models raise the intriguing possibility that exposure to word use in language also shapes the word knowledge that children amass during development. However, this possibility is strongly challenged by the fact that models use language input and learning mechanisms that may be unavailable to children. Across three studies, we found that unrealistically complex input and learning mechanisms are unnecessary. Instead, simple regularities of word use in children's language input that they have the capacity to learn can foster knowledge about word meanings. Thus, exposure to language may play a simple but powerful role in children's growing word knowledge. A video abstract of this article can be viewed at https://youtu.be/dT83dmMffnM. Research Highlights: Natural Language Processing (NLP) models can learn that words are similar in meaning from higher‐order statistical regularities of word use.Unlike NLP models, infants and children may primarily learn only simple co‐occurrences between words.We show that infants' and children's language input is rich in simple co‐occurrence that can support learning similarities in meaning between words.We find that simple co‐occurrences can explain infants' and children's knowledge that words are similar in meaning. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
39. Applying Distributional Semantic Models to a Historical Corpus of a Highly Inflected Language: the Case of Ancient Greek.
- Author
-
Keersmaekers, Alek and Speelman, Dirk
- Subjects
LEXICAL access ,SEMANTICS ,POLYSEMY ,LANGUAGE models ,MODERN languages - Abstract
So-called "distributional" language models have become dominant in research on the computational modelling of lexical semantics. This paper investigates how well such models perform on Ancient Greek, a highly inflected historical language. It compares several ways of computing such distributional models on the basis of various context features (including both bag-of-words features and syntactic dependencies). The performance is assessed by evaluating how well these models are able to retrieve semantically similar words to a given target word, both on a benchmark we designed ourselves as well as on several independent benchmarks. It finds that dependency features are particularly useful to calculate distributional vectors for Ancient Greek (although the level of granularity that these dependency features should have is still open to discussion) and discusses possible ways for further improvement, including addressing problems related to polysemy and genre differences. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
40. Comparing Word Frequency, Semantic Diversity, and Semantic Distinctiveness in Lexical Organization.
- Author
-
Minyu Chang, Jones, Michael N., and Johns, Brendan T.
- Abstract
Word frequency (WF) is a strong predictor of lexical behavior. However, much research has shown that measures of contextual and semantic diversity offer a better account of lexical behaviors than WF (Adelman et al., 2006; Jones et al., 2012). In contrast to these previous studies, Chapman and Martin (2022) recently demonstrated that WF seems to account for distinct and greater levels of variance than measures of contextual and semantic diversity across a variety of datatypes. However, there are two limitations to these findings. The first is that Chapman and Martin (2022) compared variables derived from different corpora, which makes any conclusion about the theoretical advantage of one metric over another confounded, as it could be the construction of one corpus that provides the advantage and not the underlying theoretical construct. Second, they did not consider recent developments in the semantic distinctiveness model (SDM; Johns, 2021a; Johns et al., 2020; Johns & Jones, 2022). The current paper addressed the second limitation. Consistent with Chapman and Martin (2022), our results showed that the earliest versions of the SDM were less predictive of lexical data relative to WF when derived from a different corpus. However, the later versions of the SDM accounted for substantially more unique variance than WF in lexical decision and naming data. The results suggest that context-based accounts provide a better explanation of lexical organization than repetition-based accounts. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
41. Detecting translation borrowings in huge text collections using various methods.
- Author
-
Al-Janabi, Adel, Al-Zubaidi, Ehsan Ali, and Merzah, Baqer M.
- Subjects
NATURAL language processing ,MACHINE learning ,TRANSLATING & interpreting ,MACHINE translating ,COLLECTIONS ,AMBIGUITY ,DEEP learning - Abstract
The purpose of this work is to investigate the problem of detecting transportable borrowings and text reuse. The article proposes a monolingual solution to this problem: translating the suspicious material into language collections for additional monolingual analysis. One of the major requirements for the suggested technique is robustness against machine learning ambiguities. The next step in the document analysis is split into two parts. The authors begin by retrieving documents-candidates that are similarity to other types of text recurrence. The paper proposes retrieving texts utilizing word clusters formed using distributional semantic for robustness. In the second stage, the authors use deep learning neural networks to compare the suspected document to candidates utilizing phrase embedding. The experimentation is carried out for the language pair "English-Arabic" on both articles and synthetic data. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
42. Productivity from a Metapragmatic Perspective: Measuring the Diachronic Coverage of the Low Level Lexico-Grammatical Construction Have the N (Body Part/Attitude) to ↔<Metapragmatic Comment> Using the COHA.
- Author
-
Smith, Chris A.
- Subjects
PRAGMATICS ,GRAMMATICALIZATION ,SEMANTICS ,ONOMASIOLOGY ,ENGLISH language - Abstract
This paper seeks to address the relation between semantics, pragmatics and the productivity of a low level lexico-grammatical construction, Have the N (body part/attitude) to ↔metapragmatic comment. The question posed is how semantics affects productivity, in the generative sense of extensibility of a construction (a form meaning pairing). The method identifies the specificity and variations of the Have the N (body part/attitude) to ↔metapragmatic comment construction within the pragmeme of politeness using the COHA. Hereafter, we consider how to measure the extensibility within the onomasiological frame based on the available pool of forms expressing an attitude/emotion, i.e., the coverage or attractivity of the Have the N to construction. The paper discusses the findings, namely, how to overcome methodological issues relating to a qualitative rather than quantitative approach to the constructional architecture and the relative productivity of constructions. The experimental small scale corpus study of Have the N to in the COHA suggests that a global view of constructional architecture at multiple levels should be pertinent to identifying the extensibility potential of the construction. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
43. Adjacent and Non‐Adjacent Word Contexts Both Predict Age of Acquisition of English Words: A Distributional Corpus Analysis of Child‐Directed Speech
- Author
-
Chang, Lucas M and Deák, Gedeon O
- Subjects
Cognitive and Computational Psychology ,Psychology ,Pediatric ,Clinical Research ,Behavioral and Social Science ,Basic Behavioral and Social Science ,Child ,Preschool ,Comprehension ,Humans ,Infant ,Infant ,Newborn ,Language Development ,Learning ,Speech ,Vocabulary ,Language acquisition ,Language input ,Word learning ,Syntax ,Distributional semantics ,Age of acquisition ,Statistical learning ,Artificial Intelligence and Image Processing ,Cognitive Sciences ,Experimental Psychology ,Applied and developmental psychology ,Biological psychology ,Cognitive and computational psychology - Abstract
Children show a remarkable degree of consistency in learning some words earlier than others. What patterns of word usage predict variations among words in age of acquisition? We use distributional analysis of a naturalistic corpus of child-directed speech to create quantitative features representing natural variability in word contexts. We evaluate two sets of features: One set is generated from the distribution of words into frames defined by the two adjacent words. These features primarily encode syntactic aspects of word usage. The other set is generated from non-adjacent co-occurrences between words. These features encode complementary thematic aspects of word usage. Regression models using these distributional features to predict age of acquisition of 656 early-acquired English words indicate that both types of features improve predictions over simpler models based on frequency and appearance in salient or simple utterance contexts. Syntactic features were stronger predictors of children's production than comprehension, whereas thematic features were stronger predictors of comprehension. Overall, earlier acquisition was predicted by features representing frames that select for nouns and verbs, and by thematic content related to food and face-to-face play topics; later acquisition was predicted by features representing frames that select for pronouns and question words, and by content related to narratives and object play.
- Published
- 2020
44. Modelling brain activity associated with metaphor processing with distributionalsemantic models
- Author
-
Djokic, Vesna G. and Shutova, Ekaterina
- Subjects
metaphor ,abstraction ,distributional semantics - Abstract
In this study we investigate how lexical-semantic relations as-sociated with the literal meaning (and abstract meaning) arebeing accessed across the brain during familiar metaphor com-prehension. We utilize a data-driven whole-brain searchlightsimilarity-decoding analysis. We contrast decoding metaphoricphrases (”she’s grasping the idea”) using distributional seman-tic models of the verb in the phrase (VERB model) versus thatof the more abstract verb-sense (PARAPHRASE VERB model)obtained from literal paraphrases of the metaphoric phrases(”she’s understanding the idea”). We showed successful decod-ing with the VERB model across frontal, temporal and parietallobes mainly within areas of the language and default-modenetworks. In contrast, decoding with the PARAPHRASE VERBmodel was restricted to frontal-temporal lobes within areas ofthe language-network which overlapped to some extent withsignificant decoding with the VERB model. Overall, the re-sults suggest that lexical-semantic relations closely associatedwith the abstract meaning in metaphor processing are largelylocalized to language and amodal (multimodal) semantic mem-ory systems of the brain, while those more associated withthe literal meaning are processed across a distributed seman-tic network including areas implicated in mental imagery andsocial-cognition.
- Published
- 2020
45. Controlling the retrieval of general vs specific semantic knowledge in the instancetheory of semantic memory
- Author
-
Crump, Matthew J.C., Jamieson, Randall K., Johns, Brendan T., and Jones, Michael N.
- Subjects
distributional semantics ,higher-order similarity ,instance theory ,surprise-driven learning ,retrieval - Abstract
Distributional models of semantic cognition commonly makesimplifying assumptions, such as representing word co-occurrence structure by prototype-like high-dimensional se-mantic vectors, and limit how retrieval processes may con-tribute to the construction and use of semantic knowl-edge. More recently, the instance theory of semantics (ITS,Jamieson, Avery, Johns, & Jones, 2018) reconceived a dis-tributional model in terms of instance-based memory, allow-ing context-specific construction of semantic knowledge at thetime of retrieval. By simulation, we show that additional en-coding and retrieval operations, consistent with learning andmemory theory, can play a crucial role in flexibly controllingthe construction of general versus specific semantic knowl-edge. We argue this consolidation of processing principlesholds insight for distributional theories of semantic cognition.
- Published
- 2020
46. The Typology of Polysemy: A Multilingual Distributional Framework
- Author
-
Rabinovich, Ella, Xu, Yang, and Stevenson, Suzanne
- Subjects
Semantic typology ,cross-linguistic similarity ,word meaning ,distributional semantics ,multilingual wordembeddings - Abstract
Lexical semantic typology has identified important cross-linguistic generalizations about the variation and commonal-ities in polysemy patterns—how languages package up mean-ings into words. Recent computational research has enabledinvestigation of lexical semantics at a much larger scale, butlittle work has explored lexical typology across semantic do-mains, nor the factors that influence cross-linguistic similari-ties. We present a novel computational framework that quan-tifies semantic affinity, the cross-linguistic similarity of lexicalsemantics for a concept. Our approach defines a common mul-tilingual semantic space that enables a direct comparison of thelexical expression of concepts across languages. We validateour framework against empirical findings on lexical semantictypology at both the concept and domain levels. Our resultsreveal an intricate interaction between semantic domains andextra-linguistic factors, beyond language phylogeny, that co-shape the typology of polysemy across languages.
- Published
- 2020
47. Training vs Post-training Cross-lingual Word Embedding Approaches: A Comparative Study
- Author
-
Masood Ghayoomi
- Subjects
semantic word representation ,cross-lingual context ,vector space model ,distributional semantics ,Information resources (General) ,ZA3040-5185 ,Transportation and communications ,HE1-9990 - Abstract
This paper provides a comparative analysis of cross-lingual word embedding by studying the impact of different variables on the quality of the embedding models within the distributional semantics framework. Distributional semantics is a method for the semantic representation of words, phrases, sentences, and documents. This method aims at capturing as much information as possible from the contextual information in a vector space. The early study in this domain focused on monolingual word embedding. Further progress used cross-lingual data to capture the contextual semantic information across different languages. The main contribution of this research is to make a comparative study to find out the superior impact of the learning methods, supervised and unsupervised in training and post-training approaches in different embedding algorithms, to capture semantic properties of the words in cross-lingual embedding models to be applicable in tasks that deal with multi-languages, such as question retrieval. To this end, we study the cross-lingual embedding models created by BilBOWA, VecMap, and MUSE embedding algorithms along with the variables that impact the embedding models' quality, namely the size of the training data and the window size of the local context. In our study, we use the unsupervised monolingual Word2Vec embedding model as the baseline and evaluate the quality of embeddings on three data sets: Google analogy, mono- and cross-lingual words similar lists. We further investigated the impact of the embedding models in the question retrieval task.
- Published
- 2023
- Full Text
- View/download PDF
48. Distributional Semantics of Line Charts for Trend Classification
- Author
-
Onweller, Connor, O’Brien, Andrew, Kim, Edward, McCoy, Kathleen F., Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Bebis, George, editor, Li, Bo, editor, Yao, Angela, editor, Liu, Yang, editor, Duan, Ye, editor, Lau, Manfred, editor, Khadka, Rajiv, editor, Crisan, Ana, editor, and Chang, Remco, editor
- Published
- 2022
- Full Text
- View/download PDF
49. Machine learning methods for vector-based compositional semantics
- Author
-
Maillard, Jean, Clark, Stephen, and Vlachos, Andreas
- Subjects
natural language processing ,nlp ,computational linguistics ,compositionality ,distributional semantics ,compositional semantics - Abstract
Rich semantic representations of linguistic data are an essential component to the development of machine learning algorithms for natural language processing. This thesis explores techniques to model the meaning of phrases and sentences as dense vectors, which can then be further analysed and manipulated to perform any number of tasks involving the understanding of human language. Rather than seeing this task purely as an engineering problem, this thesis will focus on linguistically-motivated approaches, based on the principle of compositionality. The first half of the thesis will be dedicated to categorial compositional models, which are based on the observation that certain types of grammars share the structure of the algebra of vector spaces. This leads to an approach where the meanings of words are modelled as multilinear maps, encoded as tensors. In this framework, the meaning of a composite linguistic phrase can be computed via the tensor multiplication of its constituents, according to the phrase's syntactic structure. I contribute two categorial compositional models: the first, an extension of a popular method for learning semantic representation of words, models the meanings of adjective-noun phrases as matrix-vector multiplications; the second uses higher-order tensors to represent the meaning of relative clauses. In contrast, the models presented in the second half of the thesis do away with traditional syntactic structures. Rather than using the standard syntax trees of linguistics to drive the compositional process, these models treat the compositional structure as a latent variable. I contribute two models that automatically induce trees for a downstream task, without ever being shown a `real' syntax tree: one model based on chart parsing, and one based on shift-reduce parsing. While these proposed approaches induce trees that do not resemble traditional syntax trees, they do lead to models with higher performance on downstream tasks - opening up avenues for future research.
- Published
- 2019
- Full Text
- View/download PDF
50. A Distributional Semantic Online Lexicon for Linguistic Explorations of Societies.
- Author
-
Dahlberg, Stefan, Axelsson, Sofia, Gyllensten, Amaru Cuba, Sahlgren, Magnus, Ekgren, Ariel, Holmberg, Sören, and Schwarz, Jonas Andersson
- Subjects
- *
LEXICON , *NATURAL languages , *COMPUTATIONAL linguistics , *CRITICAL thinking , *POLITICAL attitudes , *NATURAL language processing , *ONLINE social networks - Abstract
Linguistic Explorations of Societies (LES) is an interdisciplinary research project with scholars from the fields of political science, computer science, and computational linguistics. The overarching ambition of LES has been to contribute to the survey-based comparative scholarship by compiling and analyzing online text data within and between languages and countries. To this end, the project has developed an online semantic lexicon, which allows researchers to explore meanings and usages of words in online media across a substantial number of geo-coded languages. The lexicon covers data from approximately 140 language–country combinations and is, to our knowledge, the most extensive free research resource of its kind. Such a resource makes it possible to critically examine survey translations and identify discrepancies in order to modify and improve existing survey methodology, and its unique features further enable Internet researchers to study public debate online from a comparative perspective. In this article, we discuss the social scientific rationale for using online text data as a complement to survey data, and present the natural language processing–based methodology behind the lexicon including its underpinning theory and practical modeling. Finally, we engage in a critical reflection about the challenges of using online text data to gauge public opinion and political behavior across the world. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.