To see the other types of publications on this topic, follow the link: Semantic structure of the word.

Dissertations / Theses on the topic 'Semantic structure of the word'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Semantic structure of the word.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Tsiplakou, Stavroula. "Focus in Greek : its structure and interpretation." Thesis, SOAS, University of London, 1999. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.313433.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Karmakar, Saurav. "Syntactic and Semantic Analysis and Visualization of Unstructured English Texts." Digital Archive @ GSU, 2011. http://digitalarchive.gsu.edu/cs_diss/61.

Full text
Abstract:
People have complex thoughts, and they often express their thoughts with complex sentences using natural languages. This complexity may facilitate efficient communications among the audience with the same knowledge base. But on the other hand, for a different or new audience this composition becomes cumbersome to understand and analyze. Analysis of such compositions using syntactic or semantic measures is a challenging job and defines the base step for natural language processing. In this dissertation I explore and propose a number of new techniques to analyze and visualize the syntactic and semantic patterns of unstructured English texts. The syntactic analysis is done through a proposed visualization technique which categorizes and compares different English compositions based on their different reading complexity metrics. For the semantic analysis I use Latent Semantic Analysis (LSA) to analyze the hidden patterns in complex compositions. I have used this technique to analyze comments from a social visualization web site for detecting the irrelevant ones (e.g., spam). The patterns of collaborations are also studied through statistical analysis. Word sense disambiguation is used to figure out the correct sense of a word in a sentence or composition. Using textual similarity measure, based on the different word similarity measures and word sense disambiguation on collaborative text snippets from social collaborative environment, reveals a direction to untie the knots of complex hidden patterns of collaboration.
APA, Harvard, Vancouver, ISO, and other styles
3

Morbiato, Anna. "Word order and sentence structure in Mandarin Chinese: new perspectives." Doctoral thesis, Università Ca' Foscari Venezia, The University of Sydney, 2018. http://hdl.handle.net/10278/3716543.

Full text
Abstract:
Word order (WO) is one of the most fascinating and investigated topics in Mandarin Chinese (MC) linguistics, and many accounts have been proposed on different WO patterns and constructions. However, despite the large amount of research, several WO related issues remain rather controversial. Crucially, no unified consensus exists on the relationship between WO and the different dimensions of the language (i.e. semantics, syntax and pragmatics), and on how these levels interact with each other. The present thesis’s aim is twofold: (1) identify the categories that are useful to account for WO patterns and variations in MC; (2) examine in greater depth the syntactic, semantic and pragmatic factors that influence word order in MC, as well as how they interact and impose constraints on possible WO variations. The novelty of the approach lies on three aspects: (i) a typological, comparative perspective that benefits from cross-linguistic investigation of WO phenomena in other languages; (ii) a bottom up approach that employs cross-linguistically validated typological tools (e.g., GR tests, or constituenthood tests) aimed at conducting the analysis on a language-internal basis, and (iii) an empirical approach: the analysis avails itself of natural linguistic data, mainly drawn from corpora, and relies on acceptability checks with native speakers. Overall, the thesis highlights that WO patterns and constructions are determined by the interplay of different factors and constraints. It also highlights that, for the sake of clarity and ambiguity avoidance, WO constraints are hierarchically organized, and WO freezing phenomena occur to allow disambiguation of participants in the described event.
APA, Harvard, Vancouver, ISO, and other styles
4

Koivisto-Alanko, Päivi. "Abstract words in abstract worlds : directionality and prototypical structure in the semantic change in English nouns of cognition /." Helsinki : Société néophilologique, 2000. http://catalogue.bnf.fr/ark:/12148/cb392874530.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Romeo, Lauren Michele. "The Structure of the lexicon in the task of the automatic acquisition of lexical information." Doctoral thesis, Universitat Pompeu Fabra, 2015. http://hdl.handle.net/10803/325420.

Full text
Abstract:
La información de clase semántica de los nombres es fundamental para una amplia variedad de tareas del procesamiento del lenguaje natural (PLN), como la traducción automática, la discriminación de referentes en tareas como la detección y el seguimiento de eventos, la búsqueda de respuestas, el reconocimiento y la clasificación de nombres de entidades, la construcción y ampliación automática de ontologías, la inferencia textual, etc. Una aproximación para resolver la construcción y el mantenimiento de los léxicos de gran cobertura que alimentan los sistemas de PNL, una tarea muy costosa y lenta, es la adquisición automática de información léxica, que consiste en la inducción de una clase semántica relacionada con una palabra en concreto a partir de datos de su distribución obtenidos de un corpus. Precisamente, por esta razón, se espera que la investigación actual sobre los métodos para la producción automática de léxicos de alta calidad, con gran cantidad de información y con anotación de clase como el trabajo que aquí presentamos, tenga un gran impacto en el rendimiento de la mayoría de las aplicaciones de PNL. En esta tesis, tratamos la adquisición automática de información léxica como un problema de clasificación. Con este propósito, adoptamos métodos de aprendizaje automático para generar un modelo que represente los datos de distribución vectorial que, basados en ejemplos conocidos, permitan hacer predicciones de otras palabras desconocidas. Las principales preguntas de investigación que planteamos en esta tesis son: (i) si los datos de corpus proporcionan suficiente información para construir representaciones de palabras de forma eficiente y que resulten en decisiones de clasificación precisas y sólidas, y (ii) si la adquisición automática puede gestionar, también, los nombres polisémicos. Para hacer frente a estos problemas, realizamos una serie de validaciones empíricas sobre nombres en inglés. Nuestros resultados confirman que la información obtenida a partir de la distribución de los datos de corpus es suficiente para adquirir automáticamente clases semánticas, como lo demuestra un valor-F global promedio de 0,80 aproximadamente utilizando varios modelos de recuento de contextos y en datos de corpus de distintos tamaños. No obstante, tanto el estado de la cuestión como los experimentos que realizamos destacaron una serie de retos para este tipo de modelos, que son reducir la escasez de datos del vector y dar cuenta de la polisemia nominal en las representaciones distribucionales de las palabras. En este contexto, los modelos de word embedding (WE) mantienen la “semántica” subyacente en las ocurrencias de un nombre en los datos de corpus asignándole un vector. Con esta elección, hemos sido capaces de superar el problema de la escasez de datos, como lo demuestra un valor-F general promedio de 0,91 para las clases semánticas de nombres de sentido único, a través de una combinación de la reducción de la dimensionalidad y de números reales. Además, las representaciones de WE obtuvieron un rendimiento superior en la gestión de las ocurrencias asimétricas de cada sentido de los nombres de tipo complejo polisémicos regulares en datos de corpus. Como resultado, hemos podido clasificar directamente esos nombres en su propia clase semántica con un valor-F global promedio de 0,85. La principal aportación de esta tesis consiste en una validación empírica de diferentes representaciones de distribución utilizadas para la clasificación semántica de nombres junto con una posterior expansión del trabajo anterior, lo que se traduce en recursos léxicos y conjuntos de datos innovadores que están disponibles de forma gratuita para su descarga y uso.<br>La información de clase semántica de los nombres es fundamental para una amplia variedad de tareas del procesamiento del lenguaje natural (PLN), como la traducción automática, la discriminación de referentes en tareas como la detección y el seguimiento de eventos, la búsqueda de respuestas, el reconocimiento y la clasificación de nombres de entidades, la construcción y ampliación automática de ontologías, la inferencia textual, etc. Una aproximación para resolver la construcción y el mantenimiento de los léxicos de gran cobertura que alimentan los sistemas de PNL, una tarea muy costosa y lenta, es la adquisición automática de información léxica, que consiste en la inducción de una clase semántica relacionada con una palabra en concreto a partir de datos de su distribución obtenidos de un corpus. Precisamente, por esta razón, se espera que la investigación actual sobre los métodos para la producción automática de léxicos de alta calidad, con gran cantidad de información y con anotación de clase como el trabajo que aquí presentamos, tenga un gran impacto en el rendimiento de la mayoría de las aplicaciones de PNL. En esta tesis, tratamos la adquisición automática de información léxica como un problema de clasificación. Con este propósito, adoptamos métodos de aprendizaje automático para generar un modelo que represente los datos de distribución vectorial que, basados en ejemplos conocidos, permitan hacer predicciones de otras palabras desconocidas. Las principales preguntas de investigación que planteamos en esta tesis son: (i) si los datos de corpus proporcionan suficiente información para construir representaciones de palabras de forma eficiente y que resulten en decisiones de clasificación precisas y sólidas, y (ii) si la adquisición automática puede gestionar, también, los nombres polisémicos. Para hacer frente a estos problemas, realizamos una serie de validaciones empíricas sobre nombres en inglés. Nuestros resultados confirman que la información obtenida a partir de la distribución de los datos de corpus es suficiente para adquirir automáticamente clases semánticas, como lo demuestra un valor-F global promedio de 0,80 aproximadamente utilizando varios modelos de recuento de contextos y en datos de corpus de distintos tamaños. No obstante, tanto el estado de la cuestión como los experimentos que realizamos destacaron una serie de retos para este tipo de modelos, que son reducir la escasez de datos del vector y dar cuenta de la polisemia nominal en las representaciones distribucionales de las palabras. En este contexto, los modelos de word embedding (WE) mantienen la “semántica” subyacente en las ocurrencias de un nombre en los datos de corpus asignándole un vector. Con esta elección, hemos sido capaces de superar el problema de la escasez de datos, como lo demuestra un valor-F general promedio de 0,91 para las clases semánticas de nombres de sentido único, a través de una combinación de la reducción de la dimensionalidad y de números reales. Además, las representaciones de WE obtuvieron un rendimiento superior en la gestión de las ocurrencias asimétricas de cada sentido de los nombres de tipo complejo polisémicos regulares en datos de corpus. Como resultado, hemos podido clasificar directamente esos nombres en su propia clase semántica con un valor-F global promedio de 0,85. La principal aportación de esta tesis consiste en una validación empírica de diferentes representaciones de distribución utilizadas para la clasificación semántica de nombres junto con una posterior expansión del trabajo anterior, lo que se traduce en recursos léxicos y conjuntos de datos innovadores que están disponibles de forma gratuita para su descarga y uso.<br>Lexical semantic class information for nouns is critical for a broad variety of Natural Language Processing (NLP) tasks including, but not limited to, machine translation, discrimination of referents in tasks such as event detection and tracking, question answering, named entity recognition and classification, automatic construction and extension of ontologies, textual inference, etc. One approach to solve the costly and time-consuming manual construction and maintenance of large-coverage lexica to feed NLP systems is the Automatic Acquisition of Lexical Information, which involves the induction of a semantic class related to a particular word from distributional data gathered within a corpus. This is precisely why current research on methods for the automatic production of high- quality information-rich class-annotated lexica, such as the work presented here, is expected to have a high impact on the performance of most NLP applications. In this thesis, we address the automatic acquisition of lexical information as a classification problem. For this reason, we adopt machine learning methods to generate a model representing vectorial distributional data which, grounded on known examples, allows for the predictions of other unknown words. The main research questions we investigate in this thesis are: (i) whether corpus data provides sufficient distributional information to build efficient word representations that result in accurate and robust classification decisions and (ii) whether automatic acquisition can handle also polysemous nouns. To tackle these problems, we conducted a number of empirical validations on English nouns. Our results confirmed that the distributional information obtained from corpus data is indeed sufficient to automatically acquire lexical semantic classes, demonstrated by an average overall F1-Score of almost 0.80 using diverse count-context models and on different sized corpus data. Nonetheless, both the State of the Art and the experiments we conducted highlighted a number of challenges of this type of model such as reducing vector sparsity and accounting for nominal polysemy in distributional word representations. In this context, Word Embeddings (WE) models maintain the “semantics” underlying the occurrences of a noun in corpus data by mapping it to a feature vector. With this choice, we were able to overcome the sparse data problem, demonstrated by an average overall F1-Score of 0.91 for single-sense lexical semantic noun classes, through a combination of reduced dimensionality and “real” numbers. In addition, the WE representations obtained a higher performance in handling the asymmetrical occurrences of each sense of regular polysemous complex-type nouns in corpus data. As a result, we were able to directly classify such nouns into their own lexical-semantic class with an average overall F1-Score of 0.85. The main contribution of this dissertation consists of an empirical validation of different distributional representations used for nominal lexical semantic classification along with a subsequent expansion of previous work, which results in novel lexical resources and data sets that have been made freely available for download and use.
APA, Harvard, Vancouver, ISO, and other styles
6

Shebani, Zubaida Soliman. "Semantic word category processing." Thesis, University of Cambridge, 2012. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.610751.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Isheden, Gabriel. "Diagnostics of Semantic Word Spaces." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-142614.

Full text
Abstract:
In collaboration with Gavagai, a company that develops automated and scalable methods for retrieving actionable intelligence from dynamic data, I have been studying semantic word spaces and topology. In this bachelor’s thesis, with help from computational topology, I introduce new ways to describe properties of these semantic word spaces, so called barcodes. I develop a measure to describe barcodes of betti number zero, prove its validity and discuss its implications.
APA, Harvard, Vancouver, ISO, and other styles
8

Westum, Asbjörg. "Ris, skäver och skärva : Folklig kategorisering av några barnsjukdomar ur ett kognitivt semantiskt perspektiv." Doctoral thesis, Umeå universitet, Litteraturvetenskap och nordiska språk, 1999. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-51634.

Full text
Abstract:
In Swedish dialects we find the terms ris,skäver and skärva referring to illnesses in children. The words are also parts of various compounds which refer to variants of the illnesses. The terms are linguistic expressions denoting two folk categories of illnesses, RIS and SKÄVER/SKÄRVA. These categories are investigated from a cognitive semantic perspective. The cognitive perspective argues that we organize our understanding of reality by using Idealized Cognitive Models (ICM) based on our physical, mental and emotional experiences of the world. The aim is twofold: to demonstrate the bases on which an experienced illness is placed in a certain category, and to show how a folk conception of illness is reflected in the word formation strategies. The word formation strategies emanate from notions of characteristic symptoms, and from notions of causes of illnesses. Both categories, RIS and SKÄVER/SKÄRVA, are based on a number of ICM's. The category RIS is a radial structure, which means that the category is held together although its members have no structural criteria in common. The category SKÄVER/SKÄRVA is a concentrating structure, meaning that all members share all structural criteria. There is a strong connection between word formation strategies and the structures of the categories. Terms related to symptoms refer to members of a category which are part of a radial structure, while terms related to causes refer to members of a category which are part of a concentrating structure. This can be explained by two of the basic assumptions of cognitive semantics: semantic content is structred and symbolized overtly on the surface form of a language and categories are conventional, based on cultural assumptions about the world.<br>digitalisering@umu
APA, Harvard, Vancouver, ISO, and other styles
9

Fu, Hao. "Semantic image understanding : from pixel to word." Thesis, University of Nottingham, 2012. http://eprints.nottingham.ac.uk/12847/.

Full text
Abstract:
The aim of semantic image understanding is to reveal the semantic meaning behind the image pixel. This thesis investigates problems related to semantic image understanding, and have made the following contributions. Our first contribution is to propose the usage of histogram matching in Multiple Kernel Learning. We treat the two-dimensional kernel matrix as an image and transfer the histogram matching algorithm in image processing to kernel matrix. Experiments on various computer vision and machine learning datasets have shown that our method can always boost the performance of state of the art MKL methods. Our second contribution is to advocate the segment-then-recognize strategy in pixel-level semantic image understanding. We have developed a new framework which tries to integrate semantic segmentation with low-level segmentation for proposing object consistent regions. We have also developed a novel method trying to integrate semantic segmentation with interactive segmentation. We found this segment-then-recognize strategy also works well on medical image data, where we designed a novel polar space random field model for proposing gland-like regions. In the realm of image-level semantic image understanding, our contribution is a novel way to utilize the random forest. Most of the previous works utilizing random forest store the posterior probabilities at each leaf node, and each random tree in the random forest is considered to be independent from each other. In contrast, we store the training samples instead of the posterior probabilities at each leaf node. We consider the random forest as a whole and propose the concept of semantic nearest neighbor and semantic similarity measure. Based on these two concepts, we devise novel methods for image annotation and image retrieval tasks.
APA, Harvard, Vancouver, ISO, and other styles
10

Rabovsky, Milena. "Semantic richness effects in visual word processing." Doctoral thesis, Humboldt-Universität zu Berlin, Lebenswissenschaftliche Fakultät, 2014. http://dx.doi.org/10.18452/17073.

Full text
Abstract:
Lesen zielt darauf ab, Bedeutung aus geschriebenem Text zu extrahieren. Interessanterweise unterscheiden sich Wörter beträchtlich hinsichtlich der Menge mit ihnen assoziierter Bedeutung, und es wurde kürzlich gezeigt, dass eine hohe Bedeutungshaltigkeit lexikalische und semantische Aufgaben erleichtert. Die vorliegende Dissertation kombiniert ereigniskorrelierte Potentiale (EKPs) und konnektionistische Modellierung, um einige offene Fragen zur Rolle der Bedeutungshaltigkeit bei der Wortverarbeitung anzugehen. Hierbei wurden EKPs verwendet, um den Zeitverlauf unabhängiger Einflüsse der Anzahl semantischer Merkmale und Assoziationen beim Wortlesen zu bestimmen sowie Einflüsse von Bedeutungshaltigkeit auf implizites Wortlernen zu untersuchen. Um die zugrundeliegenden Mechanismen besser zu verstehen, wurden die Ergebnisse anschließend mittels eines semantischen Netzwerk-Modells simuliert. Es zeigten sich keine Einflüsse der Anzahl der Assoziationen, aber eine schnelle Aktivierung semantischer Merkmale, die das EKP bereits ab 190 ms beeinflussten - nur 20 bis 30 ms nach und zeitlich überlappend mit der Aktivierung orthographischer Repräsentationen, die durch N1-Lexikalitätseffekte angezeigt wurden. Im weiteren Verlauf ging eine hohe Merkmalsanzahl mit größeren N400-Amplituden einher. Zudem verstärkten semantische Merkmale Wiederholungseinflüsse auf die Akkuratheit lexikalischer Entscheidungen und N400-Amplituden, was einen ersten Hinweis auf Einflüsse von Bedeutungshaltigkeit auf implizites Wortlernen darstellt. Diese Ergebnisse stehen im Einklang mit merkmalsbasierten semantischen Netzwerk-Modellen. Simulationen legen nahe, dass semantische Aktivierung lexikalische Entscheidungen erleichtert, während Netzwerk-Fehler in engem Zusammenhang mit N400-Amplituden stehen. Da Netzwerk-Fehler psychologisch als implizite Vorhersagefehler interpretiert werden, deuten diese Ergebnisse darauf hin, dass N400-Amplituden implizite Vorhersagefehler im semantischen System widerspiegeln.<br>Language ultimately aims to convey meaning. Importantly, the amount of associated semantic information varies considerably between words. Recent evidence suggests that the richness of semantic representations facilitates performance in lexical and semantic tasks, but much remains to be learned about semantic richness effects. The present dissertation combined event-related brain potentials (ERPs) and connectionist modeling to address several unresolved issues concerning the role of semantic richness in word processing. Specifically, ERPs were employed to investigate the time course of independent influences of the number of semantic features and associates during word reading (study 1) and influences of semantic richness on implicit word learning (study 2). Aiming at advancing a mechanistic understanding of the obtained results, both studies were subsequently simulated using a network model of semantic cognition (study 3). Results showed no influences of the number of associates, but fast access to semantic features, with influences of feature-based semantic richness starting at about 190 ms - a mere 20 to 30 ms after and temporally overlapping with the activation of orthographic representations as reflected by N1 lexicality effects. Later on, a high number of semantic features induced larger N400 amplitudes. Furthermore, the number of semantic features enhanced repetition priming effects on lexical decision accuracy and N400 amplitudes, providing initial evidence for influences of semantic richness on implicit word learning. These results are in line with feature-based network models of semantic cognition. Simulations with such a model suggest that semantic activation can facilitate lexical decisions, while network error closely corresponds to N400 amplitudes. In psychological terms, network error has been conceptualized as implicit prediction error. Thus, these results are taken to suggest that N400 amplitudes reflect implicit prediction error in semantic memory.
APA, Harvard, Vancouver, ISO, and other styles
11

Lyalka, Oksana. "Mechanisms underpinning semantic priming in spoken word retrieval." Thesis, University of Newcastle upon Tyne, 2017. http://hdl.handle.net/10443/3716.

Full text
Abstract:
A number of studies have shown that speed and accuracy of word retrieval may be affected by the previous retrieval of a word with similar semantic meaning. This phenomenon is called semantic priming and includes both semantic interference or and facilitation. While there is a clear evidence for the presence of semantic priming, the mechanisms causing this effect are still under debate. Therefore, the goal of this PhD was to provide evidence regarding these mechanisms by systematically evaluating the effect of primes with different semantic relations on the speed and accuracy of spoken word retrieval in healthy subjects and people with aphasia. Five experiments were implemented with healthy participants focusing on the effects in priming of semantic coordination, association and part-whole relations on spoken word retrieval with zero or four intervening items between prime and target (lags 0 and 4). Chapter Two reports two experiments using an alternating word reading and picture naming paradigm and Chapter Three, three experiments using a continuous picture naming paradigm. Chapter Four reports two experiments with people with aphasia examining the effects of identity, semantic coordination, association and their interaction on facilitation of picture naming. The results of these two experiments were analysed at both individual subject and group levels. In Chapter Five, these experiments are placed in the context of the previous literature on semantic priming and theories of semantic representation. In this regard, the experimental results are taken to imply that semantic coordination, association, and part-whole relations can be attributed to different types of semantic relations that have different representation and organisation. Further implications of the experiments for our understanding of the mechanisms of lexical access and the nature of lexical representation are discussed.
APA, Harvard, Vancouver, ISO, and other styles
12

Eskenazi, Michael A. "Word skipping and parafoveal semantic activation during reading." Kent State University / OhioLINK, 2016. http://rave.ohiolink.edu/etdc/view?acc_num=kent1467377665.

Full text
APA, Harvard, Vancouver, ISO, and other styles
13

Montariol, Syrielle. "Models of diachronic semantic change using word embeddings." Electronic Thesis or Diss., université Paris-Saclay, 2021. http://www.theses.fr/2021UPASG006.

Full text
Abstract:
Dans cette thèse, nous étudions les changements lexico-sémantiques : les variations temporelles dans l'usage et la signification des mots, également appelé extit{diachronie}. Ces changements reflètent l'évolution de divers aspects de la société tels que l'environnement technologique et culturel.Nous explorons et évaluons des méthodes de construction de plongements lexicaux variant dans le temps afin d'analyser l'évolution du language. Nous utilisont notamment des plongements contextualisés à partir de modèles de langue pré-entraînés tels que BERT.Nous proposons plusieurs approches pour extraire et agréger les représentations contextualisées des mots dans le temps, et quantifier leur degré de changement sémantique. En particulier, nous abordons l'aspect pratique de ces systèmes: le passage à l'échelle de nos approches, en vue de les appliquer à de grands corpus ou de larges vocabulaire; leur interprétabilité, en désambiguïsant les différents usages d'un mot au cours du temps; et leur applicabilité à des problématiques concrètes, pour des documents liés au COVID19 et des corpus du domaine financier. Nous évaluons l'efficacité de ces méthodes de manière quantitative, en utilisant plusieurs corpus annotés, et de manière qualitative, en liant les variations détectées dans des corpus avec des événements de la vie réelle et des données numériques.Enfin, nous étendons la tâche de détection de changements sémantiques au-delà de la dimension temporelle. Nous l'adaptons à un cadre bilingue, pour étudier l'évolution conjointe d'un mot et sa traduction dans deux corpus de langues différentes; et à un cadre synchronique, pour détecter des variations sémantiques entre différentes sources ou communautés en plus de la variation temporelle<br>In this thesis, we study lexical semantic change: temporal variations in the use and meaning of words, also called extit{diachrony}. These changes are carried by the way people use words, and mirror the evolution of various aspects of society such as its technological and cultural environment.We explore, compare and evaluate methods to build time-varying embeddings from a corpus in order to analyse language evolution.We focus on contextualised word embeddings using pre-trained language models such as BERT. We propose several approaches to extract and aggregate the contextualised representations of words over time, and quantify their level of semantic change.In particular, we address the practical aspect of these systems: the scalability of our approaches, with a view to applying them to large corpora or large vocabularies; their interpretability, by disambiguating the different uses of a word over time; and their applicability to concrete issues, for documents related to COVID19We evaluate the efficiency of these methods quantitatively using several annotated corpora, and qualitatively by linking the detected semantic variations with real-life events and numerical data.Finally, we extend the task of semantic change detection beyond the temporal dimension. We adapt it to a bilingual setting, to study the joint evolution of a word and its translation in two corpora of different languages; and to a synchronic frame, to detect semantic variations across different sources or communities on top of the temporal variation
APA, Harvard, Vancouver, ISO, and other styles
14

Courbon, Bruno. "Étude sémantique des mots "chance", "fortune", "hasard" et "risque" du XVIIIe au XXIe siècle : perspectives sur le lexique du français et ses usages." Thesis, Lyon 2, 2009. http://www.theses.fr/2009LYO20041.

Full text
Abstract:
La recherche a pour objet la structuration du champ lexical des mots « chance », « fortune », « hasard » et « risque » du XVIIIe au XXIe siècle. Témoin de mutations qu’a connues la civilisation occidentale durant cette période, ce champ, qui se rattache à la notion de fortune / hasard, présente une relative homogénéité sémantique.Les mots (et leurs dérivés) sont étudiés à travers le déploiement, la régulation et la répartition des normes d’usages, non seulement en français hexagonal, mais aussi en français québécois. L’étude se fonde sur l’exploitation de deux types de corpus. D’une part, un corpus d’articles extraits d’une cinquantaine de dictionnaires sert à mettre en évidence la productivité morphosémantique et sémantique de ces unités dans une perspective historique large. D’autre part, un grand ensemble d’énoncés diversifiés permet, par la mise au jour de types de contextes, d’effectuer un suivi diachronique des usages. L’approche continuiste des différences d’usages s’appuie sur une représentation fréquentielle des changements sémantiques.La thèse apporte une contribution à la question de la variation des usages et du changement sémantique, qui ouvre sur plusieurs perspectives. Elle se veut d’abord une réflexion sur la théorie et la méthodologie descriptives, appréhendées à la lumière de l’analyse de la nature et du rôle des corpus. Elle met ensuite en évidence l’importance de la dimension intersubjective dans l’activité de signification, en particulier le rôle déterminant des structures syntagmatiques dans l’établissement de nouveaux usages sémantiques. Enfin, elle permet de mettre en relation le changement sémantique avec les conditions sociohistoriques et les représentations collectives<br>The present study deals with the way in which the lexical field regrouping the words “chance”, “fortune”, “hazard” and “risqué” has been structured in the French language from the eighteenth century till the present day. Revealing major changes in western societies during this period of time, the field, which corresponds to the linguistic representation of the notion of fortune / hasard, presents a certain coherence.We have examined these words and their derived forms through the display, regulation, and distribution of norms of use, not only in Hexagonal French, but also in Quebec French. Two types of corpora have been analysed. On the one hand, a corpus of articles from around 50 dictionaries has been used to emphasize the lexical and semantic productivity of the different units on a large historical scale. On the other hand, in revealing context types, a set of texts reflecting French language varieties has allowed for carrying out a diachronic analysis of lexical uses. The continuist approach to semantic differences rests upon a frequential representation of semantic changes.The thesis brings a significant contribution to the question of usage variations and semantic change, providing new perspectives. It first deals with theory and methodology of lexical description, considered through the analysis of the nature and the role of corpora. It then evidences the central role of syntagmatic structures in the setting of new semantic uses. The study has finally put into relation semantic changes with their historical background and the collective representations of the time
APA, Harvard, Vancouver, ISO, and other styles
15

Jonsson, Erik. "Semantic word classification and temporaldependency detection on cooking recipes." Thesis, Linköpings universitet, Interaktiva och kognitiva system, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-122966.

Full text
Abstract:
This thesis presents an approach to automatically infer facts about cooking recipes.The analysis focuses on two aspects: recognizing words with special semantic meaningin a recipe such as ingredients and tools; and detecting temporal dependencies betweensteps in the cooking procedure. This is done by use of machine learning classicationand several dierent algorithms are benchmarked against each other.
APA, Harvard, Vancouver, ISO, and other styles
16

Morris, Julie. "Word deafness : a comparison of auditory and semantic treatments." Thesis, University of York, 1997. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.242176.

Full text
APA, Harvard, Vancouver, ISO, and other styles
17

Moss, Adam. "Detecting Lexical Semantic Change Using Probabilistic Gaussian Word Embeddings." Thesis, Uppsala universitet, Institutionen för lingvistik och filologi, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-412539.

Full text
Abstract:
In this work, we test two novel methods of using word embeddings to detect lexical semantic change, attempting to overcome limitations associated with conventional approaches to this problem. Using a diachronic corpus spanning over a hundred years, we generate word embeddings for each decade with the intention of evaluating how meaning changes are represented in embeddings for the same word across time. Our approach differs from previous works in this field in that we encode words as probabilistic Gaussian distributions and bimodal probabilistic Gaussian mixtures, rather than conventional word vectors. We provide a discussion and analysis of our results, comparing the approaches we implemented with those used in previous works. We also conducted further analysis on whether additional information regarding the nature of semantic change could be discerned from particular qualities of the embeddings we generated for our experiments. In our results, we find that encoding words as probabilistic Gaussian embeddings can provide an enhanced degree of reliability with regard to detecting lexical semantic change. Furthermore, we are able to represent additional information regarding the nature of such changes through the variance of these embeddings. Encoding words as bimodal Gaussian mixtures however is generally unsuccessful for this task, proving to be not reliable enough at distinguishing between discrete senses to effectively detect and measure such changes. We provide potential explanations for the results we observe, and propose improvements that can be made to our approach to potentially improve performance.
APA, Harvard, Vancouver, ISO, and other styles
18

Christofaris, Lynne D. "Semantic and Phonological Relationships to Word Retrieval during Aging." Miami University / OhioLINK, 2008. http://rave.ohiolink.edu/etdc/view?acc_num=miami1209768680.

Full text
APA, Harvard, Vancouver, ISO, and other styles
19

Baker, Brett Joseph. "Word Structure in Ngalakgan." University of Sydney, Linguistics, 1999. http://hdl.handle.net/2123/408.

Full text
Abstract:
Ngalakgan is an Australian language of the Gunwinyguan family, spoken fluently by just a few people in the mid Roper River area of the Top End. The thesis is a description and examination of the phonology, prosody, and morphology of Ngalakgan, based on several years of fieldwork. Ngalakgan is a language with a rich inventory of classically Gunwinyguan morphological features, including noun class agreement for all major and some minor word classes, compounding of both nouns and verbs, and a rich array of modifying and inflectional prefixes and suffixes. In Ngalakgan, there is a distinction between two kinds or 'levels' of morphology: 'root'-level and 'word'-level. Root-level morphology is lexicalised and unproductive. It is restricted to the tense/aspect/mood inflection of the small closed class of 'finite' verb roots, and to the large closed class of compounds of these roots. Word-level morphology is productive, and includes almost all prefixes, all (non-tensed) suffixes and all clitics. Only word-level structure is consistently reflected in prosodic structure; forms which are complex only at the root-level are treated as prosodic units. I show that all word-level morphemes constitute prosodic domains: every word-level stem, affix and clitic potentially begins a new domain for metrical foot structure. Geminates and glottal stops are over-represented at morpheme boundaries in complex words. In addition, they are subject to complex, non-local alternations with simple stops and zero, respectively, in Ngalakgan and related languages. The alternations are conditioned by preceding geminates and voiceless obstruent clusters, as well as by prosodic and morphological structure. I propose that voiceless obstruent clusters constitute 'boundary signals' to morphological structure, in a similar fashion to stress and, like stress, are 'licensed' by the organisation of intonation. Ngalakgan displays a quantitive-sensitive stress system in roots which is apparently unique to languages of this area. Heavy syllables in Ngalakgan are those which are articulatorily and perceptually complex: those in which the coda is followed by a consonant with a distinct place of articulation. Geminates, homorganic nasal+stop clusters and glottal stops interact with this distinction in ways which are not predicted by current prosodic theories.
APA, Harvard, Vancouver, ISO, and other styles
20

Baker, Brett. "Word Structure in Ngalakgan." Thesis, The University of Sydney, 1999. http://hdl.handle.net/2123/408.

Full text
Abstract:
Ngalakgan is an Australian language of the Gunwinyguan family, spoken fluently by just a few people in the mid Roper River area of the Top End. The thesis is a description and examination of the phonology, prosody, and morphology of Ngalakgan, based on several years of fieldwork. Ngalakgan is a language with a rich inventory of classically Gunwinyguan morphological features, including noun class agreement for all major and some minor word classes, compounding of both nouns and verbs, and a rich array of modifying and inflectional prefixes and suffixes. In Ngalakgan, there is a distinction between two kinds or 'levels' of morphology: 'root'-level and 'word'-level. Root-level morphology is lexicalised and unproductive. It is restricted to the tense/aspect/mood inflection of the small closed class of 'finite' verb roots, and to the large closed class of compounds of these roots. Word-level morphology is productive, and includes almost all prefixes, all (non-tensed) suffixes and all clitics. Only word-level structure is consistently reflected in prosodic structure; forms which are complex only at the root-level are treated as prosodic units. I show that all word-level morphemes constitute prosodic domains: every word-level stem, affix and clitic potentially begins a new domain for metrical foot structure. Geminates and glottal stops are over-represented at morpheme boundaries in complex words. In addition, they are subject to complex, non-local alternations with simple stops and zero, respectively, in Ngalakgan and related languages. The alternations are conditioned by preceding geminates and voiceless obstruent clusters, as well as by prosodic and morphological structure. I propose that voiceless obstruent clusters constitute 'boundary signals' to morphological structure, in a similar fashion to stress and, like stress, are 'licensed' by the organisation of intonation. Ngalakgan displays a quantitive-sensitive stress system in roots which is apparently unique to languages of this area. Heavy syllables in Ngalakgan are those which are articulatorily and perceptually complex: those in which the coda is followed by a consonant with a distinct place of articulation. Geminates, homorganic nasal+stop clusters and glottal stops interact with this distinction in ways which are not predicted by current prosodic theories.
APA, Harvard, Vancouver, ISO, and other styles
21

Bloom, Paul 1963. "Semantic structure and language development." Thesis, Massachusetts Institute of Technology, 1990. http://hdl.handle.net/1721.1/13686.

Full text
APA, Harvard, Vancouver, ISO, and other styles
22

McNeill, Allan. "Semantic structure of personal information." Thesis, Connect to e-thesis, 2002. http://theses.gla.ac.uk/840/.

Full text
Abstract:
Thesis (Ph.D.) - University of Glasgow, 2002.<br>Ph.D. thesis submitted to the Department of Psychology, University of Glasgow, 2002. Includes bibliographical references. Print version also available.
APA, Harvard, Vancouver, ISO, and other styles
23

Bose, Sougata. "On decision problems on word transducers with origin semantics." Thesis, Bordeaux, 2021. http://www.theses.fr/2021BORD0073.

Full text
Abstract:
La sémantique d'origine pour les transducteurs de mots a été introduite par Bojańczyk en 2014 afin d'obtenir une caractérisation indépendante de la machine pour les fonctions mot à mot définies par les transducteurs. Notre objectif principal était d'étudier certains problèmes de décision classiques pour les transducteurs dans la sémantique d'origine, tels que le problème d'inclusion et d'équivalence. Nous avons montré que ces problèmes deviennent décidables dans la sémantique d'origine, même si la version classique est indécidable.Motivé par l'observation que la sémantique d'origine est plus fine que la sémantique classique, nous avons défini les resynchroniseurs comme un moyen de décrire les distorsions d'origine et d'étudier les problèmes ci-dessus de manière relaxée. Nous avons étendu le modèle des resynchroniseurs rationnels, introduit par Filiot et al. pour les transducteurs unidirectionnels, aux resynchroniseurs réguliers, qui fonctionnent pour des classes de transducteurs plus grandes.Nous avons étudié les deux variantes du problème d’inclusion relative à une resynchronisation, qui demande si un transducteur est contenu dans un autre jusqu'à une distorsion spécifiée par un resynchroniseur. Nous avons montré que le problème peut être résolu lorsque le resynchroniseur fait partie de l'entrée. Lorsque le resynchroniseur n'est pas spécifié dans l'entrée, nous avons cherché à synthétiser un tel resynchroniseur, chaque fois que cela était possible. Nous appelons cela le problème de synthèse pour les resynchroniseurs et nous montrons qu'il est indécidable en général. Nous avons identifié quelques cas restreints où le problème devient décidable. Nous avons également étudié le problème de resynchronisabilité unidirectionnelle, qui demande si un transducteur bidirectionnel donné est resynchronisable dans un transducteur unidirectionnel, et nous avons montré que ce problème est également décidable<br>The origin semantics for word transducers was introduced by Bojańczyk in 2014 in order to obtain a machine-independent characterization for word-to-word functions defined by transducers. Our primary goal was to study some classical decision problems for transducers in the origin semantics, such as the containment and the equivalence problem. We showed that these problems become decidable in the origin semantics, even though the classical version is undecidable.Motivated by the observation that the origin semantics is more fine-grained than classical semantics, we defined resynchronizers as a way to describe distortions of origins, and to study the above problems in a more relaxed way. We extended the model of rational resynchronizers, introduced by Filiot et al. for one-way transducers, to regular resynchronizers, which work for larger classes of transducers.We studied the two variants of the containment up to resynchronizer problem, which asks if a transducer is contained in another up to a distortion specified by a resynchronizer. We showed that the problem is decidable when the resynchronizer is given as part of the input. When the resynchronizer is not specified in the input, we aimed to synthesize such a resynchronizer, whenever possible. We call this the synthesis problem for resynchronizers and show that it is undecidable in general. We identified some restricted cases when the problem becomes decidable. We also studied the one-way resynchronizability problem, which asks whether a given two-way transducer is resynchronizable in a one-way transducer, and showed that this problem is decidable as well
APA, Harvard, Vancouver, ISO, and other styles
24

Fagerlund, Martin. "Computing Word Senses by Semantic Mirroring and Spectral Graph Partitioning." Thesis, Linköping University, Scientific Computing, 2010. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-57103.

Full text
Abstract:
<p>In this thesis we use the method of Semantic Mirrors to create a graph of words that are semantically related to a seed word. Spectral graph partitioning methods are then used to partition the graph into subgraphs, and thus dividing the words into different word senses. These methods are applied to a bilingual lexicon of English and Swedish adjectives. A panel of human evaluators have looked at a few examples, and evaluated consistency within the derived senses and synonymy with the seed word.</p>
APA, Harvard, Vancouver, ISO, and other styles
25

Seifert, Lauren Sue. "The role of semantic memory in picture and word processing /." The Ohio State University, 1994. http://rave.ohiolink.edu/etdc/view?acc_num=osu1487849377296571.

Full text
APA, Harvard, Vancouver, ISO, and other styles
26

You, Huiling. "Unsupervised Lexical Semantic Change Detection with Context-Dependent Word Representations." Thesis, Uppsala universitet, Institutionen för lingvistik och filologi, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-444871.

Full text
Abstract:
In this work, we explore the usefulness of contextualized embeddings from language models on lexical semantic change (LSC) detection. With diachronic corpora spanning two time periods, we construct word embeddings for a selected set of target words, aiming at detecting potential LSC of each target word across time. We explore different systems of embeddings to cover three topics: contextualized vs static word embeddings, token- vs type-based embeddings, and multilingual vs monolingual language models. We use a multilingual dataset covering three languages (English, German, Swedish) and explore each system of embedding with two subtasks, a binary classification task and a ranking task. We compare the performance of different systems of embeddings, and seek to answer our research questions through discussion and analysis of experimental results. We show that contextualized word embeddings are on par with static word embeddings in the classification task. Our results also show that it is more beneficial to use the contextualized embeddings from a multilingual model than from a language specific model in most cases. We present that token-based setting is strong for static embeddings, and type-based setting for contextual embeddings, especially for the ranking task. We provide some explanation for the results we achieve, and propose improvements that can be made to our experiments for future work.
APA, Harvard, Vancouver, ISO, and other styles
27

Franks, Bradley William. "Criteria and concepts : an anti-realist approach to word meaning." Thesis, University of Edinburgh, 1990. http://hdl.handle.net/1842/26523.

Full text
APA, Harvard, Vancouver, ISO, and other styles
28

À, Beckett Margaret. "Gender assignment & word-final pronunciation in French two semantic systems." Muenchen LINCOM Europa, 2010. http://d-nb.info/99962945X/04.

Full text
APA, Harvard, Vancouver, ISO, and other styles
29

Lau, Ka-po Natalie. "Semantic interference of Chinese words in the picture-word interference task /." View the Table of Contents & Abstract, 2006. http://sunzi.lib.hku.hk/hkuto/record/B36846181.

Full text
APA, Harvard, Vancouver, ISO, and other styles
30

Esin, Yunus Emre. "Improvement Of Corpus-based Semantic Word Similarity Using Vector Space Model." Master's thesis, METU, 2009. http://etd.lib.metu.edu.tr/upload/12610759/index.pdf.

Full text
Abstract:
This study presents a new approach for finding semantically similar words from corpora using window based context methods. Previous studies mainly concentrate on either finding new combination of distance-weight measurement methods or proposing new context methods. The main difference of this new approach is that this study reprocesses the outputs of the existing methods to update the representation of related word vectors used for measuring semantic distance between words, to improve the results further. Moreover, this novel technique provides a solution to the data sparseness of vectors which is a common problem in methods which uses vector space model. The main advantage of this new approach is that it is applicable to many of the existing word similarity methods using the vector space model. The other and the most important advantage of this approach is that it improves the performance of some of these existing word similarity measuring methods.
APA, Harvard, Vancouver, ISO, and other styles
31

Lau, Ka-po Natalie, and 劉家寶. "Semantic interference of Chinese words in the picture-word interference task." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2006. http://hub.hku.hk/bib/B45007597.

Full text
APA, Harvard, Vancouver, ISO, and other styles
32

Tissier, Julien. "Improving methods to learn word representations for efficient semantic similarites computations." Thesis, Lyon, 2020. http://www.theses.fr/2020LYSES008.

Full text
Abstract:
De nombreuses applications en traitement du langage naturel (TALN) reposent sur les représentations de mots, ou “word embeddings”. Ces représentations doivent capturer à la fois de l’information syntaxique et sémantique pour donner des bonnes performances dans les tâches en aval qui les utilisent. Cependant, les méthodes courantes pour les apprendre utilisent des textes génériques comme Wikipédia qui ne contiennent pas d’information sémantique précise. De plus, un espace mémoire important est requis pour pouvoir les sauvegarder car le nombre de représentations de mots à apprendre peut être de l’ordre du million. Le sujet de ma thèse est de développer de nouveaux algorithmes pour améliorer l’information sémantique dans les word embeddings tout en réduisant leur taille en mémoire lors de leur utilisation dans des tâches en aval de TALN.La première partie de mes travaux améliore l’information sémantique contenue dans les word embeddings. J’ai développé dict2vec, un modèle qui utilise l’information des dictionnaires linguistiques lors de l’apprentissage des word embeddings. Les word embeddings appris par dict2vec obtiennent des scores supérieurs d’environ 15% par rapport à ceux appris avec d’autres méthodes sur des tâches de similarités sémantiques de mots. La seconde partie de mes travaux consiste à réduire la taille mémoire des word embeddings. J’ai développé une architecture basée sur un auto-encodeur pour transformer des word embeddings à valeurs réelles en vecteurs binaires, réduisant leur taille mémoire de 97% avec seulement une baisse de précision d’environ 2% dans des tâches de TALN en aval<br>Many natural language processing applications rely on word representations (also called word embeddings) to achieve state-of-the-art results. These numerical representations of the language should encode both syntactic and semantic information to perform well in downstream tasks. However, common models (word2vec, GloVe) use generic corpus like Wikipedia to learn them and they therefore lack specific semantic information. Moreover it requires a large memory space to store them because the number of representations to save can be in the order of a million. The topic of my thesis is to develop new learning algorithms to both improve the semantic information encoded within the representations while making them requiring less memory space for storage and their application in NLP tasks.The first part of my work is to improve the semantic information contained in word embeddings. I developed dict2vec, a model that uses additional information from online lexical dictionaries when learning word representations. The dict2vec word embeddings perform ∼15% better against the embeddings learned by other models on word semantic similarity tasks. The second part of my work is to reduce the memory size of the embeddings. I developed an architecture based on an autoencoder to transform commonly used real-valued embeddings into binary embeddings, reducing their size in memory by 97% with only a loss of ∼2% in accuracy in downstream NLP tasks
APA, Harvard, Vancouver, ISO, and other styles
33

Cooper, Elisa. "Pronouncing printed words : investigating a semantic contribution to adult word reading." Thesis, University of East London, 2013. http://roar.uel.ac.uk/3042/.

Full text
Abstract:
When considering print-to-sound word reading, orthography and phonology are obviously involved. However, another system, that of semantic memory, might also be involved in orthography-to-phonology computation. Whether this occurs is debated in the literature both in the interpretation of behavioural results (e.g., Monaghan & Ellis, 2002; Strain et al., 1995) and in the implementation of semantic memory within computational models of word reading (Coltheart et al., 2001; Plaut et al., 1996). The central aim of this thesis was to investigate whether there is a semantic contribution to orthography-to-phonology computation in healthy adult word reading. Experiments 1-4 used a semantic priming design in which a picture prime was followed either two trials later (Experiments 1, 3, and 4) or one trial later (Experiment 2) by a word target, and this investigated priming of various word types. Regression investigations explored whether semantic features and imageability were unique significant predictors of ELexicon single word reading reaction times while statistically controlling for age-of-acquisition. The two ERP experiments (Experiments 5 and 6) investigated the neurocorrelates of imageability and semantic features and whether there are semantic effects early in the time-course of low frequency word reading. Experiments 1, 2, 4, 5, and 6 and the regression investigations show evidence of a semantic contribution to low frequency regular and low frequency exception word reading. There is also some suggestion of a semantic contribution to high frequency word reading (Experiment 2 and Regression analyses). From the results of the three lines of investigation, it is concluded that semantic information is involved in healthy adult word reading, and these results are best accommodated by the connectionist triangle model of word reading. These investigations also provide information concerning various word types and factors that contribute to “easy” and “difficult” words, semantic memory models and their accounts of priming, and the measures, age-of-acquisition, imageability, and semantic features.
APA, Harvard, Vancouver, ISO, and other styles
34

Rost, Gwyneth Campbell. "Object categories provide semantic representation for 3-year-olds' word learning." Diss., University of Iowa, 2011. https://ir.uiowa.edu/etd/2764.

Full text
Abstract:
Word learning implies learning of both a phonological form and its referent. For nouns, the referent is typically a category of objects, with variability between objects within the category but an overlying similarity that allows them to be categorized together, to function similarly, and to be called by the same label. We hypothesized that by strengthening knowledge of the category of referents a word refers to, we could strengthen learning and use of the word. Three-year old children were provided with elaborated referent category information in the form of multiple exemplars of the referent category. In the first manipulation, children were trained on identical exemplars or variable exemplars. A second manipulation provided children with variable exemplars that had been distributed to support a prototype. Children in the third condition, who were provided with a prototype plus variants, learned words best in expressive and receptive tasks, when tested on trained and untrained items, and at two time points. In a second manipulation, we asked if simultaneous presentation of multiple exemplars leads to better learning of the object label than sequential presentation. Results indicated little difference. We conclude that 3-year-olds learn words best in the presence of variability distributed to highlight both invariant elements of the referent category and those elements that are allowed to vary.
APA, Harvard, Vancouver, ISO, and other styles
35

Guthormsen, Amy. "Conceptual integration of mathematical and semantic knowledge /." Thesis, Connect to this title online; UW restricted, 2007. http://hdl.handle.net/1773/8995.

Full text
APA, Harvard, Vancouver, ISO, and other styles
36

Flynn, Michael J. "Structure building operations and word order." New York : Garland Pub, 1985. http://catalog.hathitrust.org/api/volumes/oclc/12285682.html.

Full text
APA, Harvard, Vancouver, ISO, and other styles
37

Hughes, Richard Sylvester. "The conceptual structure of product semantic models." Thesis, Brunel University, 1999. http://bura.brunel.ac.uk/handle/2438/4969.

Full text
Abstract:
The study is concerned with the conceptual structure and content of the framework for characterising user-product interaction, proposed under the title – ‘Product Semantics’. The sources for the critique of design, from which the framework is derived, are identified and analysed, and the substantive theoretical and methodological content given initial consideration in terms of the deployment of the central concept of ‘meaning’, and the principal theoretical approaches adopted in the analysis of meaning and semantic concepts generally. The commitment to a cognitive and experiential approach to user-interaction is established and the concepts central to the framework, and requiring more detailed analysis, are identified. The core of the study consists in an analysis of the sequence of concepts and contexts that are chiefly used in the theoretical articulation of the framework, including - function, affordance, categorisation, artefacts, meaning and expression - of which the concept of affordance is central to the structure. On the basis of the initial consideration of the structure and content of the scheme, and in the light of the analysis of concepts, the explanatory structure of the framework is established. It is argued that the core commitment to an experiential and cognitive account, and the form of the explanatory structure, are jointly incompatible with the conceptual content of the framework, particularly in respect of the pivotal role of the concept of affordance. Proposals are advanced for an alternative interpretation which addresses the central issues of consistency and coherence, and which suggests an alternative approach to the conceptual characterisation of the framework and the form of the explanatory hierarchy. The implications of the framework, and the proposed alternative interpretation, are considered in respect of their application in shaping approaches to the development of design theory and methodology, and the experiential aspect of semantics and cognition.
APA, Harvard, Vancouver, ISO, and other styles
38

Hope, Thomas. "On the Structure of Semantic Number Knowledge." Doctoral thesis, Università degli studi di Padova, 2008. http://hdl.handle.net/11577/3425575.

Full text
Abstract:
This doctoral thesis present the results of five distinct projects which all share a common focus - the structure of cognitive number processing. Two of the projects employ psychophysical methods to (a) establish some perceptual constraints on the processing of very large numbers (or very long digit strings) and (b) demonstrate a doissociation between subjects' sensitivities to these (large) numbers' holistic content from their sensitivities to the numbers' single-digit components. The other three projects are primarily computational; each proposes a novel model-building method, then applies it to a problem of relevance to numerical cognition. The results include (a) a novel proposal for the format of semantic number knowledge, which is supported by recent neurophysiological data, (b) a model-space analysis which suggests that this new format may be more effective than its conventional counterparts in driving models of multi-digit number comparison, and (c) a demonstration of model-building with minimal assumptions, which captures the optimal and near-optimal features of neural information processing in a well-known decision problem.
APA, Harvard, Vancouver, ISO, and other styles
39

Lilliehöök, Hampus. "Extraction of word senses from bilingual resources using graph-based semantic mirroring." Thesis, Linköpings universitet, Interaktiva och kognitiva system, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-91880.

Full text
Abstract:
In this thesis we retrieve semantic information that exists implicitly in bilingual data. We gather input data by repeatedly applying the semantic mirroring procedure. The data is then represented by vectors in a large vector space. A resource of synonym clusters is then constructed by performing K-means centroid-based clustering on the vectors. We evaluate the result manually, using dictionaries, and against WordNet, and discuss prospects and applications of this method.<br>I det här arbetet utvinner vi semantisk information som existerar implicit i tvåspråkig data. Vi samlar indata genom att upprepa proceduren semantisk spegling. Datan representeras som vektorer i en stor vektorrymd. Vi bygger sedan en resurs med synonymkluster genom att applicera K-means-algoritmen på vektorerna. Vi granskar resultatet för hand med hjälp av ordböcker, och mot WordNet, och diskuterar möjligheter och tillämpningar för metoden.
APA, Harvard, Vancouver, ISO, and other styles
40

Manion, Steve Lawrence. "Unsupervised Knowledge-based Word Sense Disambiguation: Exploration & Evaluation of Semantic Subgraphs." Thesis, University of Canterbury. Department of Mathematics & Statistics, 2014. http://hdl.handle.net/10092/10016.

Full text
Abstract:
Hypothetically, if you were told: Apple uses the apple as its logo . You would immediately detect two different senses of the word apple , these being the company and the fruit respectively. Making this distinction is the formidable challenge of Word Sense Disambiguation (WSD), which is the subtask of many Natural Language Processing (NLP) applications. This thesis is a multi-branched investigation into WSD, that explores and evaluates unsupervised knowledge-based methods that exploit semantic subgraphs. The nature of research covered by this thesis can be broken down to: 1. Mining data from the encyclopedic resource Wikipedia, to visually prove the existence of context embedded in semantic subgraphs 2. Achieving disambiguation in order to merge concepts that originate from heterogeneous semantic graphs 3. Participation in international evaluations of WSD across a range of languages 4. Treating WSD as a classification task, that can be optimised through the iterative construction of semantic subgraphs The contributions of each chapter are ranged, but can be summarised by what has been produced, learnt, and raised throughout the thesis. Furthermore an API and several resources have been developed as a by-product of this research, all of which can be accessed by visiting the author’s home page at http://www.stevemanion.com. This should enable researchers to replicate the results achieved in this thesis and build on them if they wish.
APA, Harvard, Vancouver, ISO, and other styles
41

Chan, Alexander Mark. "Extracting Spatiotemporal Word and Semantic Representations from Multiscale Neurophysiological Recordings in Humans." Thesis, Harvard University, 2012. http://dissertations.umi.com/gsas.harvard:10251.

Full text
Abstract:
With the recent advent of neuroimaging techniques, the majority of the research studying the neural basis of language processing has focused on the localization of various lexical and semantic functions. Unfortunately, the limited time resolution of functional neuroimaging prevents a detailed analysis of the dynamics involved in word recognition, and the hemodynamic basis of these techniques prevents the study of the underlying neurophysiology. Compounding this problem, current techniques for the analysis of high-dimensional neural data are mainly sensitive to large effects in a small area, preventing a thorough study of the distributed processing involved for representing semantic knowledge. This thesis demonstrates the use of multivariate machine-learning techniques for the study of the neural representation of semantic and speech information in electro/magneto-physiological recordings with high temporal resolution. Support vector machines (SVMs) allow for the decoding of semantic category and word-specific information from non-invasive electroencephalography (EEG) and magnetoenecephalography (MEG) and demonstrate the consistent, but spatially and temporally distributed nature of such information. Moreover, the anteroventral temporal lobe (avTL) may be important for coordinating these distributed representations, as supported by the presence of supramodal category-specific information in intracranial recordings from the avTL as early as 150ms after auditory or visual word presentation. Finally, to study the inputs to this lexico-semantic system, recordings from a high density microelectrode array in anterior superior temporal gyrus (aSTG) are obtained, and the recorded spiking activity demonstrates the presence of single neurons that respond specifically to speech sounds. The successful decoding of word identity from this firing rate information suggests that the aSTG may be involved in the population coding of acousto-phonetic speech information that is likely on the pathway for mapping speech-sounds to meaning in the avTL. The feasibility of extracting semantic and phonological information from multichannel neural recordings using machine learning techniques provides a powerful method for studying language using large datasets and has potential implications for the development of fast and intuitive communication prostheses.<br>Engineering and Applied Sciences
APA, Harvard, Vancouver, ISO, and other styles
42

Miller, Matthew Adam. "Semantic spaces : behavior, language and word learning in the Human Speechome corpus." Thesis, Massachusetts Institute of Technology, 2011. http://hdl.handle.net/1721.1/69805.

Full text
Abstract:
Thesis (S.M.)--Massachusetts Institute of Technology, School of Architecture and Planning, Program in Media Arts and Sciences, 2011.<br>Cataloged from PDF version of thesis.<br>Includes bibliographical references (p. 149-152).<br>The Human Speechome Project is an unprecedented attempt to record, analyze and understand the process of language acquisition. It is composed of over 90,000 hours of video and 150,000 hours of audio, capturing roughly 80% of the waking hours of a single child from his birth until age 3. This thesis proposes and develops a method for representing and analyzing a video corpus of this scale that is both compact and efficient, while retaining much of the important information about large scale behaviors of the recorded subjects. This representation is shown to be useful for the unsupervised modeling, clustering and exploration of the data, particularly when it is combined with text transcripts of the speech. Novel methods are introduced to perform Spatial Latent Semantic Analysis - extending the popular framework for topic modeling to cover behavior as well. Finally, the representation is used to analyze the inherent "spatiality" of individual words. A surprising connection is demonstrated between the uniqueness of a word's spatial distribution and how early it is learned by the child.<br>by Matthew Miller.<br>S.M.
APA, Harvard, Vancouver, ISO, and other styles
43

Burrill, Katheryn Elizabeth. "The effect of phonological and semantic cues on word retrieval in adults." Scholarly Commons, 2008. https://scholarlycommons.pacific.edu/uop_etds/694.

Full text
Abstract:
Word retrieval difficulties can affect individuals who have had strokes or head trauma (Goodglass & Wingfield, 1997) and to a lesser extent, typically aging adults. This can affect an individual's ability to name pictures accurately and quickly. Cues are used to help individuals with word retrieval difficulties in fmding specific words. Two commonly used cues are semantic and phonological cues. Semantic cues can be information about the word the person is trying to retrieve, such as its definition, and/or its functions. Phonological cues are usually the initial sound of a word that a person is attempting to retrieve. Previous research has suggested that both of these cues, in isolation, are effective in stimulating word retrieval during naming tasks (Li & Williams 1989; Stirnley & Noll 1991 ). However, research has not investigated the effects of combining these two cues during picture naming tasks. The current study observed participants under four different cueing conditions during a picture naming task with the Boston Naming Test. The four conditions include a control group (received no cues), a semantically cued group (received a semantic cue before being asked to name a picture), a phonologically cued group (received a phonological cue before being asked to name a picture), and a semanticallyphonologically cued group (received a semantic and phonological cue before being asked to name a picture). Each group was compared on number of items correctly named and response times. The results indicated that there was no statistically significant difference between the groups with regard to number of items named. There was a statistically significant difference found between the groups with regards to response times. These findings are discussed and compared to previous research and current word retrieval theories.
APA, Harvard, Vancouver, ISO, and other styles
44

Pillay, Bhavani S. "Semantic feature analysis for word retrieval in a small aphasia-group setting." Diss., University of Pretoria, 2016. http://hdl.handle.net/2263/58976.

Full text
Abstract:
Background: Word retrieval deficits are a common, pervasive feature of aphasia. Semantic feature analysis (SFA) is a popular treatment technique for word retrieval impairment. Preliminary evidence of its use in small aphasia-group settings suggests improved word retrieval in discourse-based tasks with improvements in communicative informativeness (Antonucci, 2009; Falconer & Antonucci, 2012). Objective: The aim of this study was to investigate the training of SFA within discourse during a small (two-member) aphasia group treatment to ascertain if gains would be made in word retrieval abilities and communicative informativeness. Method: Two female participants with chronic aphasia, aged seventy seven and sixty eight, participated weekly in group treatment for five consecutive weeks where stimuli were used to elicit increasingly naturalistic discourse. A multiple-baseline design was used and a case series analysis conducted. Baseline measures, treatment data and data from the six week follow-up (SWFU) session were analysed using the protocols developed by Nicholas and Brookshire (1993) and Mayer and Murray (2003). Results: Both participants demonstrated certain modest gains in overall communicative informativeness but these were not maintained at the SWFU session. Gains in informativeness ranged from 12% to 13.4% for participants, which agree with improvements in other studies (Antonucci, 2009; Boyle & Coelho, 1995; Falconer & Antonucci, 2012). P2 (presented with anomic aphasia) performed better than P1 (presented with Broca’s aphasia and a component of apraxia of speech [AOS]) with regards to communicative informativeness (%CIUs) and had greater access to semantic knowledge (i.e. could access lexical forms at an improved rate). Conclusion: The small group setting may offer advantages to facilitate communication skills and increase participation in everyday conversation. Further research is needed to identify benefits for individuals with differing aphasia types and severity, optimal intensity and frequency of group treatment and the role of peer support during group interaction. Key Words Aphasia, stroke rehabilitation, word retrieval, semantic feature analysis, connected speech, discourse, group therapy, communicative informativeness, functional communication.<br>Dissertation (MA)--University of Pretoria, 2016.<br>Speech-Language Pathology and Audiology<br>MA<br>Unrestricted
APA, Harvard, Vancouver, ISO, and other styles
45

Lee, Wai-ling Janise. "The effectiveness of semantic and syllabic cues for Cantonese aphasic patients with naming difficulties." Click to view the E-thesis via HKUTO, 2001. http://sunzi.lib.hku.hk/hkuto/record/B36207937.

Full text
Abstract:
Thesis (B.Sc)--University of Hong Kong, 2001.<br>"A dissertation submitted in partial fulfilment of the requirements for the Bachelor of Science (Speech and Hearing Sciences), The University of Hong Kong, 4 May 2001." Also available in print.
APA, Harvard, Vancouver, ISO, and other styles
46

Hopkins, Paul Stanley. "The phonological structure of the Kashubian word." Thesis, National Library of Canada = Bibliothèque nationale du Canada, 2001. http://www.collectionscanada.ca/obj/s4/f2/dsk3/ftp04/NQ58570.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
47

Haugan, Jens. "Old Norse Word Order and Information Structure." Doctoral thesis, Trondheim : Norwegian University of science and technology, 2000. http://catalogue.bnf.fr/ark:/12148/cb38802334g.

Full text
APA, Harvard, Vancouver, ISO, and other styles
48

Shaik, Arshad. "Biomedical Semantic Embeddings: Using Hybrid Sentences to Construct Biomedical Word Embeddings and Their Applications." Thesis, University of North Texas, 2019. https://digital.library.unt.edu/ark:/67531/metadc1609064/.

Full text
Abstract:
Word embeddings is a useful method that has shown enormous success in various NLP tasks, not only in open domain but also in biomedical domain. The biomedical domain provides various domain specific resources and tools that can be exploited to improve performance of these word embeddings. However, most of the research related to word embeddings in biomedical domain focuses on analysis of model architecture, hyper-parameters and input text. In this paper, we use SemMedDB to design new sentences called `Semantic Sentences'. Then we use these sentences in addition to biomedical text as inputs to the word embedding model. This approach aims at introducing biomedical semantic types defined by UMLS, into the vector space of word embeddings. The semantically rich word embeddings presented here rivals state of the art biomedical word embedding in both semantic similarity and relatedness metrics up to 11%. We also demonstrate how these semantic types in word embeddings can be utilized.
APA, Harvard, Vancouver, ISO, and other styles
49

Shaik, Arshad. "Biomedical Semantic Embeddings: Using Hybrid Sentences to Construct Biomedical Word Embeddings and its Applications." Thesis, University of North Texas, 2019. https://digital.library.unt.edu/ark:/67531/metadc1609064/.

Full text
Abstract:
Word embeddings is a useful method that has shown enormous success in various NLP tasks, not only in open domain but also in biomedical domain. The biomedical domain provides various domain specific resources and tools that can be exploited to improve performance of these word embeddings. However, most of the research related to word embeddings in biomedical domain focuses on analysis of model architecture, hyper-parameters and input text. In this paper, we use SemMedDB to design new sentences called `Semantic Sentences'. Then we use these sentences in addition to biomedical text as inputs to the word embedding model. This approach aims at introducing biomedical semantic types defined by UMLS, into the vector space of word embeddings. The semantically rich word embeddings presented here rivals state of the art biomedical word embedding in both semantic similarity and relatedness metrics up to 11%. We also demonstrate how these semantic types in word embeddings can be utilized.
APA, Harvard, Vancouver, ISO, and other styles
50

Noble, Claire. "Early comprehension of argument structure and semantic roles." Thesis, University of Liverpool, 2009. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.526846.

Full text
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!