Dissertations / Theses on the topic 'Modèle lexical'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Modèle lexical.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Jousse, Anne-Laure. "Modèle de structuration des relations lexicales fondé sur le formalisme des fonctions lexicales." Thèse, Paris 7, 2010. http://hdl.handle.net/1866/4347.
Full textThis thesis proposes a model for structuring lexical relations, based on the concept of lexical functions (LFs) proposed in Meaning-Text Theory [Mel’cuk, 1997]. The lexical relations taken into account include semantic derivations and collocations as defined within this theoretical framework, known as Explanatory and Combinatorial Lexicology [Mel’cuk et al., 1995]. Considering the assumption that lexical relations are neither encoded nor made available in lexical databases in an entirely satisfactory manner, we assume the necessity of designing a new model for structuring them. First of all, we justify the relevance of devising a system of lexical functions rather than a simple classification. Next, we present the four perspectives developped in the system: a semantic perspective, a combinatorial one, another one targetting the parts of speech of the elements involved in a lexical relation, and, finally, a last one emphasizing which element of the relation is focused on. This system covers all LFs, even non-standard ones, for which we have proposed a normalization of the encoding. Our system has already been implemented into the DiCo relational database. We propose three further applications that can be developed from it. First, it can be used to build browsing interfaces for lexical databases such as the DiCo. It can also be directly consulted as a tool to assist lexicographers in encoding lexical relations by means of lexical functions. Finally, it constitutes a reference to compute lexicographic information which will, in future work, be implemented in order to automatically fill in some fields within the entries in lexical databases.
Thèse réalisée en cotutelle avec l'Université Paris Diderot (Paris 7)
Quang, Vu Minh. "Exploitation de la prosodie pour la segmentation et l'analyse automatique de signaux de parole." Grenoble INPG, 2007. http://www.theses.fr/2007INPG0104.
Full textThis thesis work is at the frontier between multimedia information retrieval and automatic speech processing. During the last years, a new task en speech processing: the rich transcription of an audio document. An important meta-data for rich transcription is the information on sentence type sentence of interrogative or affirmative type). The study on the prosodie differences between these two types of sentences in Vietnamese languaç detection and classification of sentence type in French language and in Vietnamese language is the main subject of this research work. Our depar1 study on French language. We've realized a system for segmentation and automatic detection of sentence type based on both prosodie and lexica information. The system has been validated on real world spontaneous speech corpus which are recording of conversations via telephone, betwee and a tourism office staff, recruiting interview, project meeting. After this first study on French, we've extended our research in Vietnamese langui language where ail studies until now on prosodie system are still preliminary. We've carried a study on the prosodie differences between interroga affirmative sentences in both production and perception levels. Next, based on these results, a classification motor has been built
Abdel, Jalil Mohamed ali. "Approche polysémique et traductologique du Coran : la sourate XXII (Al-Hajj [le pèlerinage]) comme modèle." Thesis, Aix-Marseille, 2017. http://www.theses.fr/2017AIXM0348.
Full textAccording to Islamic tradition, one of the core characteristics of the Quran is that it is a polysemic text par excellence (ḥammāl dhū wujūh, bearer of several faces). To say that the Quranic text is polysemic implies that its various exegeses are as many possible readings of it, which implies in turn that its translations are also as many readings that complete each other. The accumulation of translations is thus another expression of the polysemy of the original text, even if the diversity of these translations does not match that of the exegeses.The thesis deals with the analysis of the Surah of Al-Ḥajj and it is based on two research axes:I. a study of the polysemy of the original text (Surah of Al-Ḥajj).II. A study of the polysemy of the final text (18 French translations) to show how translation reduces and/or modifies polysemy. The corpus of translations (18 translations) covers all the periods of the history of the translation of the Koran from 1647 until 2010 in order to see the evolution of the translation of the Koranic text.As a closed space that evolves independently from exegesis to more literality, the translations meet and complement each other, reflecting in their diversity with slight modification a large part of the polysemy united and concentrated in the original text but sporadic, sparse and Dispersed in the translations
Ghoul, Dhaou. "Classifications et grammaires des invariants lexicaux arabes en prévision d’un traitement informatique de cette langue. Construction d’un modèle théorique de l’arabe : la grammaire des invariants lexicaux temporels." Thesis, Paris 4, 2016. http://www.theses.fr/2016PA040184.
Full textThis thesis focuses on the classification and the treatment of Arabic lexical invariants that express a temporal aspect. Our aim is to create a diagram of grammar (finite state machine) for each invariant. In this work, we limited our treatment to 20 lexical invariants. Our assumption is that the lexical invariants are located at the same structural level (formal) as the schemes in the language quotient (skeleton) of the Arabic language. They hide much information and involve syntactic expectations that make it possible to predict the structure of the sentence.In the first part of our research tasks, we present the concept of “invariant lexical” by exposing the various levels of invariance. Then, we classify the invariants according to several criteria.The second part is the object of our own study concerning the temporal lexical invariants. We present our linguistic method as well as our approach of modelling using diagrams of grammars. Then, we analyze the simple lexical invariants such “ḥattā, baʿda” and the complexes ones such “baʿdamā, baynamā”.Finally, an experimental application “Kawâkib” was used to detect and identify the lexical invariants by showing their strong points as well as their gaps. We also propose a new vision of the next version of “Kawâkib” that can represent a teaching application of Arabic without lexicon
Desalle, Yann. "Réseaux lexicaux, métaphore, acquisition : une approche interdisciplinaire et inter-linguistique du lexique verbal." Phd thesis, Université Toulouse le Mirail - Toulouse II, 2012. http://tel.archives-ouvertes.fr/tel-00714834.
Full textMaheux, Marie-Andrée. "Description lexicale du français québécois un modèle prédictionnairique." Mémoire, Université de Sherbrooke, 1994. http://hdl.handle.net/11143/10036.
Full textRomeo, Lauren Michele. "The Structure of the lexicon in the task of the automatic acquisition of lexical information." Doctoral thesis, Universitat Pompeu Fabra, 2015. http://hdl.handle.net/10803/325420.
Full textLa información de clase semántica de los nombres es fundamental para una amplia variedad de tareas del procesamiento del lenguaje natural (PLN), como la traducción automática, la discriminación de referentes en tareas como la detección y el seguimiento de eventos, la búsqueda de respuestas, el reconocimiento y la clasificación de nombres de entidades, la construcción y ampliación automática de ontologías, la inferencia textual, etc. Una aproximación para resolver la construcción y el mantenimiento de los léxicos de gran cobertura que alimentan los sistemas de PNL, una tarea muy costosa y lenta, es la adquisición automática de información léxica, que consiste en la inducción de una clase semántica relacionada con una palabra en concreto a partir de datos de su distribución obtenidos de un corpus. Precisamente, por esta razón, se espera que la investigación actual sobre los métodos para la producción automática de léxicos de alta calidad, con gran cantidad de información y con anotación de clase como el trabajo que aquí presentamos, tenga un gran impacto en el rendimiento de la mayoría de las aplicaciones de PNL. En esta tesis, tratamos la adquisición automática de información léxica como un problema de clasificación. Con este propósito, adoptamos métodos de aprendizaje automático para generar un modelo que represente los datos de distribución vectorial que, basados en ejemplos conocidos, permitan hacer predicciones de otras palabras desconocidas. Las principales preguntas de investigación que planteamos en esta tesis son: (i) si los datos de corpus proporcionan suficiente información para construir representaciones de palabras de forma eficiente y que resulten en decisiones de clasificación precisas y sólidas, y (ii) si la adquisición automática puede gestionar, también, los nombres polisémicos. Para hacer frente a estos problemas, realizamos una serie de validaciones empíricas sobre nombres en inglés. Nuestros resultados confirman que la información obtenida a partir de la distribución de los datos de corpus es suficiente para adquirir automáticamente clases semánticas, como lo demuestra un valor-F global promedio de 0,80 aproximadamente utilizando varios modelos de recuento de contextos y en datos de corpus de distintos tamaños. No obstante, tanto el estado de la cuestión como los experimentos que realizamos destacaron una serie de retos para este tipo de modelos, que son reducir la escasez de datos del vector y dar cuenta de la polisemia nominal en las representaciones distribucionales de las palabras. En este contexto, los modelos de word embedding (WE) mantienen la “semántica” subyacente en las ocurrencias de un nombre en los datos de corpus asignándole un vector. Con esta elección, hemos sido capaces de superar el problema de la escasez de datos, como lo demuestra un valor-F general promedio de 0,91 para las clases semánticas de nombres de sentido único, a través de una combinación de la reducción de la dimensionalidad y de números reales. Además, las representaciones de WE obtuvieron un rendimiento superior en la gestión de las ocurrencias asimétricas de cada sentido de los nombres de tipo complejo polisémicos regulares en datos de corpus. Como resultado, hemos podido clasificar directamente esos nombres en su propia clase semántica con un valor-F global promedio de 0,85. La principal aportación de esta tesis consiste en una validación empírica de diferentes representaciones de distribución utilizadas para la clasificación semántica de nombres junto con una posterior expansión del trabajo anterior, lo que se traduce en recursos léxicos y conjuntos de datos innovadores que están disponibles de forma gratuita para su descarga y uso.
Lexical semantic class information for nouns is critical for a broad variety of Natural Language Processing (NLP) tasks including, but not limited to, machine translation, discrimination of referents in tasks such as event detection and tracking, question answering, named entity recognition and classification, automatic construction and extension of ontologies, textual inference, etc. One approach to solve the costly and time-consuming manual construction and maintenance of large-coverage lexica to feed NLP systems is the Automatic Acquisition of Lexical Information, which involves the induction of a semantic class related to a particular word from distributional data gathered within a corpus. This is precisely why current research on methods for the automatic production of high- quality information-rich class-annotated lexica, such as the work presented here, is expected to have a high impact on the performance of most NLP applications. In this thesis, we address the automatic acquisition of lexical information as a classification problem. For this reason, we adopt machine learning methods to generate a model representing vectorial distributional data which, grounded on known examples, allows for the predictions of other unknown words. The main research questions we investigate in this thesis are: (i) whether corpus data provides sufficient distributional information to build efficient word representations that result in accurate and robust classification decisions and (ii) whether automatic acquisition can handle also polysemous nouns. To tackle these problems, we conducted a number of empirical validations on English nouns. Our results confirmed that the distributional information obtained from corpus data is indeed sufficient to automatically acquire lexical semantic classes, demonstrated by an average overall F1-Score of almost 0.80 using diverse count-context models and on different sized corpus data. Nonetheless, both the State of the Art and the experiments we conducted highlighted a number of challenges of this type of model such as reducing vector sparsity and accounting for nominal polysemy in distributional word representations. In this context, Word Embeddings (WE) models maintain the “semantics” underlying the occurrences of a noun in corpus data by mapping it to a feature vector. With this choice, we were able to overcome the sparse data problem, demonstrated by an average overall F1-Score of 0.91 for single-sense lexical semantic noun classes, through a combination of reduced dimensionality and “real” numbers. In addition, the WE representations obtained a higher performance in handling the asymmetrical occurrences of each sense of regular polysemous complex-type nouns in corpus data. As a result, we were able to directly classify such nouns into their own lexical-semantic class with an average overall F1-Score of 0.85. The main contribution of this dissertation consists of an empirical validation of different distributional representations used for nominal lexical semantic classification along with a subsequent expansion of previous work, which results in novel lexical resources and data sets that have been made freely available for download and use.
De, la Garza Bernardo. "Creating lexical models: do foreign language learning techniques affect lexical organization in fluent bilinguals?" Diss., Kansas State University, 2012. http://hdl.handle.net/2097/14127.
Full textDepartment of Psychology
Richard J. Harris
The use of different language learning methods for the purposes of acquiring foreign language vocabulary has long been explored but studies have often failed to take into account the potential effects on lexical processing. The current study examined the effectiveness of the Keyword, Context and Paired-Associate learning methods in acquiring foreign language vocabulary, but primarily focusing on the lexical and conceptual organization effects that each method may have on a foreign language learner. Three main theories/models (i.e., Word Association, Concept Mediated and Revised Asymmetrical Hierarchical) have been used to explain the organization of bilingual lexical, conceptual stores and connections between each store, but studies have not examined the addition of a third language (i.e., L3) and the potential connections created between new L3 and the two existing language stores. It was predicted that since low-proficiency bilinguals would create lexical models which heavily rely on translation equivalents, thus, the use of non-elaborative learning methods would assist in creating only lexical translation links, while more sophisticated elaborative methods would be successful in creating direct access to the conceptual meaning. The current study further explored the potential effects of language learning methods on comprehension ability, requiring the creation of situation models for comprehension. Finally, the present study explored the immediate and delayed effects of language learning methods on both vocabulary acquisition and comprehension ability. Results from the current study indicated that all learning methods were successful in creating and conceptual connections between the languages and the conceptual store, while Keyword learners had significantly better scores on certain trial types. Differences in terms in lexical and conceptual strength are suggested since differences in RTs and scores were found between some of the learning methods. Furthermore, in terms of comparisons across time, repeated testing learners attained better scores on all trial types in comparison to learners who were only tested at Time 2. Lastly, when assessing if lexical links could be created to a non-associated highly fluent second language known by the bilingual, results indicated that each language learning method successfully created such lexical connections, but these links were weaker in strength than those of the base language that was used during learning. Based on the current results, new models of lexical access are proposed which vary based on the use of language learning methods. The current findings also have strong implications and applications to the field of foreign language acquisition, primarily for bilingual language learners acquiring an L3.
Rivière, Laura. "Etude de l'importance relative des contraintes linguistiques et extralinguistiques conduisant à la compréhension de l'ironie." Electronic Thesis or Diss., Aix-Marseille, 2019. http://www.theses.fr/2019AIXM0284.
Full textThe objective of this thesis was, using the framework of the Constraints Satisfaction model, to determine, for the first time in French, the role played by several types of constraints (i.e., pragmatic, linguistic and sociocultural) in the understanding of ironic criticisms and ironic praises.The results of a first experiment, in which we used a listening task, showed that the incongruity between the context and the utterance was a stronger cue than prosody in the understanding of ironic critics. Indeed, we showed that while all participants, in their interpretations, relied on contextual information, only some participants also used prosodic cues. The results of the two subsequent experiments, consisting of written tasks, confirmed the main role played by pragmatic constraints in irony understanding, and particularly in understanding of ironic criticisms. Our results also highlighted the contribution, while at a lower level than pragmatic constraints, of sociocultural constraints of the participants in the irony understanding. Our results also confirmed the asymmetry of irony and showed that the pragmatic constraints contributing to the understanding of ironic praises would be different from those contributing to the understanding of ironic criticism
De, Nadai Patrick. "De l'image dictionnairique au modèle lexicographique : la systémique lexicale." Paris 8, 1992. http://www.theses.fr/1993PA080771.
Full textMonolingual language dictionaries contain a lot of informations describing the language they are dealing with. Since this description is based on a given corpus, dictionaries provide a certain 'image' of this language. In order to improve the quality of dictionaries, it would be necessary for lexicographers to have a 'model' providing a description of the language as objective as possible. If language is conceived of as consisting in a 'system' of relations (both lexical and syntactical), the model should present all relations established in the language. The purpose of our work is to propose a method so as to build up such a model, from the informations contained in an extensive dictionary. After having established what a 'language dictionary' and a 'word definition' are, we analyse the definitions of our corpus, in order to extract lexical relations. Then, with these extracted relations, we try to outline a linguistic analysis of the 'prefix' anti-
Clark, Stephen. "Class-based statistical models for lexical knowledge acquisition." Thesis, University of Sussex, 2001. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.341541.
Full textHagiwara, Masato, Yasuhiro Ogawa, and Katsuhiko Toyama. "AUTOMATIC ACQUISITION OF LEXICAL KNOWLEDGE USING LATENT SEMANTIC MODELS." INTELLIGENT MEDIA INTEGRATION NAGOYA UNIVERSITY / COE, 2006. http://hdl.handle.net/2237/10444.
Full textMendling, Jan, Fabian Pittke, and Henrik Leopold. "Automatic detection and resolution of lexical ambiguity in process models." Gesellschaft für Informatik e.V, 2016. https://dl.gi.de/handle/20.500.12116/730.
Full textVial, Loïc. "Modèles neuronaux joints de désambiguïsation lexicale et de traduction automatique." Thesis, Université Grenoble Alpes, 2020. http://www.theses.fr/2020GRALM032.
Full textWord Sense Disambiguation (WSD) and Machine Translation (MT) are two central and among the oldest tasks of Natural Language Processing (NLP). Although they share a common origin, WSD being initially conceived as a fundamental problem to be solved for MT, the two tasks have subsequently evolved very independently of each other. Indeed, on the one hand, MT has been able to overcome the explicit disambiguation of terms thanks to statistical and neural models trained on large amounts of parallel corpora, and on the other hand, WSD, which faces some limitations such as the lack of unified resources and a restricted scope of applications, remains a major challenge to allow a better understanding of the language in general.Today, in a context in which neural networks and word embeddings are becoming more and more important in NLP research, the recent neural architectures and the new pre-trained language models offer not only some new possibilities for developing more efficient WSD and MT systems, but also an opportunity to bring the two tasks together through joint neural models, which facilitate the study of their interactions.In this thesis, our contributions will initially focus on the improvement of WSD systems by unifying the ressources that are necessary for their implementation, constructing new neural architectures and developing original approaches to improve the coverage and the performance of these systems. Then, we will develop and compare different approaches for the integration of our state of the art WSD systems and language models into MT systems for the overall improvement of their performance. Finally, we will present a new architecture that allows to train a joint model for both WSD and MT, based on our best neural systems
Schwab, Didier. "Approche hybride - lexicale et thématique - pour la modélisation, la détection et l'exploitation des fonctions lexicales en vue de l'analyse sémantique de texte." Phd thesis, Université Montpellier II - Sciences et Techniques du Languedoc, 2005. http://tel.archives-ouvertes.fr/tel-00333334.
Full textLes informations lexicales sont issues de la base lexicale sémantique dont nous introduisons l'architecture à trois niveaux d'objets lexicaux (item lexical, acception, lexie). Elles sont matérialisées sous la forme de Relations Lexicales Valuées qui traduisent la probabilité d'existence de la relation entre les objets. L'utilité de ces relations a pu être mis en évidence pour l'analyse sémantique grâce à l'utilisation du paradigme des algorithmes à fourmis. Le modèle introduit dans cette thèse, utilise à la fois les vecteurs conceptuels et les relations du réseau lexical pour résoudre une partie des problèmes posés lors d'une analyse sémantique.
Tous nos outils ont été implémentés en Java. Ils reposent sur Blexisma (Base LEXIcale Sémantique Multi-Agent) une architecture multi-agent élaborée au cours de cette thèse dont l'objectif est d'intégrer tout élément lui permettant de créer, d'améliorer et d'exploiter une ou plusieurs Bases Lexicales Sémantiques. Les expériences menées ont montré la faisabilité de cette approche, sa pertinence en termes d'amélioration globale de l'analyse et ouvert des perspectives de recherches fort intéressantes.
Aravena, Sandra. "Dynamics of language induced cortical motor activity : determining the linguistic contexts that trigger motor activation during lexical semantic processing." Thesis, Lyon 2, 2014. http://www.theses.fr/2014LYO20010/document.
Full textThe present dissertation was conducted in order to specify the relationship between motor and language structures as cooperative systems in lexical meaning construction. Specifically, this thesis aimed at deepening our understanding of how the linguistic context coordinates the recruitment of motor structures during lexical semantic processing. Although the involvement of motor activity in action-related language comprehension is now sufficiently documented, the specific role that motor structures play in action language processing is still unclear. “Embodied” and “disembodied” theories debate the nature of meaning representation in terms of the necessity of motor structures, neglecting the fact that the conditions of their activation during language processing are not well-described. Very recent research has begun to note the necessity of exploring the context under which words trigger modality-specific cortical activity. However, this trend is at odds with implicit theoretical assumptions that have been made in research on motor-language crosstalk, which are based on the “two-step” model of semantic processing and the “dictionary-like” view of lexical meaning representation. Within such framework, word meaning recognition is taken to proceed in a modular fashion. Only after this process has concluded is the context thought to exert its effects. These assumptions have biased the debate on the role of language induced motor activity. The discussion has been centered on whether motor activation should be considered an integral part of the lexical access process or taken as the result of an ensuing “higher order” operation (i.e., situation model construction). A large body of work evidences that lexical semantic processing and semantic context are far more integrated and interdependent. It seems crucial to integrate this knowledge gained from psycholinguistics into the research on the role of language induced motor activity. In an effort to liberate the debate from the “lexical vs. post-lexical” discussion, this thesis aimed at determining the conditions under which language triggers motor activity. To accomplish these objectives, we introduced a novel tool that analyzes on-line modulations of grip-force while participants listened to specific target words embedded within different types of contexts. Our results show that when the target word was a hand action verb and the sentence focus centered on that action (“John signs the contract”), an increase of grip force was observed in the temporal window classically associated with lexical semantic processing. No comparable increase in grip force was detected when the same action word was embedded in negative sentences (“John doesn’t sign the contract”) or in sentences which focus was shifted towards the agent’s mental state (“John wants to sign the contract”). Our results suggest that the presence of an action word in an
Trybocki, Christine. "Elaboration d'un modèle conceptuel pour les bases de données lexicales." Aix-Marseille 3, 1995. http://www.theses.fr/1995AIX30088.
Full textMuniz, Juliana Aguiar. "Processos de indeterminação lexical em conversas telefônicas interceptadas." Universidade do Estado do Rio de Janeiro, 2013. http://www.bdtd.uerj.br/tde_busca/arquivo.php?codArquivo=5340.
Full textO objetivo principal deste trabalho é estudar estratégias de indeterminação de sentido em um corpus de conversas telefônicas interceptadas, considerando que a produção de sentido é um processo cognitivo dependente do contexto. Delimitamos a linguística cognitiva como a área na qual essa pesquisa se encontra inserida, para melhor compreender os fundamentos e os pressupostos norteadores da Teoria dos Modelos Cognitivos Idealizados (TMCI) e da Teoria da Mesclagem Conceptual (blending), tendo como base, principalmente, os estudos de Lakoff (1987), Fauconnier (1997) e Fauconnier e Turner (2002). No decorrer do trabalho propomo-nos responder às seguintes questões de pesquisa: a) que estratégias de indeterminação de sentido são mais frequentemente usadas nestas conversas? b) que elementos do contexto e do cotexto permitem a delimitação do sentido do item lexical em determinada conversa? c) como funcionam, no corpus, as estratégias de indeterminação de sentido e de que forma elas contribuem para sustentar determinado tipo de relação interpessoal? Para responder a estas questões de pesquisa, das 22 gravações de conversas telefônicas de atores sociais envolvidos com tráfico de armas e drogas, sequestro e extorsão, fornecidas pela Coordenadoria de Segurança e Inteligência do Ministério Público do Rio de Janeiro, selecionamos 10 conversas, em função da sua qualidade sonora, para serem transcritas e para proceder à análise qualitativa do uso da polissemia e da vagueza lexical. A partir das discussões teóricas e das análises desenvolvidas, concluímos que a polissemia representa a estratégia de indeterminação de sentido mais frequente no corpus desta pesquisa e que a mesma pode ser entendida como um processo de mesclagem conceptual, que sofre influências sociais e culturais: é a dinamicidade do pensamento e da linguagem que geram a polissemia. Concluímos também que a vagueza lexical é utilizada, no corpus, como um recurso linguístico para referência a assuntos ilícitos. Os itens lexicais analisados instanciam esquemas mentais abstratos que têm seus sentidos realizados a partir de pistas linguísticas e extralinguísticas que apontam para um processo interacional que pode ser entendido como um enquadre de transações comerciais (tráfico de drogas)
The main objective of this research is to study strategies of indeterminacy of meaning in a corpus of intercepted telephone conversations by social actors involved with the trafficking of drugs and weapons, with kidnapping and extortion. We elected Cognitive Linguistics as the area in which this research should be developed, as we understand the process of meaning production as a cognitive process, dependent on the context. Within Cognitive Linguistics, we adopted the principles and assumptions guiding the Theory of Idealized Cognitive Models (TMCI) and Conceptual Blending Theory, based principally on studies by Lakoff (1987), Fauconnier (1997) and Fauconnier and Turner (2002). Throughout the paper our purpose is to answer the following research questions: a) what strategies for the indeterminacy of meaning are most often used in these conversations? b) what elements of context and co-text (the immediate grammatical context ) trigger the instantiation of the meaning of a lexical item in a particular conversation? c) how do the strategies of indetermination of meaning operate , in the corpus, and how do they contribute to the creation of a particular kind of interpersonal relationship? In order to answer these questions, from the 22 recordings provided by the Coordinator of Intelligence and Security of the Public Ministry of Rio de Janeiro, we selected 10 conversations, on the basis of their sound quality. We further transcribed them and submitted them to qualitative analysis, investigating the use of lexical polysemy and vagueness. From the theoretical discussions and analyzes undertaken, we conclude that polysemy represents the strategy of indeterminacy of meaning that is most often used in the corpus and that it can be understood as a process of conceptual blending, under the influence of social and cultural factors: it is the association between the use of language and the real dynamics of thought and language that generate polysemy. We also conclude that lexical vagueness is used as a language resource to refer to illicit affairs. The lexical items studied instantiate abstract mental schemas whose meanings are triggered by the use of particular linguistic and extralinguistic cues, within the domain, or frame, of a commercial transaction (drug trafficking)
Lara, Leandro Zanetti. "Um estudo acerca da representação semântico-lexical no modelo da gramática discursivo-funcional." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2012. http://hdl.handle.net/10183/49685.
Full textThis dissertation aims to present a study on lexical representation, in general, and on the treatment of lexical semantics, in particular, within the framework of Functional Discourse Grammar (HENGEVELD, 2000, 2004a, 2006; HENGEVELD; MACKENZIE, 2006, 2008) using data of Brazilian Portuguese, especially in relation to the syntactic behavior of adjectives regarding their semantic content. Since Functional Discourse Grammar is based on pragmatic adequacy, pursued since the FG by Dik (1978, 1997), which stipulates that the phonological and morphosyntatic configuration, that is, the structural encoding, is a result of pragmatic-semantic representations, our inquiry discussed the role performed by the semantic of lexical items in this model. The reason for this choice is the identification of linguistic phenomena of syntactic encoding, for instance, which are more linked to the semantic content of the lexical items instead of to phrasal or text semantics, that is, compositional semantics. Our point of support for the analysis was the syntactic behavior of Brazilian Portuguese adjectives which is closely related to the sense they present. We collected a corpus of art criticism in Brazilian Portuguese in order to analyze adjectival semantic subclasses that were acting in the examples and the way they were expressed syntactically. The data indicate that the lexical representation in Functional Discourse Grammar would be more precise in terms of explanatory power if it showed the internal organization of the lexicon, which seems to present rules of lexical formation as well as meaning definitions that directly motivate the syntactic behavior of the adjectival lexical items.
Esta tesis tiene como objetivo presentar un estudio sobre la representación léxica, en general, y el tratamiento de la semántica léxica, en particular, en la Gramática Discursivo-Funcional (HENGEVELD, 2000, 2004a, 2006; HENGEVELD; MACKENZIE, 2006, 2008), tomando como base para este estudio datos del portugués de Brasil, especialmente en relación con el comportamiento morfosintáctico de los adjetivos con respecto a su contenido semántico. Una vez que la Gramática Discursivo-Funcional fue construida bajo los auspicios de la adecuación pragmática, perseguida desde los días de la GF de Dik (1978, 1997), que estipula que la configuración fonológica y morfosintáctica, es decir, la estructura de codificación es el resultado de las representaciones pragmático-semánticas, nuestra investigación ha girado en torno al papel que la semántica de las unidades léxicas presentan en este modelo. Esto se debe al hecho de que identificamos fenómenos lingüísticos de codificación sintáctica, por ejemplo, que están vinculados directamente al contenido de la semántica lexical, y no solo de la semántica composicional. Nuestro punto de apoyo para el análisis fue el comportamiento sintáctico de los adjetivos del portugués de Brasil, que está estrechamente ligada al influjo del contenido léxico. Hemos recogido un corpus de crítica de arte en portugués de Brasil, para analizar las subclases semánticas presentes en los ejemplos y la forma en que se expresan sintácticamente. Los datos indican que la representación léxica de la Gramática Discursivo- Funcional ganaría en poder explicativo se tuviera también una representación de la organización interna del léxico, que parece contener reglas léxicas de formación, así como definiciones semánticas que motivan el comportamiento sintáctico de los elementos léxicos adjetivales.
Humphreys, Jane. "WIRDS, WERDS, WYRDZ : visual wordlikeness, lexical phonology, and models of visual word recognition." Thesis, University of Bristol, 2008. http://hdl.handle.net/1983/7f7f66b6-f4e7-420e-ab74-af32bbec6ce0.
Full textCoté, Myriam. "Utilisation d'un modele d'accès lexical et de concepts perceptifs pour la reconnaissance d'images de mots cursifs." Paris, ENST, 1997. http://www.theses.fr/1997ENST0009.
Full textKharrazen, Essaïd. "PSILISP, un modèle d'interprétation parallèle de programmes LISP." Paris 11, 1986. http://www.theses.fr/1986PA112385.
Full textPSILISP comprises the definition of a language derived from LISP and its implementation on an MIMD parallel architecture with shared memory. The main features of PSILISP are: lexically scoped variables, call by value, explicit parallel evaluation of the arguments of an application, primitives with no side effects. PSILISP language extends LISP by the new “parallel application” construct. Its evaluation leads to intensive use of the processors for the parallel computation of the arguments. PSILISP uses lexically scoped variables. This choice avoids the semantical defects that are common to the usual implementations of LISP. Furthermore, the implementation in this case lends itself better to the management of parallel evaluation. PSILISP brings a solution to the Funarg problem by environment retention. Functions become thus members of first class citizens. The PSILISP experience demonstrates that the efficiency of LISP programs can be increased considerably by use of the parallelism
He, Yanzhang. "Segmental Models with an Exploration of Acoustic and Lexical Grouping in Automatic Speech Recognition." The Ohio State University, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=osu1429881253.
Full textSwaileh, Wassim. "Des modèles de langage pour la reconnaissance de l'écriture manuscrite." Thesis, Normandie, 2017. http://www.theses.fr/2017NORMR024/document.
Full textThis thesis is about the design of a complete processing chain dedicated to unconstrained handwriting recognition. Three main difficulties are adressed: pre-processing, optical modeling and language modeling. The pre-processing stage is related to extracting properly the text lines to be recognized from the document image. An iterative text line segmentation method using oriented steerable filters was developed for this purpose. The difficulty in the optical modeling stage lies in style diversity of the handwriting scripts. Statistical optical models are traditionally used to tackle this problem such as Hidden Markov models (HMM-GMM) and more recently recurrent neural networks (BLSTM-CTC). Using BLSTM we achieve state of the art performance on the RIMES (for French) and IAM (for English) datasets. The language modeling stage implies the integration of a lexicon and a statistical language model to the recognition processing chain in order to constrain the recognition hypotheses to the most probable sequence of words (sentence) from the language point of view. The difficulty at this stage is related to the finding the optimal vocabulary with minimum Out-Of-Vocabulary words rate (OOV). Enhanced language modeling approaches has been introduced by using sub-lexical units made of syllables or multigrams. The sub-lexical units cover an important portion of the OOV words. Then the language coverage depends on the domain of the language model training corpus, thus the need to train the language model with in domain data. The recognition system performance with the sub-lexical units outperformes the traditional recognition systems that use words or characters language models, in case of high OOV rates. Otherwise equivalent performances are obtained with a compact sub-lexical language model. Thanks to the compact lexicon size of the sub-lexical units, a unified multilingual recognition system has been designed. The unified system performance have been evaluated on the RIMES and IAM datasets. The unified multilingual system shows enhanced recognition performance over the specialized systems, especially when a unified optical model is used
Neff, Kathryn Joan Eggers. "Neural net models of word representation : a connectionist approach to word meaning and lexical relations." Virtual Press, 1991. http://liblink.bsu.edu/uhtbin/catkey/832999.
Full textDepartment of English
Dugua, Céline. "Liaison, segmentation lexicale et schémas syntaxiques entre 2 et 6 ans : un modèle développemental basé sur l'usage." Grenoble 3, 2006. https://hal.archives-ouvertes.fr/tel-01272976.
Full textThis thesis focuses on the acquisition of liaison by french children aged between 2 and 6. Through cognitive functional approaches, more specifically on usage-based models and construction grammars, our analyses highlight how linguistic levels (phonological, lexical, syntactic) interact during the development. From 8 corpora studies as well as a measurement of errors in liaison contexts, taken from a child's utterances, we elaborated 6 experimental study protocols, in particular, a four-year longitudinal follow up of 20 children, as well as 2 cross-studies with larger samples (122 and 200 subjects). We suggest a 3 stage developmental model integrating liaison phenomenon, lexical segmentation and constructional schemas emergence. Precociously, the child would retrieve concrete linguistic sequences in her linguistic environment. She would then memorise these sequences and store them in her lexicon in same form as that one heard. For example, she could memorise sequences like un âne (a donkey), l'âne (with determiners), zâne, nâne (with consonant liaison on the initial). These concrete sequences constitute the base from which more abstract schemas emerge progressively. The first are general, integrating a determiner (the pivot) and a slot which can receive any lexical forms. They are like un (a/an) + X and they explain early frequent substitution errors (like un zâne). Gradually, these schemas become more specific, integrating the phonetic nature of the liaison consonant: un + nX. Their application explains progress in liaison contexts and overgeneralization errors on words starting with a consonant (like un nèbre instead of un zèbre (zebra))
Zhang, Ying. "Modèles et outils pour des bases lexicales "métier" multilingues et contributives de grande taille, utilisables tant en traduction automatique et automatisée que pour des services dictionnairiques variés." Thesis, Université Grenoble Alpes (ComUE), 2016. http://www.theses.fr/2016GREAM017/document.
Full textOur research is in computational lexicography, and concerns not only the computer support to lexical resources useful for MT (machine translation) and MAHT (Machine Aided Human Translation), but also the linguistic architecture of lexical databases supporting these resources in an operational context (CIFRE thesis with L&M).We begin with a study of the evolution of ideas in this area, since the computerization of classical dictionaries to platforms for building up true "lexical databases" such as JIBIKI-1 [Mangeot, M. et al., 2003 ; Sérasset, G., 2004] and JIBIKI-2 [Zhang, Y. et al., 2014]. The starting point was the PIVAX-1 system [Nguyen, H.-T. et al., 2007 ; Nguyen, H. T. & Boitet, C., 2009] designed for lexical bases for heterogeneous MT systems with a lexical pivot, able to support multiple volumes in each "lexical space", be it natural or artificial (as UNL). Considering the industrial context, we focused our research on some issues, in informatics and lexicography.To scale up, and to add some new features enabled by JIBIKI-2, such as the "rich links", we have transformed PIVAX-1 into PIVAX-2, and reactivated the GBDLEX-UW++ project that started during the ANR TRAOUIERO project, by re-importing all (multilingual) data supported by PIVAX-1, and making them available on an open server.Hence a need for L&M for acronyms, we expanded the "macrostructure" of PIVAX incorporating volumes of "prolexemes" as in PROLEXBASE [Tran, M. & Maurel, D., 2006]. We also show how to extend it to meet new needs such as those of the INNOVALANGUES project. Finally, we have created a "lemmatisation middleware", LEXTOH, which allows calling several morphological analyzers or lemmatizers and then to merge and filter their results. Combined with a new dictionary creation tool, CREATDICO, LEXTOH allows to build on the fly a "mini-dictionary" corresponding to a sentence or a paragraph of a text being "post-edited" online under IMAG/SECTRA, which performs the lexical proactive support functionality foreseen in [Huynh, C.-P., 2010]. It could also be used to create parallel corpora with the aim to build MOSES-based "factored MT systems"
Coté, Myriam. "Utilisation d'un modèle d'accès lexical et de concepts perceptifs pour la reconnaissance d'images de mots récursifs /." Paris : École nationale supérieure des télécommunications, 1997. http://catalogue.bnf.fr/ark:/12148/cb367038172.
Full textHasan, Saša [Verfasser]. "Triplet lexicon models for statistical machine translation / Sasa Hasan." Aachen : Hochschulbibliothek der Rheinisch-Westfälischen Technischen Hochschule Aachen, 2012. http://d-nb.info/1028004060/34.
Full textChan, May T. M. "Alveolarization in Hong Kong Cantonese : a sociophonetic study of neogrammarian and lexical diffusion models of sound change." Thesis, University of Oxford, 2017. https://ora.ox.ac.uk/objects/uuid:2d40e687-83cd-4d93-9c3e-fa6e5569cf6b.
Full textSouza, Marcos Antônio de. "O dicionário de hebraico bíblico de Brown, Driver e Briggs (BDB) como modelo de sistema lexical bilíngüe." reponame:Repositório Institucional da UFSC, 2012. http://repositorio.ufsc.br/xmlui/handle/123456789/90857.
Full textMade available in DSpace on 2012-10-23T16:05:59Z (GMT). No. of bitstreams: 1 273667.pdf: 1385993 bytes, checksum: 2c3ea5f466a7c2a004fa49dd996aa441 (MD5)
Esta dissertação de mestrado foi concebida como um trabalho original relacionado à lexicografia bíblica hebraica. O problema abordado neste estudo pode ser formulado da seguinte maneira: O estudo procura investigar o popular dicionário de hebraico bíblico de Brown, Driver e Briggs (BDB) baseado em um dos mais antigos dicionários e editado por volta de 1810 por Wilhelm Gesenius como um modelo de sistema lexical bilíngüe no contexto da polêmica envolvendo glosas, definições e domínios semânticos. Partindo da hipótese de que ciência é qualquer conhecimento obtido pelo Método Científico, este estudo está estruturado em uma série de três passos. No primeiro passo observação a lexicografia hebraica bíblica é investigada mediante uma identificação dos principais dicionários de hebraico bíblico publicados em língua inglesa e portuguesa quanto as suas macro-estrutura e micro-estrutura. No segundo passo, uma hipótese é formulada. Partindo do conceito de que lexicografia é uma disciplina de aplicação em que o propósito vem primeiro e posteriormente a teoria, formula-se a hipótese de que um dicionário de hebraico bíblico é um modelo de um sistema lexical bilíngüe. Como fundamento teórico para esta hipótese, é apresentada uma analogia com um dos mais bem sucedido modelo de sistema e utilizado pelos engenheiros de telecomunicações o modelo de Erlang e o conceito de cadeia de transferência desenvolvido a partir do triângulo da significação de Ogden & Richards. O terceiro e último passo consiste em um experimento apropriado para verificação da validade da hipótese. Neste experimento, quatro poemas hebraicos (dois bíblicos, um medieval e um moderno) são traduzidos segundo as glosas fornecidas pelo BDB e comparadas às glosas de quatro outros dicionários de hebraico, além de uma comparação com antigas traduções da Bíblia Hebraica (Septuaginta e Vulgata) para os dois poemas bíblicos.
This Masters degree dissertation was conceived as an original work mainly concerned to Biblical Hebrew Lexicography. The basic problem in this study is to be formulated as follows: The study seeks to investigate the well-know Biblical Hebrew dictionary by Brown, Driver, and Briggs (BDB) based in one of the oldest Hebrew dictionary published around 1810 by Wilhelm gesenius as a model of a lexical bilingual system in the context of the polemic involving glosses, definitions and semantic domains. Proceeding on the assumption that Science is any knowledge that is arrived at by the Scientific Method, this study is structured on a series of three definite steps. In the first step observation modern Biblical Hebrew Lexicography is observed by a survey of Hebrew dictionaries published in English and Portuguese languages with a analysis of their macrostructure and microstructure. In the second step, a hypothesis is formulated. On the assumption that lexicography is an applied discipline in which the purpose comes first and the theory comes last, a hypothesis is formulated in which a Biblical Hebrew Dictionary is a model of a lexical bilingual system. As a framework for this hypothesis, a analogy is made with one of the most successful system model used by telecommunication engineers the Erlangs model and the concept of chain of transference is developed based on the Ogden & Richards triangle of signification. The third and final step is an appropriate experiment to see if the hypothesis is ubstantiated. In this experiment, four Hebrew poems (two biblical poems, one medieval poem and one modern poem) are translated according the glosses provided by BDB and compared to the glosses of four other Hebrew dictionaries besides a comparison to Hebrew Bible ancient versions (Septuagint and Vulgate) for the two biblical poems.
Laporte, Elena-Mirabela. "La traduction automatique statistique factorisée : une application à la paire de langues français - roumain." Thesis, Strasbourg, 2014. http://www.theses.fr/2014STRAC022/document.
Full textOur first aim is to build linguistic resources for a French - Romanian factored phrase - based statistical machine translation system. Our second aim is to study the impact of exploited linguistic information in the lexical alignment and translation process. On the one hand, this study is motivated by the lack of such systems for the studied languages. On the other hand, it is motivated by the high number of errors provided by the current machine translation systems. The linguistic resources required by the system are tokenized, lemmatized, tagged, word, and sentence - aligned parallel corpora
Zhang, Xuelu. "Les tons lexicaux du chinois mandarin en voix modale et en voix chuchotée." Thesis, Strasbourg, 2017. http://www.theses.fr/2017STRAC041/document.
Full textOur research is a contribution to studies on secondary acoustic cues in Mandarin tone identification, by comparing acoustic data collected in modal speech and in whispered speech. According to the literature on the same issue, theses cues could be found in acoustic dimensions other than in the fundamental frequency, as a set of attributes. We have analyzed these attributes in the temporal domain, at the intensity level and in the spectrum, as well as their relations with tones. Our results show that some temporal parameters and the fourth resonance in the spectrum are very closely related to tones. These relations are dependent on the intrinsic characteristics of the vowel that carries the tone (which is equivalent to the rime in our research)
Azevedo, Luciana de Oliveira Faria. ""Uma flor tapoja e uma casa jufosa: o papel da nomeação e de propriedades morfofonológicas no processo de identificação de novos adjetivos por crianças brasileiras"." Universidade Federal de Juiz de Fora (UFJF), 2008. https://repositorio.ufjf.br/jspui/handle/ufjf/4607.
Full textApproved for entry into archive by Adriana Oliveira (adriana.oliveira@ufjf.edu.br) on 2017-05-19T15:55:07Z (GMT) No. of bitstreams: 1 lucianadeoliveirafariaazevedo.pdf: 2448500 bytes, checksum: eaee6df3b32bee6f42619f40a226e17f (MD5)
Made available in DSpace on 2017-05-19T15:55:07Z (GMT). No. of bitstreams: 1 lucianadeoliveirafariaazevedo.pdf: 2448500 bytes, checksum: eaee6df3b32bee6f42619f40a226e17f (MD5) Previous issue date: 2008-08-29
Esta dissertação aborda o processo de aquisição lexical por crianças brasileiras e investiga, particularmente, a relação entre categoria conceitual e categoria lingüística, e propriedades morfofonológicas do adjetivo. A hipótese que orienta esta dissertação é a de que a nomeação dos objetos e a presença de morfemas característicos de adjetivos são pistas robustas usadas pelas crianças no processo de aquisição de novos adjetivos. Adota-se uma perspectiva psicolingüística da aquisição da linguagem que pretende a conciliação de um modelo de processamento lingüístico (modelos de Bootstrapping Fonológico e Sintático), com um modelo de língua proposto pela Teoria Gerativa. A conciliação entre os modelos visa a explicar, satisfatoriamente, a forma pela qual a criança se torna capaz de, uma vez exposta a uma língua natural, extrair do material lingüístico ao qual é apresentada os elementos formadores do léxico de sua língua. Foram desenvolvidas duas atividades experimentais, usando-se a técnica de identificação de objeto, com crianças de dois e três anos. A primeira avalia o reconhecimento de novos adjetivos, comparando-se a apresentação de objetos nomeados (uma flor tapoja) ou com nomes vagos (uma coisa tapoja). No segundo experimento, foram acrescentados aos pseudo-adjetivos os sufixos -oso/a e –ado/a (uma casa jufosa) / uma coisa jufosa), em vista de investigar o papel do sufixo juntamente com a nomeação dos objetos como facilitadores na identificação do adjetivo pela criança. Adjetivos acompanhados de nome (Exper. 1) são mais facilmente identificados, mas quando acrescidos de sufixo (Exper. 2) são reconhecidos mesmo na presença de nomes vagos. Os resultados são compatíveis com nossa hipótese, pois sugerem que a nomeação e a marca morfofonológica são pistas robustas usadas pelas crianças para identificar novos adjetivos.
This dissertation approaches the process of lexical acquisition for Brazilian children and it investigates, particularly, the relationship among conceptual category and linguistic category, and properties morphophonological of the adjective. The hypothesis that guides this dissertation is the one that the nomination of the objects and the presence of morphemes characteristic of adjectives are robust tracks used by the children in the process of acquisition of new adjectives. A perspective psycholinguistic of the acquisition of the language is adopted that intends the conciliation of a model of linguistic processing (models of Phonological and Syntactic Bootstrapping), with a language model proposed by the Generative Theory. The conciliation among the models seeks to explain, satisfactorily, the form for the which the child becomes capable of, once exposed to a natural language, to extract of the linguistic material to which is presented the elements that form the lexicon of your language. Two experimental activities were developed, being used the technique of object identification, with two three year-old children. The first evaluates the recognition of new adjectives, being compared the presentation of nominated objects (a tapoja flower) or with vague names (a tapoja thing). In the second experiment, they were increased to the pseudo-adjectives the suffixes -oso/a and -ado/a (a jufosa house) / a jufosa thing), in view of investigating the paper of the suffix together with the nomination of the objects as facilitators in the identification of the adjective for the child accompanied Adjectives of name (Exper. 1) they are more easily identified, but when added of suffix (Exper. 2) they are recognized even in the presence of vague names. The results are compatible with our hypothesis, because they suggest that the nomination and the mark morphophonological are robust tracks used by the children to identify new adjectives.
Séguéla, Patrick. "Construction de modèles de connaissances par analyse lingustiques de relations lexicales dans les documents techniques." Toulouse 3, 2001. http://www.theses.fr/2001TOU30210.
Full textLowry, Jonathan E. "The Language of Team: Building a lexicon integrating multiple disciplines for effective project management." University of Cincinnati / OhioLINK, 2011. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1306499898.
Full textEklund, Robert. "A Probabilistic Tagging Module Based on Surface Pattern Matching." Thesis, Stockholm University, Department of Computational Linguistics, Institute of Linguistics, 1993. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-135294.
Full textSHIMA, Yoshihiro, and 義弘 島. "内的作業モデルが情報処理に及ぼす影響 : プライムされた関係との関連." 名古屋大学大学院教育発達科学研究科, 2012. http://hdl.handle.net/2237/16160.
Full textAdelstein, Andreina. "Unidad léxica y significado especializado: modelo de representación a partir del nombre relacional madre." Doctoral thesis, Universitat Pompeu Fabra, 2007. http://hdl.handle.net/10803/7505.
Full textEl trabajo presenta una revisión crítica de las propuestas lingüísticas y terminológicas acerca de las propiedades semánticas del léxico científico. Luego, analiza las propiedades del significado léxico especializado, los factores y los mecanismos de generación semántica, a partir del análisis contrastivo del comportamiento de madre en corpora textuales. La tesis proporciona, a su vez, criterios de reconocimiento formal de información semántica especializada, útiles para desarrollar diversos tipos de aplicaciones. Finalmente, propone una generalización de la semántica especializada de los nombres relacionales y una representación de entrada dinámica, que contempla componentes de conocimiento lingüístico y extralingüístico que interactúan en la generación del significado léxico.
This dissertation studies the specificity of specialized lexical meanings, based on an analysis of the behaviour of relational nouns; it offers an integral explanation of specialized and non-specialized semantics and a modellization of a unified lexical entry. Thus, it combines the communicative model of terminology with polysemic models of generation of lexical meaning.
This work starts by presenting a critical review of linguistic and terminological approaches to the semantic properties of scientific lexicon. Then it procedes to analize the properties of specialized lexical meaning, the factors that influence semantic generation and its mechanism, based on the behaviour of madre in text corpora. It also provides criteria for the formal recognition of specialized semantic information which can help develop different kinds of applications. Finally, the dissertation puts forward a generalization of the specialized semantics of relational nouns and a representation of a dynamic entry, which contemplates components of linguistic and extralinguistic knowledge interacting in the generation of lexical meaning.
Jacob, Bruno. "Un outil informatique de gestion de modèles de Markov cachés : expérimentations en reconnaissance automatique de la parole." Toulouse 3, 1995. http://www.theses.fr/1995TOU30240.
Full textNobrega, Karliane Fernandes. "A interpreta??o sem?ntica dos auxiliares modais poder, precisar e dever: uma abordagem da sem?ntica cognitiva." Universidade Federal do Rio Grande do Norte, 2007. http://repositorio.ufrn.br:8080/jspui/handle/123456789/16359.
Full textApresentamos, neste trabalho, com base na sem?ntica cognitiva, uma an?lise do significado, em contexto, dos auxiliares modais poder, precisar e dever. Analisamos 120 textos produzidos por candidatos ao vestibular e por alunos do ensino fundamental, como resposta da quest?o n?mero tr?s da prova discursiva de L?ngua Portuguesa do vestibular 2005 da UFRN, que pede aos candidatos para explicitar a diferen?a de sentido entre tr?s frases, observando o uso desses tr?s verbos. Consideramos que um item lexical n?o ? incorporado a uma representa??o ling??stica sem?ntica fixa, limitada e ?nica, mas antes, ? ligado a uma representa??o ling??stica sem?ntica flex?vel e aberta que prov? acesso a muitas concep??es e sistemas conceituais dependente de cada contexto determinado. Com base em seu significado, um item lexical evoca um grupo de dom?nios cognitivos, que por sua vez, apresentam um determinado conte?do conceitual. Isto implica em afirmar que a rede de significados lexicais vai variar conforme o conhecimento de mundo de cada um (LANGACKER, 2000). A relev?ncia deste trabalho ? proporcionar uma contribui??o para a descri??o sem?ntica do portugu?s
We present, in this work, based on cognitive semantics, an analysis of the meaning in context of the modal auxiliaries can, need and must. We analysed 120 texts produced by applicants for university entrance examinations and primary school students as answer to question number three of the Portuguese Language discursive test, in the entrance examinations for UFRN, that asked the candidates to explicit the difference in meaning between three sentences, observing the use of those three verbs. We consider that a lexical item is not incorporated by a steady semantic structure, limited and unique, but instead, it is linked to an open and flexible linguistic semantic representation that provides access to many conceptions and conceptual systems depending on each determined context. Based on its meaning, a lexical item evokes a group of cognitive domains, which present a determined conceptual content. This makes possible to affirm that the net of lexical meanings will vary according to the world knowledge each one has (LANGACKER, 2000). The relevance of this work is provide a understanding of the semantic decription of portuguese
Carter, Kelli Patrice. "Investigating Student Conceptual Understanding of Structure and Function by Using Formative Assessment and Automated Scoring Models." Scholar Commons, 2019. https://scholarcommons.usf.edu/etd/7761.
Full textBelkacem, Thiziri. "Neural models for information retrieval : towards asymmetry sensitive approaches based on attention models." Thesis, Toulouse 3, 2019. http://www.theses.fr/2019TOU30167.
Full textThis work is situated in the context of information retrieval (IR) using machine learning (ML) and deep learning (DL) techniques. It concerns different tasks requiring text matching, such as ad-hoc research, question answering and paraphrase identification. The objective of this thesis is to propose new approaches, using DL methods, to construct semantic-based models for text matching, and to overcome the problems of vocabulary mismatch related to the classical bag of word (BoW) representations used in traditional IR models. Indeed, traditional text matching methods are based on the BoW representation, which considers a given text as a set of independent words. The process of matching two sequences of text is based on the exact matching between words. The main limitation of this approach is related to the vocabulary mismatch. This problem occurs when the text sequences to be matched do not use the same vocabulary, even if their subjects are related. For example, the query may contain several words that are not necessarily used in the documents of the collection, including relevant documents. BoW representations ignore several aspects about a text sequence, such as the structure the context of words. These characteristics are important and make it possible to differentiate between two texts that use the same words but expressing different information. Another problem in text matching is related to the length of documents. The relevant parts can be distributed in different ways in the documents of a collection. This is especially true in large documents that tend to cover a large number of topics and include variable vocabulary. A long document could thus contain several relevant passages that a matching model must capture. Unlike long documents, short documents are likely to be relevant to a specific subject and tend to contain a more restricted vocabulary. Assessing their relevance is in principle simpler than assessing the one of longer documents. In this thesis, we have proposed different contributions, each addressing one of the above-mentioned issues. First, in order to solve the problem of vocabulary mismatch, we used distributed representations of words (word embedding) to allow a semantic matching between the different words. These representations have been used in IR applications where document/query similarity is computed by comparing all the term vectors of the query with all the term vectors of the document, regardless. Unlike the models proposed in the state-of-the-art, we studied the impact of query terms regarding their presence/absence in a document. We have adopted different document/query matching strategies. The intuition is that the absence of the query terms in the relevant documents is in itself a useful aspect to be taken into account in the matching process. Indeed, these terms do not appear in documents of the collection for two possible reasons: either their synonyms have been used or they are not part of the context of the considered documents. The methods we have proposed make it possible, on the one hand, to perform an inaccurate matching between the document and the query, and on the other hand, to evaluate the impact of the different terms of a query in the matching process. Although the use of word embedding allows semantic-based matching between different text sequences, these representations combined with classical matching models still consider the text as a list of independent elements (bag of vectors instead of bag of words). However, the structure of the text as well as the order of the words is important. Any change in the structure of the text and/or the order of words alters the information expressed. In order to solve this problem, neural models were used in text matching
Balza, Tardaguila Irene. "Syntactic structure and modal interpretation : the case of Basque "behar"." Thesis, Bordeaux 3, 2018. http://www.theses.fr/2018BOR30070.
Full textThis dissertation is an investigation of the syntactic structure and modal interpretation of clauses involving the denominal necessity predicate behar ‘need’ and an infinitival complement. On the one hand, it analyses the syntactic status of non-finite complements of denominal behar by examining their interaction with syntactic phenomena sensitive to different structural and locality conditions, and concludes that the infinitival complements of behar can correspond to different underlying structures. The largest type of infinitive is a non-restructuring infinitive that projects a full clausal architecture (i.e. a CP), and the smallest one is a reduced restructuring infinitive that projects up to vP. There is evidence for intermediate types projecting up to the inflectional domain (IP/TP). On the other hand, the dissertation examines the thematic and scope properties of the subjects in each of the different structural types and the modal interpretation that they can give rise to. On the basis of this analysis it is argued that modal interpretation is not constrained by any single factor (the presence of restructuring, the referential status of the subject and its relative scope vis-à-vis the modal predicate, among other frequently mentioned ones), but depends on the cumulative effect of several factors working together. The dissertation also shows the necessity of adopting a more fine-grained view of root modality, one that allows a simpler mapping of syntactic structures into modal meanings
Bosse, Marie-Line. "L'acquisition et la mobilisation des connaissances lexicales orthographiques : tests d'hypothèses développementales issues du modèle de lecture de Ans, Carbonnel et Valdois (1998)." Grenoble 2, 2004. http://www.theses.fr/2004GRE29038.
Full textThis research studies the acquisition of orthographic knowledge. Hypotheses have been made from the multi-trace model of expert reading of Ans, Carbonnel, and Valdois (1998). A first series of experiments permitted to confirm the existence of an analogical process on children. The studies showed that children can acquire lexical orthographic knowledge from the beginning of literacy learning, and that this knowledge can be activated during the processing of new words. The next part of the research tests the hypothesis that the acquisition of orthographic knowledge depends not only on phonological processing but also on visual attentional processing efficacy. To do so, a second series of experiments studies the impairment of visual attentional processing on dyslexic children. They evidence, on some dyslexic children, both a visual attentional processing impairment and an important deficit of lexical orthographic knowledge. A last series of studies examined the involvement of visual attentional processing in the normal acquisition of orthographic knowledge. Results show that visual attentional processing is highly predictive of orthographic knowledge, for children from first to fifth grade, and even after the control of the part predicted by phonological processing, IQ and verbal short term memory. These studies on large samples of children, with coherent results on reading and spelling, provide convincing arguments for the hypothesis that the acquisition of orthographic knowledge depends not only on phonological processing but also on visual attentional processing efficacy
Santos, Anderson Roberto Santos dos. "A computational investigation of verbs during aging with and without Alzheimer’s disease." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2011. http://hdl.handle.net/10183/119124.
Full textAlzheimer’s disease produces alterations of cognitive functions and of processes that are responsible for language and memory. In order to have a better understanding of language changes, we investigate the characteristics of the semantic networks of patients diagnosed with probable Alzheimer, focusing on verbs. The results of comparisons with networks of healthy individuals and patients with Alzheimer disease highlight some topological differences among them. We also constructed classifiers that could capture the differences between the various profiles of speakers, and that can be used to classify unknown speakers according to the closest profile. We made this effort in order to help the diagnosis of diseases that affect language, such as the Alzheimer’s disease.
Aymoré, Debora de Sá Ribeiro. "O modelo de Historiografia da ciência Kuhniano: da obra A estrutura das revoluções científicas aos ensaios tardios." Universidade de São Paulo, 2010. http://www.teses.usp.br/teses/disponiveis/8/8133/tde-26102010-093744/.
Full textThe central aim of our work is to critically examine the central aspects of the historiography of science proposed by Thomas Kuhn (1922-1996). To achieve this goal, we will begin our examination with The structure of scientific revolutions (1962), which contains the first more complete formulation about the structure of scientific development, along with the Postscript of 1969. Then we will also investigate some essays in the collections The essential tension (1977) and The road since Structure (2000). After the analysis we realize that Kuhn\'s historiography is based on the postulate of the real history of science and the assumptions of the relationship between history and philosophy of science, the centrality of the paradigm, the plurality of readings of text and the relationship between internal and external history of science.
Nobre, Alexandre de Pontes. "Processamento léxico-semântico : relações com reconhecimento visual de palavras e compreensão de leitura textual." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2013. http://hdl.handle.net/10183/101860.
Full textThe aim of this dissertation was to investigate the relationships between word recognition and reading comprehension with lexical-semantic processing. The dissertation is composed of two studies. In the first study, models of word reading and reading comprehension are reviewed in order to examine the role of lexical-semantic processing in visual word recognition and in reading comprehension. The semantic priming paradigm is presented as an instrument for the investigation of relationships between lexical-semantic processing and the components of reading examined. The main theoretical models of semantic priming are presented and a review of studies which relate semantic priming and reading is conducted, and some conclusions and perspectives for investigation are presented. In the second study, relations between lexical-semantic processing and reading (visual word recognition and reading comprehension) were investigated empirically in a sample of 68 children, aged seven to twelve years, from private schools in Porto Alegre, Brazil. Lexical-semantic processing was evaluated by a lexical decision task in the semantic priming paradigm and reading abilities were assessed with a word/nonword reading task and a reading comprehension task (questionnaire and story retelling). Correlations between semantic priming effects and word reading and reading comprehension were investigated, as well as if semantic priming effects predict performance on the reading task. Results showed that semantic priming correlates with both groups of reading measures, and that word reading partially mediates the relation between lexical-semantic processing and reading comprehension.
Oliveira, Ana Flávia Souto de. "A multiplicidade semântica em learners' dictionaries : por uma abordagem semântico-cognitiva para a organização das acepções." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2015. http://hdl.handle.net/10183/130772.
Full textSemantic multiplicity can be defined as a case in which a single linguistic form presents more than one meaning or contextual reading. Even though this phenomenon usually does not pose serious challenges for everyday communication, it certainly brings about important issues to lexical semantic theories and its linguistic applications. The present dissertation aims at evaluating the treatment semantic multiplicity receives in English monolingual advanced learners’ dictionaries from a cognitive-semantic point of view. Therefore, the consequences of semantic multiplicity for the organization of learners’ dictionaries are presented, mainly with regard to the procedures of lumping and splitting, the solutions applied for structuring entries (homonymic or polysemous solutions), and the criteria used in arranging senses. First, it is demonstrated that there are no solid methodological bases on which to decide on how many (nor which) senses a lexical item has. Second, it is shown that there is no agreement on the solution to be applied to this type of dictionary. Third, it is advocated that the criterion used for sense arrangement (frequency) is not as objective as one would expect and it has not yet been proven to bring any advantage for the users of learners’ dictionaries. Because of the interpretative nature of lexical semantic description, which is reflected in the different treatments the dictionaries provide to these matters, a distinct approach is sought. To this end, the cognitive-semantic conception of semasiological structure is introduced. With its origins linked to Prototype Theory tenets, this notion highlights that semantic salience and overlapping are structural characteristics of the lexicon, which reflect the flexibility and instability of lexical meaning. In order to search for methods that could ground an alternative proposal for these lexicographic issues, the cognitive-semantic descriptive models of semantic structure are assessed: the radial model, the schematic model, and the overlapping sets model. The semasiological structure of the lexical item case is described and a new proposal for its organization is provided, in tune with cognitive-semantic tenets and with what is known about this type of dictionary and its users’ needs. Regarding homonymic and polysemous solutions, four different arrangements are suggested, which represent the coherent structure of polysemy. Regarding sense arrangement, it is shown that even through the use of a hierarchical structure, it is possible to represent the semantic overlapping and salience found to be useful for the users of learners’ dictionaries. By exploring the hierarchical resources themselves and by manipulating the wording of definitions, it is feasible to accentuate attributes shared by two or more senses that cannot be related in the linear structure of the dictionary entry. Thus, Cognitive Semantics presents itself as a useful approach to guide lexicographic practices related to the structuring of semantic multiplicity information in learners’ dictionaries.
Khuc, Vinh Ngoc. "Approaches to Automatically Constructing Polarity Lexicons for Sentiment Analysis on Social Networks." The Ohio State University, 2012. http://rave.ohiolink.edu/etdc/view?acc_num=osu1343187623.
Full text