Academic literature on the topic 'Lexical analyze'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Lexical analyze.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Lexical analyze"

1

Hijjah, Shifany Maulida, Mukhshon Nawawi та Fatima El Zahraa. "تحليل التماسك المعجمي في سورة القيامة". Kalimatuna: Journal of Arabic Research 2, № 2 (2024): 215–24. http://dx.doi.org/10.15408/kjar.v2i2.28412.

Full text
Abstract:
This research is motivated by Sayyid Qutub’s assumption in his book (Fii Zilaal Al-Qur’an) that meaningful cohesion study has been carried out in the Al-Qur’an, but has not been studied literally and lexically. There was the impression that the lexical cohesion was the important thing to understand the meaning of Al-Qur’an. Lexical cohesion is a cohesion that associated with meaning and formed by the semantic relationship between another lexicon. In addition to the researcher chose the Surah Al-Qiyamah as the object of research to analyze the lexical cohesion devices that contained in Surah Al-Qiyamah by using Halliday and Ruqaiya’s theory of lexical cohesion. Halliday and Ruqaiya see lexical cohesion as being formed of two aspects: reiteration (repitition, synonym, antonym, hyponym) and collocation. Therefore, this research is trying to answer how many devices of lexical cohesion are contained in Surah Al-Qiyamah.
APA, Harvard, Vancouver, ISO, and other styles
2

Lestariningsih, Depi, and Prembayun Miji Lestari. "Lexical semantic study of the Bubakan Manten Tradition in Mojodelik Village, Bojonegoro, East Java, Indonesia." LITE: Jurnal Bahasa, Sastra, dan Budaya 20, no. 2 (2024): 149–63. http://dx.doi.org/10.33633/lite.v20i2.11119.

Full text
Abstract:
This research aims to analyze the lexicon contained in the bubakan manten traditional procession in Mojodelik village on AM Production's YouTube account and analyze the lexical and cultural meaning contained in the bubakan manten traditional procession in Mojodelik village on AM Production's YouTube account. This research uses a qualitative descriptive method with a lexical semantic approach. The data in this research is in the form of a lexicon contained in the bubakan manten tradition, which contains lexical and cultural meanings. The primary data source in this research is the video of the bubakan manten tradition in Mojodelik Village, which is available on the AM Production Purwosari YouTube account, entitled Live Streaming Season 2 Langen Tayub Mustiko Laras Mojodelik Gayam, as well as interviews with the resource person who served as the leader in the implementation of bubakan manten, namely Mr. Jimin and Mr. Pandholi. Data collection techniques in this research used note-taking and interview techniques. Data analysis techniques used include sorting data for analysis, classifying lexicon forms, explaining the lexical and cultural meaning of the lexicon, and conclusion. The results of the research are a lexicon of places, equipment, food, and plants. The lexical meaning of the bubakan tradition is to open the way for marriage, while the cultural meaning contained in the bubakan tradition is to ask for safety and to be kept away from bad things.
APA, Harvard, Vancouver, ISO, and other styles
3

Hameed, Samia, Imran Ali, and Khalid -. "Analysis of Lexical Density and Readability of Pakistani English Newspaper Dawn." Global Mass Communication Review IX, no. II (2024): 80–92. http://dx.doi.org/10.31703/gmcr.2024(ix-ii).07.

Full text
Abstract:
This research study aimed to analyze the lexical density and readability of Pakistani English Newspaper and so, to suggest their suitability for the level of students. Eight newspapers from four genres were selected as sample. The theory of Systemic Functional Linguistics based on the theory of lexical density by Ure (1971) was applied for finding lexical density and readability indices by using content analysis design through an online text analyzer software. Readability values were compared with the FleschmReading Ease Scale. The results of this study showed that lexical density values of all genres were greater than 40% according to the Flesch Reading Ease Scale. All the texts were lexically dense having a greater number of contents carrying lexical items as compared to grammatical items. The findings of study could also be applied to academic textbooks in order to check complexity and suitability according to the level of the students.
APA, Harvard, Vancouver, ISO, and other styles
4

De Raad, Boele, B. F. Mulder, and Dick P. H. Barelds. "Psycho-lexically based Openness to Experience." International Journal of Personality Psychology 7 (December 17, 2021): 48–57. http://dx.doi.org/10.21827/ijpp.7.38194.

Full text
Abstract:
We investigated whether NEO-PI-R Openness to Experience (Costa & McCrae, 1992) and its six facets could be identified in the natural trait lexicon. To represent the NEO-PI-R Openness, a list of 113 items was selected from a lexically derived trait list developed for the eight-factor trait model of De Raad and Barelds (2008). We used ratings from two samples. The first (N=271) filled out the lexical Openness scales, the NEO-PI-R Openness scales, and scales measuring the eight-factor model. From the second sample (N=1,466), ratings were used to analyze the lexical Openness scales. Correlations between the eight-factor scales and the two sets of Openness scales indicated that Openness scales are fairly covered by the eight factors, except for the Ideas and Values facets of the NEO-PI-R. The lexical Openness scales correlated well with the NEO-PI-R Openness scales. Openness to Experience and its six facets were identified in the natural trait lexicon, but exploratory factor analyses did not support the six-facet structure of the NEO-PI-R Openness, neither did they lead to a similar six-facet structure across samples. Moreover, it did not consistently support a proposed two-facet structure emphasizing internal openness (fantasy, aesthetics) and external openness (ideas, change).
APA, Harvard, Vancouver, ISO, and other styles
5

Dr., Mohyeldeen Mohamed Babiker, and Ahmed Hamdan Hamdan. "Linguistic Stylistic Analysis of Lexical and Grammatical Categories in Arrow of God." International Journal of Arts, Humanities and Social Studies 4, no. 2 (2022): 95–99. https://doi.org/10.5281/zenodo.6466089.

Full text
Abstract:
The Study attempts to Shed light on linguistic Stylistics Analysis of lexical and Grammatical Categories of Chinua Achebe’s novel (Arrow of God). It aims to highlight the techniques of using lexical and Grammatical Categories in the novel. The researchers have adopted analytical checklist to explore lexical and Grammatical Categories in the functional linguistic field. These lexical and grammatical features have been used as primary tools for data collection. The sample composed of hundred words and two hundred and fifty of different sentences types that have been used in the context so as to analyze Achebe’s style. The finding revealed that studying linguistic stylistic alone cannot be satisfied to understand the novel, rather identifies significant linguistic features that mostly contributes in interpreting the novel. The study recommended that researchers should use linguistic features for   interpreting and tackling their literary works
APA, Harvard, Vancouver, ISO, and other styles
6

Tereshchuk, A. "Lexical and semantic characteristics of “Memoirs about the Last Carlist war” by M. Vayreda." Key Issues of Contemporary Linguistics, no. 3 (February 7, 2023): 153–62. https://doi.org/10.18384/2310-712x-2023-3-153-162.

Full text
Abstract:
Aim. To analyze the lexical characteristics of “Memoirs about the Last Carlist War” by the Catalan writer Marian Vayreda.Methodology. The method of contextual lexical and semantic analysis was used in the work. Dialectisms, lexical interferences from Spanish, and sociolectisms are studied in the article. The meaning and area of the analyzed lexical units were specified according to the “Large Dictionary of the Catalan language” and the “Catalan-Valencian-Balearic Dialect Dictionary”.Results. The analyzed text contains lexemes, the use of which is characteristic for the author's native North Catalan subdialect. In addition, the book contains lexical units that are absent in North Catalan, but can be found in other diatopic language varieties. The dialectisms used by the author function as stylistic devices. The article analyzes the lexical interferences from Spanish present in the text, which are divided into italicized foreign language inclusions and Spanish interferences in Catalan. The Spanish language is used by M. Vayreda to create a verbal portrait of negative characters, as well as to characterize certain concepts that the author considers alien to Catalonia and its culture. The lexical units related to the sociolect of the Catalan Carlists are considered. Some features of the political discourse of the Carlists are noted.Research implications. The study is a continuation of a number of publications dedicated to M. Vayreda's books and the first study that comprehensively analyzes the lexical features of “Memoirs of the Last Carlist War”. The article can serve as a starting point for further research on the history of the Catalan language and on the language of Spanish Carlism.
APA, Harvard, Vancouver, ISO, and other styles
7

Бобчинець, Любов Іванівна. "ЛЕКСИКО-СЕМАНТИЧНІ Й СТРУКТУРНІ ОСОБЛИВОСТІ ЛЕКСИКИ КАРТЯРСТВА У СУЧАСНІЙ ІСПАНСЬКІЙ МОВІ". Збірник наукових праць ХНПУ імені Г.С. Сковороди "Лінгвістичні дослідження", № 41 (29 березня 2016): 10–16. https://doi.org/10.5281/zenodo.48527.

Full text
Abstract:
<em>The aim of the article is to analyze lexical, semantic and structural peculiarities of card games vocabulary in the modern Spanish language, to classify the words according to lexical and semantic criteria, to describe synonymy, polysemy within the lexicon. Topicality of the investigation is conditioned by frequent use of card games idioms in modern mass media. Synonymy is caused by parallel existence of Spanish and borrowed lexical units. Polysemy does not impede the players&rsquo; understanding because the context and the game situation actualize certain meaning of the word. According to the structure the majority of the lexemes are two member word combinations, mainly nominative. Card games vocabulary in the Spanish language is an open system. The figurative character of card games vocabulary contributes to its penetration into general lexicon as expressive means. Card games idioms enrich general lexicon. Expressive characteristics of card game idioms in works of fiction and journalistic style in Spanish need further investigation.</em>
APA, Harvard, Vancouver, ISO, and other styles
8

González Sopeña, Inmaculada. "Indications of lexical-semantic changes in the Arabisms of the Kingdom of Granada (1493-1612): the lexical competition between almádena and martillo." Onomázein Revista de lingüística filología y traducción, no. 51 (2021): 83–98. http://dx.doi.org/10.7764/onomazein.51.09.

Full text
Abstract:
The objective of the present study is to analyze a couple of lexical items formed by an Arabism and a Romance voice (almádena and martillo) through a corpus of documents linked to the ancient Kingdom of Granada (i.e. the current provinces of Málaga, Almería and Granada) in the late 15th century to 17th century. That documentation includes different types of texts as the correspondence of Hernando de Zafra or texts linked to the building construction at that time. Due to the historical, political, social and religious peculiarities of the Kingdom of Granada there is a persistence of lexicon of Arab origin over the above two centuries with regard to other Spanish-speaking territories. However, Arabic loanwords in Spanish lexicon are subjected to specialization processes or semantic restrictions, as exemplified in the case of almádena. This voice lexically competed with the Romance voice martillo, but, finally, the first one suffered a process of semantic restriction and it is actually cornered in some dialectal areas.
APA, Harvard, Vancouver, ISO, and other styles
9

Makarov, N. S. "The Church lexicon as a lexical category." Linguistics & Polyglot Studies 10, no. 2 (2024): 38–49. http://dx.doi.org/10.24833/2410-2423-2024-2-39-38-49.

Full text
Abstract:
The article focuses on the study of church lexicon as a lexical category. The aim of this article is to analyze and compare different linguistic points of view on the key stages of church lexicon development in the Russian and English languages, to determine a working definition of the category “church lexicon”, as well as its most important elements. In order to achieve this goal, the methods of analyzing sources, their subsequent comparison and contrast, identification of analogies and similarities were used, as well as synthesis of information extracted from these sources.The material includes dissertations, scientific articles and other works of modern Russian scholars considering the concepts “church lexicon” and “religious lexicon”, comparing and juxtaposing them, as well as singling out constituent elements of the categories “church lexicon”, “religious lexicon”, “sacral lexicon”, and “liturgical lexicon”. In this article are used the works of such scholars as A.A. Azarov, I.V. Bugaeva, S.V. Bulavina, N.S. Valgina, R.I. Goryushina, E.V. Kakorina, I.A. Koroleva, T.I. Kosheleva and others.The results of the research show that the interest in the church lexicon as a lexical category among Russian linguists has increased since the late twentieth and early twenty-first centuries due to the revival of the Christian Church of various denominations, as well as other religious organizations. The long period of persecution of the Christian Church in Russia has affected church vocabulary, marginalizing it and relegating it to the periphery of the language. In the UK, USA and other English-speaking linguistic communities, the activities of the Christian Church have never been banned or persecuted, and the history of church vocabulary in the English language as such has never faced marginalization similar to that in Russia.It has also been established that church vocabulary is part of the broader category of “religious vocabulary” and includes basic concepts from the sphere of the life of the Christian Church. Although scholars argue about the constituents of the category “church lexicon”, many agree that it includes the names of sacraments and rites, elements of the church calendar, names of church buildings and their elements, items of decoration, elements of vestments, as well as names of persons in relation to the Christian faith and the Christian Church.This article may be of interest to philologists, linguists, and teachers of the Russian and English languages.
APA, Harvard, Vancouver, ISO, and other styles
10

Rosyadi ZA, Dede, Bima Kurniawan, Nisrinatus Zahiyah, Rosiana Sukma Anjani, and Elsa Alfida Maharani. "Lexicon Dialect Variations of Madurese in the Bangkalan and Sumenep Sociolinguistics Study." ELS Journal on Interdisciplinary Studies in Humanities 8, no. 1 (2025): 21–26. https://doi.org/10.34050/els-jish.v8i1.42006.

Full text
Abstract:
This research investigates the lexical variations within the dialects of the Madurese language, with a specific focus on the Bangkalan and Sumenep regions. The study seeks to identify and analyze the ways in which lexical items diverge between these dialects, considering how geographic factors play a role in shaping these variations. Through a descriptive qualitative approach, the study employs a combination of observation, interviews, and field notes to gather comprehensive data. The findings demonstrate that although certain lexical items exhibit differences in form across the dialects, they largely retain the same meanings. These lexical variations are primarily attributed to the geographical separation between the regions, which has led to distinct developments in the Madurese lexicon over time. The research underscores the influence of geographic isolation on the evolution of language and provides a deeper understanding of the dynamic relationship between language and region.
APA, Harvard, Vancouver, ISO, and other styles
More sources

Dissertations / Theses on the topic "Lexical analyze"

1

Morlane-Hondère, François. "Une approche linguistique de l'évaluation des ressources extraites par analyse distributionnelle automatique." Phd thesis, Université Toulouse le Mirail - Toulouse II, 2013. http://tel.archives-ouvertes.fr/tel-00937926.

Full text
Abstract:
Dans cette thèse, nous abordons du point de vue linguistique la question de l'évaluation des bases lexicales extraites par analyse distributionnelle automatique (ADA). Les méthodes d'évaluation de ces ressources qui sont actuellement mises en œuvre (comparaison à des lexiques de référence, évaluation par la tâche, test du TOEFL...) relèvent en effet d'une approche quantitative des données qui ne laisse que peu de place à l'interprétation des rapprochements générés. De ce fait, les conditions qui font que certains couples de mots sont extraits alors que d'autres ne le sont pas restent mal connues. Notre travail vise une meilleure compréhension des fonctionnements en corpus qui régissent les rapprochements distributionnels. Pour cela, nous avons dans un premier temps adopté une approche quantitative qui a consisté à comparer plusieurs ressources distributionnelles calculées sur des corpus différents à des lexiques de références (le Dictionnaire électronique des synonymes du CRISCO et le réseau lexical JeuxDeMots). Cette étape nous a permis, premièrement, d'avoir une estimation globale du contenu de nos ressources, et, deuxièmement, de sélectionner des échantillons de couples de mots à étudier d'un point de vue qualitatif. Cette deuxième étape constitue le cœur de la thèse. Nous avons choisi de nous focaliser sur les relations lexico-sémantiques que sont la synonymie, l'antonymie, l'hyperonymie et la méronymie, que nous abordons en mettant en place quatre protocoles différents. En nous appuyant sur les relations contenues dans les lexiques de référence, nous avons comparé les propriétés distributionnelles des couples de synonymes/antonymes/hyperonymes/méronymes qui ont été extraits par l'ADA avec celles des couples qui ne l'ont pas été. Nous mettons ainsi au jour plusieurs phénomènes qui favorisent ou bloquent la substituabilité des couples de mots (donc leur extraction par l'ADA). Ces phénomènes sont considérés au regard de paramètres comme la nature du corpus qui a permis de générer les bases distributionnelles étudiées (corpus encyclopédique, journalistique ou littéraire) ou les limites des lexiques de référence. Ainsi, en même temps qu'il questionne les méthodes d'évaluation des bases distributionnelles actuellement employées, ce travail de thèse illustre l'intérêt qu'il y a à considérer ces ressources comme des objets d'études linguistiques à part entière. Les bases distributionnelles sont en effet le résultat d'une mise en œuvre à grande échelle du principe de substituabilité, ce qui en fait un matériau de choix pour la description des relations lexico-sémantiques.
APA, Harvard, Vancouver, ISO, and other styles
2

Boizou, Loïc. "Analyse lexicale automatique du lituanien." Paris, INALCO, 2009. http://www.theses.fr/2009INAL0004.

Full text
Abstract:
La présente thèse a pour objet l'analyse automatique des formes lexicales dans les textes écrits en lituanien, sur la base d'une heuristique forme - valeur qui s'inscrit dans une approche symbolique du traitement automatique des langues. Cette étude accorde une attention spécifique à l'éxploitation optimale des indices formels et s'appuie principalement sur deux domaines de la linguistique, la graphématique et la morphologie. Le point d'entrée formaliste couplé à l'objectif d'automatisation a réclamé une révision de la perspective grammaticale traditionnelle, qui nous a conduit à esquisser un renouvellement de la description relative à plusieurs aspects du système linguistique, notamment les parties du discours, la structure lexicale et la suffixation. Le modèle linguistique, qui reste à développer, a servi de fondement à la réalisation d'un analyseur de formes lexicales nommé ALeksas. Ce logiciel possède une structure hybride principalement basée sur des automates à nombre fini d'états. ALeksas, qui est encore à l'état expérimental, assure l'analyse grammaticale des mots formes selon une approche indépendante d'une base de données lexicale permettant de formuler des hypothèses d'interprétation sur des critères formels. Le prototype a fait l'objet d'une mise à l'épreuve par confrontation à un corpus de textes authentiques variés, afin d'évaluer ses capacités, notamment par rapport aux outils comparables, et de mieux cerner les améliorations nécessaires<br>The aim of this thesis is to carry out lexical analysis of written texts in Lithuanian by automatic means, according to a heuristics from form to content based on symbolic methods. This study attempts to make an expanded use of marks given by linguistic forms, drawing on graphemic and morphological aspects. This formal starting point in conjunction with automation of linguistic tasks required a revision of the traditional grammatical point of view, concerning mainly parts of speech, lexical structure and suffixation. This linguistic model, which needs further expansion, served as a basis for ALeksas, an analyzer of lexical forms. This software implements a hybrid structure expanding a system of finite state automata. The prototype computes the analysis of word forms, giving grammatical interpretations according to a set of formal criteria, instead of making use of a lexical database. The results of the analysis of a corpus complied from various texts allowed us to delineate more precisely the advantages and shortcomings of Aleksas, as compared with other similar tools, and to also suggest possible enhancements
APA, Harvard, Vancouver, ISO, and other styles
3

Adam, Clémentine. "Voisinage lexical pour l'analyse du discours." Phd thesis, Université Toulouse le Mirail - Toulouse II, 2012. http://tel.archives-ouvertes.fr/tel-00784531.

Full text
Abstract:
Cette thèse s'intéresse au rôle de la cohésion lexicale dans différentes approches de l'analyse du discours. Nous yexplorons deux hypothèses principales:- l'analyse distributionnelle, qui permet de rapprocher des unités lexicales sur la base des contextes syntaxiques qu'ellespartagent, met au jour des relations sémantiques variées pouvant être exploitées pour la détection de la cohésion lexicaledes textes;- les indices lexicaux constituent des éléments de signalisation de l'organisation du discours pouvant être exploités aussibien à un niveau local (identification de relations rhétoriques entre constituants élémentaires du discours) qu'à un niveauglobal (repérage ou caractérisation de segments de niveau supérieur dotés d'une fonction rhétorique et garantissant lacohérence et la lisibilité du texte, par exemple passages à unité thématique).Concernant le premier point, nous montrons la pertinence d'une ressource distributionnelle pour l'appréhension d'une largegamme de relations impliquées dans la cohésion lexicale des textes. Nous présentons les méthodes de projection et defiltrage que nous avons mises en œuvre pour la production de sorties exploitables.Concernant le second point, nous fournissons une série d'éclairages qui montrent l'apport d'une prise en compte réfléchiede la cohésion lexicale pour une grande variété de problématiques liées à l'étude et au repérage automatique del'organisation textuelle: segmentation thématique de textes, caractérisation des structures énumératives, étude de lacorrélation entre lexique et structure rhétorique du discours et enfin détection de réalisations d'une relation de discoursparticulière, la relation d'élaboration.
APA, Harvard, Vancouver, ISO, and other styles
4

Audibert, Laurent. "Outil d'exploration de corpus et désambiguïté lexicale automatique." Aix-Marseille 1, 2003. https://tel.archives-ouvertes.fr/tel-00004475.

Full text
Abstract:
Ce travail de thèse adresse le problème de la désambigui͏̈sation lexicale automatique à l'aide de méthodes d'apprentissage supervisé. Dans une première partie, nous proposons un ensemble de puissants outils de manipulation de corpus linguistiques étiquetés. Pour réaliser ces outils, nous avons développé une bibliothèque C++ qui implémente un langage élaboré et expressif d'interrogation de corpus, basé sur des méta-expressions régulières. Dans une seconde partie, nous comparons divers algorithmes d'apprentissage supervisé, que nous utilisons ensuite pour mener à bien une étude systématique et approfondie de différents critères de désambigui͏̈sation, basés sur la cooccurrence de mots et plus généralement de n-grammes. Nos résultats vont parfois à l'encontre de certaines pratiques dans le domaine. Par exemple, nous montrons que la suppression des mots grammaticaux dégrade les performances et que les bigrammes permettent d'obtenir de meilleurs résultats que les unigrammes.
APA, Harvard, Vancouver, ISO, and other styles
5

Henestroza, Anguiano Enrique. "Efficient large context dependency parsing and correction with distributional lexical resources." Paris 7, 2013. http://www.theses.fr/2013PA077066.

Full text
Abstract:
Cette thèse présente des méthodes pour améliorer l'analyse syntaxique probabiliste en dépendances. Nous employons l'analyse à base de transitions avec une modélisation effectuée par des machines à vecteurs supports (Cortes and Vapnik, 1995), et nos expériences sont réalisées sur le français. L'analyse a base de transitions est rapide, de par la faible complexité des algorithmes sous-jacents, eux mêmes fondés sur une optimisation locale des décisions d'attachement. Ainsi notre premier fil directeur est d'élargir le contexte syntaxique utilisé. Partant du système de transitions arc-eager (Nivre, 2008), nous proposons une variante qui considère simultanément plusieurs gouverneurs candidats pour les attachements à droite. Nous testons aussi la correction des analyses, inspirée par Hall and Novâk (2005), qui révise chaque attachement en choisissant parmi plusieurs gouverneurs alternatifs dans le voisinage syntaxique. Nos approches i améliorent légèrement la précision globale ainsi que celles de l'attachement des groupes prépositionnels et de la coordination. Notre deuxième fil explore des approches semi-supervisées. Nous testons l'auto-entrainement avec un analyseur en deux étapes, basé sur McClosky et al. (2006), pour le domaine journalistique ainsi que pour l'adaptation au domaine médical. Nous passons ensuite à la modélisation lexicale à base de corpus, avec des classes lexicales généralisées pour réduire la dispersion des données, et des préférences lexicales de l'attachement des groupes prépositionnels pour aider à la désambiguïsation. Nos approches améliorent, dans certains cas, la précision et la couverture de l'analyseur, sans augmenter sa complexité théorique<br>This thesis explores ways to improve the accuracy and coverage of efficient statistical dependency parsing. We employ transition-based parsing with models learned using Support Vector Machines (Cortes and Vapnik, 1995), and our experiments are carried out on French. Transition-based parsing is very fast due to the computational efficiency of its underlying algorithms, which are based on a local optimization of attachment décisions. Our first research thretd is thus to increase thé syntactic context used. From the arc-eager transition System (Nivre, 2008) we propose a variant that simultaneously considers multiple candidate governors for right-directed attachments. We also test parse correction, inspired by Hall and Novâk (2005), which revises each attachment in a parse by considering multiple alternative governors in the local syntactic neighborhood. We find that multiple-candidate approaches slightly improve parsing accuracy overall as well as for prepositional phrase attachment and coordination, two linguistic phenomena that exhibit high syntactic ambiguity. Our second research thread explores semi-supervised approaches for improving parsing accuracy and coverage. We test self-training within the journalistic domain as well as for adaptation to the medical domain, using a two-stage parsing approach based on that of McClosky et al. (2006). We then turn to lexical modeling over a large corpus: we model generalized lexical classes to reduce data sparseness, and prepositional phrase attachment préférence to improve disambiguation. We find that semi-supervised approaches can sometimes improve parsing accuracy and coverage, without increasing time complexity
APA, Harvard, Vancouver, ISO, and other styles
6

Sagot, Marie-France. "Ressemblance lexicale et structurale entre macromolécules : formalisation et approches combinatoires." Marne-la-Vallée, 1996. http://www.theses.fr/1996MARN0049.

Full text
Abstract:
Ce travail présente une formalisation du problème de la ressemblance entre macromolécules biologiques, ADN/ARN et protéines. Une première partie est ainsi composée par une exploration de cette ressemblance. D'une manière plus précise, ces macromolécules présentent la caractéristique commune d'être constituées par l'assemblage linéaire d'unités monomères distinctes (nucléotides pour l'ADN/ARN ou acides aminés pour les protéines) pouvant être représenté par une chaîne de symboles. Ces chaînes que nous allons vouloir comparer peuvent être vues comme des 'copies', ou des copies de copies différentes d'une même macromolécule ou d'une même forme ancestrale, chaque copie étant le résultat d'une suite particulière de transformations et de mutations effectuées sur des copies antérieures. Certaines parties de ces copies resteront cependant plus conservées que d'autres, et il est probable qu'il s'agit justement là de celles qui ont le plus de chance d'être associées, soit à une activité de la chaîne polymérique, soit à un élément structural qui sert de charpente à la structure générale de la macromolécule dans l'espace. Ce sont essentiellement ces parties qui nous intéressent dans ce travail et notre analyse des chaînes porte sur l'identification des mots présents dans ces chaînes et similaires entre eux d'une certaine façon. Outre l'idée d'une comparaison locale (mots), deux notions sont fondamentales dans les définitions de ressemblance que nous établissons : l'une est celle d'une comparaison multiple, l'autre celle de modèle. L'idée de la première repose sur l'observation que la comparaison simultanée d'un grand nombre d'objets permet d'être beaucoup plus sensible, c'est-à-dire, de détecter de plus faibles similarités entre ces objets. Le second concept (modèle) nous permet alors de réaliser cette comparaison simultanée de manière efficace. Un tel objet, qui est externe aux chaînes que l'on compare, peut être soit un mot sur le même alphabet, soit un produit cartésien de sous-ensembles des symboles de cet alphabet. La définition de ressemblance 'multiple' entre les mots des chaînes se trouve donc ramenée à cet objet de référence. Plus précisément, nous dirons que les mots d'un ensemble sont similaires entre eux s'il existe au moins un modèle auquel tous ressemblent. Notre recherche de mots similaires communs se résume alors à une recherche efficace de ces modèles présents, c'est-à-dire ayant des occurrences, dans au moins un certain nombre des chaînes de l'ensemble étudié. Finalement, le problème de la définition de la ressemblance entre mots se ramène à la définition de la ressemblance entre des mots et un modèle. Dans ce travail, nous proposons plusieurs de ces définitions avec le souci constant de leur conserver un caractère mathématiquement précis. Cette condition est en effet importante afin de donner au biologiste les moyens d'interpréter sans ambiguité les résultats fournis. Dans le même ordre d'idée, il est également important que les algorithmes de recherche des modèles soient exhaustifs. La seconde partie de ce travail a ainsi porté sur l'élaboration d'algorithmes combinatoires qui explorent tout l'espace des chaînes que l'on compare et qui fournissent en résultat une liste de tous les modèles, d'une longueur donnée ou de longueur maximale, présents, selon une des définitions de ressemblance entre modèle et mots, dans au moins un certain nombre des chaînes. Chacune de ces définitions a donné lieu à un algorithme spécifique. Le principe de base de ces algorithmes est cependant le même et s'appuie sur une formule de récurrence permettant la construction des modèles d'une certaine longueur à partir de ceux de longueur plus petite. Nous proposons ici des algorithmes qui sont tous linéaires dans la longueur totale des chaînes, et éventuellement exponentiels uniquement dans le degré de souplesse autorisé dans la ressemblance entre mots et modèles. Ce facteur exponentiel est intrinsèque au problème et représente toujours une situation de pire cas. Enfin, l'intérêt et les limitations actuelles de ces algorithmes sont illustrés sur des exemples biologiques
APA, Harvard, Vancouver, ISO, and other styles
7

Boissard, Aubrée. "Voix passive et passif lexical en innu : une analyse aspectuelle et sémantique." Master's thesis, Université Laval, 2017. http://hdl.handle.net/20.500.11794/27940.

Full text
Abstract:
Tableau d'honneur de la Faculté des études supérieures et postdoctorales, 2017-2018<br>L’innu est une langue algonquienne parlée au Québec et au Labrador par environ 10 000 locuteurs. Ce mémoire examine l’aspectualité de la voix passive ainsi que le phénomène du passif lexical en innu, un sous-type de passif peu documenté au sein des langues algonquiennes. L’étude de cas menée dans ce but révèle que certaines formes verbales passives en innu peuvent osciller entre une sémantique dynamique et stative/résultative. Ce mémoire propose d’envisager les formes statives/résultatives comme des formes passives lexicalisées dans la langue. L’étude met aussi en lumière des formes verbales apparentées à la voix passive et à la voix moyenne également présentes dans d’autres langues algonquiennes qui sont traitées comme des passifs lexicaux en tant que tels dans le présent travail. Les annexes fournissent les données de l’étude de cas menée dans le cadre de ce mémoire ainsi qu’une sélection de phrases passives tirées de l’ouvrage Eukuan nin matshimanitu innu-iskueu (Kapesh, 1976). Les annexes fournissent également une sélection de formes verbales passives transitives ainsi qu’une sélection des formes verbales traitées comme des passifs lexicaux. Ces deux sélections sont extraites à partir du dictionnaire innu en ligne.<br>Innu is an Algonquian language spoken in Québec and Labrador by around 10 000 speakers. This thesis investigates the aspectual properties of the passive voice as well as the phenomenon of lexical passive, a subtype of passive that remains poorly documented in Algonquian languages. The case study carried out with this aim reveals that certain passive verbal forms in Innu can have either a dynamic or a stative/resultative semantics. This thesis explores forward the idea of considering stative/resultative forms as lexicalized passive forms in the language. Moreover, the study highlights a number of verbal forms related to passive and middle voice also found in other Algonquian languages. These forms are considered as lexical passives in this thesis. Appendices show the data of the case study study as well as a selection of passive sentences from the book Eukuan nin matshimanitu innu-iskueu (Kapesh, 1976). Appendices show also a selection of transitive passive verbal forms and a selection of verbal forms analyzed as lexical passives in this study. Those selection are all extracted from the Innu online dictionnary.
APA, Harvard, Vancouver, ISO, and other styles
8

Conrath, Juliette. "Unsupervised extraction of semantic relations using discourse information." Thesis, Toulouse 3, 2015. http://www.theses.fr/2015TOU30202/document.

Full text
Abstract:
La compréhension du langage naturel repose souvent sur des raisonnements de sens commun, pour lesquels la connaissance de relations sémantiques, en particulier entre prédicats verbaux, peut être nécessaire. Cette thèse porte sur la problématique de l'utilisation d'une méthode distributionnelle pour extraire automatiquement les informations sémantiques nécessaires à ces inférences de sens commun. Des associations typiques entre des paires de prédicats et un ensemble de relations sémantiques (causales, temporelles, de similarité, d'opposition, partie/tout) sont extraites de grands corpus, par l'exploitation de la présence de connecteurs du discours signalant typiquement ces relations. Afin d'apprécier ces associations, nous proposons plusieurs mesures de signifiance inspirées de la littérature ainsi qu'une mesure novatrice conçue spécifiquement pour évaluer la force du lien entre les deux prédicats et la relation. La pertinence de ces mesures est évaluée par le calcul de leur corrélation avec des jugements humains, obtenus par l'annotation d'un échantillon de paires de verbes en contexte discursif. L'application de cette méthodologie sur des corpus de langue française et anglaise permet la construction d'une ressource disponible librement, Lecsie (Linked Events Collection for Semantic Information Extraction). Celle-ci est constituée de triplets: des paires de prédicats associés à une relation; à chaque triplet correspondent des scores de signifiance obtenus par nos mesures.Cette ressource permet de dériver des représentations vectorielles de paires de prédicats qui peuvent être utilisées comme traits lexico-sémantiques pour la construction de modèles pour des applications externes. Nous évaluons le potentiel de ces représentations pour plusieurs applications. Concernant l'analyse du discours, les tâches de la prédiction d'attachement entre unités du discours, ainsi que la prédiction des relations discursives spécifiques les reliant, sont explorées. En utilisant uniquement les traits provenant de notre ressource, nous obtenons des améliorations significatives pour les deux tâches, par rapport à plusieurs bases de référence, notamment des modèles utilisant d'autres types de représentations lexico-sémantiques. Nous proposons également de définir des ensembles optimaux de connecteurs mieux adaptés à des applications sur de grands corpus, en opérant une réduction de dimension dans l'espace des connecteurs, au lieu d'utiliser des groupes de connecteurs composés manuellement et correspondant à des relations prédéfinies. Une autre application prometteuse explorée dans cette thèse concerne les relations entre cadres sémantiques (semantic frames, e.g. FrameNet): la ressource peut être utilisée pour enrichir cette structure par des relations potentielles entre frames verbaux à partir des associations entre leurs verbes. Ces applications diverses démontrent les contributions prometteuses amenées par notre approche permettant l'extraction non supervisée de relations sémantiques<br>Natural language understanding often relies on common-sense reasoning, for which knowledge about semantic relations, especially between verbal predicates, may be required. This thesis addresses the challenge of using a distibutional method to automatically extract the necessary semantic information for common-sense inference. Typical associations between pairs of predicates and a targeted set of semantic relations (causal, temporal, similarity, opposition, part/whole) are extracted from large corpora, by exploiting the presence of discourse connectives which typically signal these semantic relations. In order to appraise these associations, we provide several significance measures inspired from the literature as well as a novel measure specifically designed to evaluate the strength of the link between the two predicates and the relation. The relevance of these measures is evaluated by computing their correlations with human judgments, based on a sample of verb pairs annotated in context. The application of this methodology to French and English corpora leads to the construction of a freely available resource, Lecsie (Linked Events Collection for Semantic Information Extraction), which consists of triples: pairs of event predicates associated with a relation; each triple is assigned significance scores based on our measures. From this resource, vector-based representations of pairs of predicates can be induced and used as lexical semantic features to build models for external applications. We assess the potential of these representations for several applications. Regarding discourse analysis, the tasks of predicting attachment of discourse units, as well as predicting the specific discourse relation linking them, are investigated. Using only features from our resource, we obtain significant improvements for both tasks in comparison to several baselines, including ones using other representations of the pairs of predicates. We also propose to define optimal sets of connectives better suited for large corpus applications by performing a dimension reduction in the space of the connectives, instead of using manually composed groups of connectives corresponding to predefined relations. Another promising application pursued in this thesis concerns relations between semantic frames (e.g. FrameNet): the resource can be used to enrich this sparse structure by providing candidate relations between verbal frames, based on associations between their verbs. These diverse applications aim to demonstrate the promising contributions provided by our approach, namely allowing the unsupervised extraction of typed semantic relations
APA, Harvard, Vancouver, ISO, and other styles
9

Miniconi, Roger. "La langue de la mer en Corse : Patrimoine lexical des pêcheurs de l'Ile." Corte, 1996. http://www.theses.fr/1996CORT2003.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Schwab, Didier. "Approche hybride - lexicale et thématique - pour la modélisation, la détection et l'exploitation des fonctions lexicales en vue de l'analyse sémantique de texte." Phd thesis, Université Montpellier II - Sciences et Techniques du Languedoc, 2005. http://tel.archives-ouvertes.fr/tel-00333334.

Full text
Abstract:
Utilisée à la fois pour l'apprentissage et l'exploitation des vecteurs conceptuels, l'analyse sémantique de texte est centrale à nos recherches. L'amélioration qualitative du processus d'analyse entraîne celle des vecteurs. En retour, cette meilleure pertinence a un effet positif sur l'analyse. Parmi les différentes voies à explorer pour obtenir ce cercle vertueux, l'une des pistes les plus intéressantes semble être la découverte puis l'exploitation des relations lexicales entre les mots du texte. Ces relations, parmi lesquelles la synonymie, l'antonymie, l'hyperonymie, la bonification ou l'intensification, sont modélisables sous la forme de fonctions lexicales. Énoncées essentiellement dans un cadre de production par Igor Mel'čuk, nous cherchons, dans cette thèse, à les adapter à un cadre d'analyse. Nous introduisons ici deux classes de Fonctions Lexicales d'Analyse. Les premières, les FLA de construction permettent de fabriquer un vecteur conceptuel à partir des informations lexicales disponibles. Les secondes, les FLA d'évaluation permettent de mesurer la pertinence d'une relation lexicale entre plusieurs termes. Ces dernières sont modélisables grâce à des informations thématiques (vecteurs conceptuels) et/ou grâce à des informations lexicales (relations symboliques entre les objets lexicaux).<br /><br />Les informations lexicales sont issues de la base lexicale sémantique dont nous introduisons l'architecture à trois niveaux d'objets lexicaux (item lexical, acception, lexie). Elles sont matérialisées sous la forme de Relations Lexicales Valuées qui traduisent la probabilité d'existence de la relation entre les objets. L'utilité de ces relations a pu être mis en évidence pour l'analyse sémantique grâce à l'utilisation du paradigme des algorithmes à fourmis. Le modèle introduit dans cette thèse, utilise à la fois les vecteurs conceptuels et les relations du réseau lexical pour résoudre une partie des problèmes posés lors d'une analyse sémantique.<br /><br />Tous nos outils ont été implémentés en Java. Ils reposent sur Blexisma (Base LEXIcale Sémantique Multi-Agent) une architecture multi-agent élaborée au cours de cette thèse dont l'objectif est d'intégrer tout élément lui permettant de créer, d'améliorer et d'exploiter une ou plusieurs Bases Lexicales Sémantiques. Les expériences menées ont montré la faisabilité de cette approche, sa pertinence en termes d'amélioration globale de l'analyse et ouvert des perspectives de recherches fort intéressantes.
APA, Harvard, Vancouver, ISO, and other styles
More sources

Books on the topic "Lexical analyze"

1

Université de Paris IV: Paris-Sorbonne, ed. Analyse contrastive des systèmes lexicaux du Français et du Portugais. A.N.R.T, Université de Lille III, 1987.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
2

Zanola, Maria Teresa. L' emprunt lexical anglais dans le français contemporain: Analyse d'un corpus de press (1982-1989). Scuola, 1991.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
3

Qi, Chong. Néologie et terminologie: Analyse et classification des emprunts lexicaux en chinois. A.N.R.T, Université de Lille III, 1997.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
4

III, Université de Paris, ed. Transitivité et diathèse en tarifit: Analyse de quelques relations de dépendances lexicale et syntaxique. A.N.R.T, Université de Lille III, 1990.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
5

Université de Paris IV: Paris-Sorbonne, ed. Le français régional de Provence, étude phonétique, phonologique, lexicale, syntaxique: Analyse du substrat provençal. A.N.R.T, Université de Lille III, 1986.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
6

Institut royal de la culture amazighe. Centre de l'aménagement linguistique, ed. Transitivité et diathèse en tarifite: Analyse de quelques relations de dépendances lexicale et syntaxique. Publications de l'Institut royal de la culture amazighe, 2005.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
7

Bonucci, Alexandre. Analyse phonologique et indexation figurative pour une base de données d'entrées lexicales de la LSF. A.N.R.T, Université de Lille III, 1998.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
8

Delmonte, Rodolfo. Computational linguistic text processing: Lexicon, grammar, parsing and anaphora resolution. Nova Science Publishers, 2008.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
9

Descartes, Université René, ed. La dynamique d'un parler poitevin dans le sud des deux-sèvres: Étude sociolinguistique : description phonologique et lexicale : analyse des tendances évolutives et des interférences avec le français. A.N.R.T, Université de Lille III, 1994.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
10

Behnke, Dorothea. " dass dem weiblichen Geschlechte an Tapfferkeit, Klugheit, Gelehrsamkeit und andern Haupt-Tugenden gar nichts fehle": Lexika zu Schriftstellerinnen aus dem deutschsprachigen Raum : Bestandsaufnahme und Analyse. Zeller, 1999.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
More sources

Book chapters on the topic "Lexical analyze"

1

Rodriguez-Cuadrado, Sara, and Carlos Romero Rivas. "Chapter 8. Effects of foreign-accented speech on language comprehension processes and social cognition." In Bilingual Processing and Acquisition. John Benjamins Publishing Company, 2023. http://dx.doi.org/10.1075/bpa.17.08rod.

Full text
Abstract:
In this chapter, we review the literature on how foreign-accented speech modulates cognitive processes related to language comprehension and describe the interactions between native and non-native speakers. First, we focus on how language comprehension processes are modulated when native listeners listen to foreign-accented speakers. Whilst the results about acoustic-phonetic and syntactic processing are relatively consistent throughout the literature on the subject, there is currently a debate on what the effects of foreign-accented speech are on lexical-semantic processing. Secondly, we analyze the literature on how foreign listeners process foreign-accented speech, a field of research that has grown recently. Then, we discuss how native and non-native speakers align with one another during verbal interactions. Finally, we examine the cognitive biases that native listeners show when interacting with non-native speakers that modulate social cognition processes.
APA, Harvard, Vancouver, ISO, and other styles
2

de Almeida, Roberto G., and Caitlyn Antal. "How Can Semantics Avoid the Troubles with the Analytic/Synthetic Distinction?" In Language, Cognition, and Mind. Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-50200-3_5.

Full text
Abstract:
AbstractAt least since Quine (From a logical point of view. Harvard University Press, Cambridge, MA, 1953) it has been suspected that a semantic theory that rests on defining features, or on what are taken to be “analytic” properties bearing on the content of lexical items, rests on a fault line. Simply put, there is no criterion for determining which features or propertiesFeatures are to be analytic and which ones are to be synthetic or contingent on experience. Deep down, our concern is what cognitive science and its several competing semantic theories have to offer in terms of solution. We analyze a few cases, which run into trouble by appealing to analyticity, and propose our own solution to this problem: a version of atomism cum inferences, which we think it is the only way out of the dead-end of analyticity. We start off by discussing several guiding assumptions regarding cognitive architecture and on what we take to be methodological imperatives for doing semantics within cognitive science—that is a semantics that is concerned with accounting for mental states. We then discuss theoretical perspectives on lexical causatives and the so-called “coercion” phenomenon or, in our preferred terminology, indeterminacy. And we advance, even if briefly, a proposal for the representation and processing of conceptual content that does away with the analytic/synthetic distinction. We argue that the only account of mental content that does away with the analytic/synthetic distinction is atomism. The version of atomism that we sketch accounts for the purported effects of analyticity with a system of inferences that are in essence synthetic and, thus, not content constitutive.
APA, Harvard, Vancouver, ISO, and other styles
3

Sugimura, Ryôichi, Kôji Akasaka, Yukihiro Kubo, and Yûji Matsumoto. "Logic based lexical analyser LAX." In Logic Programming '88. Springer Berlin Heidelberg, 1989. http://dx.doi.org/10.1007/3-540-51564-x_64.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Villarini, Andrea. "Lo sviluppo della competenza lessicale: incursioni tra le ipotesi teoriche e le applicazioni glottodidattiche." In Competenza lessicale e apprendimento dell’Italiano L2. Firenze University Press, 2021. http://dx.doi.org/10.36253/978-88-5518-403-8.05.

Full text
Abstract:
The paper analyses the issue of the development of lexical competence in Italian as L2 learners. The notion of lexicon is considered from three different points of view: the point of view of the speaker/learner, the point of view of the researcher interested in language issues and the point of view of the foreign language teacher. The question of what to intend and how to manage the lexicon within a second language courses is an apparently trivial one that might be solved with a simple: "teaching Italian words"; while, it is one of the most rich of theoretical implications and methodological choices. This contribution aims at investigating the complex relationship between teaching language and lexicon, providing also a small list of instructions on how to teach the words of a new language for teachers of Italian language as L2.
APA, Harvard, Vancouver, ISO, and other styles
5

Aszyk-Treppa, Karolina, and Zbigniew Treppa. "10. Comparative Analyses." In Sindonological lexicon. Vandenhoeck & Ruprecht, 2025. https://doi.org/10.13109/9783666573606.116.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Vlasov, Sergej V., and Leonid V. Moskovkin. "Лексические омонимы и квазиомонимы (équivoques) в Грамматике французской и русской 1730 года." In Biblioteca di Studi Slavistici. Firenze University Press, 2024. https://doi.org/10.36253/979-12-215-0585-6.13.

Full text
Abstract:
Lexical Homonyms and Quasi-Homonyms (“équivoques”) in the Grammar of French and Russian of 1730. The article analyzes Russian homonyms and quasi-homonyms with a French translation from the section “Équivoques” of Grammaire Françoise et Russe of 1730. The history of the French term “équivoques” and the related terms “homonym”, “homophone”, “homograph” and “quasi-homonym” are considered. It is concluded that among the words in the section “Équivoques” there are both homonyms and quasi-homonyms. Correspondences of several words to similar words from the grammars of J. W. Paus and M. Schwanwitz were identified. Errors in the spelling of Russian words and their translation into French are analyzed. It is concluded that the author described the lexical difficulties of the Russian language from the point of view of a native German speaker.
APA, Harvard, Vancouver, ISO, and other styles
7

Nilges, Edward G. "The Lexical Analyzer for the QuickBasic Compiler." In Build Your Own .NET Language and Compiler. Apress, 2004. http://dx.doi.org/10.1007/978-1-4302-0698-9_5.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Sharma, Shikhar, Piyush Kumar, and Krishan Kumar. "LEXER: LEXicon Based Emotion AnalyzeR." In Lecture Notes in Computer Science. Springer International Publishing, 2017. http://dx.doi.org/10.1007/978-3-319-69900-4_47.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Kucewicz, Wojciech, and Jakub S. Bechcicki. "2. Physical Analyses of the Shroud." In Sindonological lexicon. Vandenhoeck & Ruprecht, 2025. https://doi.org/10.13109/9783666573606.88.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Jaworski, Jan S. "7. Chemical Analyses of the Shroud." In Sindonological lexicon. Vandenhoeck & Ruprecht, 2025. https://doi.org/10.13109/9783666573606.103.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Lexical analyze"

1

S P, Prahalya, Sankepalli Sai Samhitha, Vadapalli Sanjana Ratan, and Meena Belwal. "Malicious Query Detection Using Lexical Analyzer and Machine Learning." In 2024 Third International Conference on Electrical, Electronics, Information and Communication Technologies (ICEEICT). IEEE, 2024. http://dx.doi.org/10.1109/iceeict61591.2024.10718448.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Muthukumar, Sai Smrithi, Bhavika Gandham, Trisha Vijayekkumaran, Nagasarapu Sarayu Krishna, and Niharika Panda. "Implementing a Multilingual Lexical Analyser for Java, C and C++." In 2024 5th IEEE Global Conference for Advancement in Technology (GCAT). IEEE, 2024. https://doi.org/10.1109/gcat62922.2024.10923959.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Williams, Christopher. "The Interplay Between Syntactic and Lexical Complexity." In Výzkum v didaktice cizích jazyků. Masaryk University Press, 2023. http://dx.doi.org/10.5817/cz.muni.p280-0534-2023-11.

Full text
Abstract:
This article reports on research into the relationship between syntactic and lexical complexity, the aim being to discover whether the presence of one measure correlates with the presence of another. 100 argumentative essays written by L1-Czech L2-English students were analysed using the LSA and L2SCA. The results indicate that use of complex nominals and clause length, both features of syntactic complexity, have a strong positive correlation with lexical complexity measures.
APA, Harvard, Vancouver, ISO, and other styles
4

Hoque, Mohammed Moshiul, Mohammad Jahedur Rahman, and Pranab KumarDhar. "Lexical Semantics: A Newapproach to Analyze the Bangla Sentence with Semantic Features." In 2007 International Conference on Information and Communication Technology. IEEE, 2007. http://dx.doi.org/10.1109/icict.2007.375348.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Dankova, Natalia. "Lexical Richness of French and Quebec Journalistic Texts." In Computational Linguistics in Bulgaria. Institute for Bulgarian Language, 2024. https://doi.org/10.47810/clib.24.20.

Full text
Abstract:
This paper presents some results of a quantitative study that focuses on the variety and word frequency in texts from a comparative perspective. The study aims to analyze and compare French and Quebec journalistic texts on political and cultural topics written in French and recently published in major newspapers such as Le Monde, le Figaro, Le Devoir, etc. The statistical analysis concerns the number of different words in the text, the number of different adjectives, the number of different verbs (and also passive structures, participles and gerunds which contribute to syntactic and stylistic sophistication), and the number of hapaxes. French texts from France exhibit greater lexical richness and sophistication: they contain more adjectives, a greater variety of adjectives, as well as more participles and gerunds compared to French texts from Quebec. The originality of the study lies in the fact that it analyzes variation in French using a lexicometric approach.
APA, Harvard, Vancouver, ISO, and other styles
6

Rutecki, Krzysztof, and Ryszard Makowski. "Lexical model for ASR of fluent Polish speech - analyze of polish phonology and phonetics." In 2008 International Conference on Signals and Electronic Systems. IEEE, 2008. http://dx.doi.org/10.1109/icses.2008.4673479.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Dumitran, Angela. "TRANSLATION ERROR IN GOOGLE TRANSLATE FROM ENGLISH INTO ROMANIAN IN TEXTS RELATED TO CORONAVIRUS." In eLSE 2021. ADL Romania, 2021. http://dx.doi.org/10.12753/2066-026x-21-078.

Full text
Abstract:
Both the emergence of the pandemic and lack of knowledge and/or time needed to translate texts related to this topic brought about an increased interest in analyzing Neural Machine Translation (NMT) performance. This study aims to identify and analyze lexical and semantic errors of language aspects that appear in medical texts translated by Google Translate from English into Romanian. The data used for investigation comprises official prospects of 5 vaccines that were approved to be used against the current coronavirus. The focus is on the lexical and semantic errors, as researchers state that these errors made by Machine Translation have the highest frequency compared to morphological or syntactic errors. Moreover, the lexical errors may affect the meaning, the message, and may easily lead to mistranslation, misunderstanding and, therefore, misinformation. The texts to be analyzed are collected from official websites and translated using Google Translate and Google Languages Tools. From the data analyzed, there are 22 lexical and semantic errors that are approached through descriptive methodology. By examining types of errors in translation from English into Romanian and analyzing the potential causes of errors, the results will be used to illustrate the quality and accuracy of Google Translate when translating public health information from English into Romanian, to observe how much the message is affected by the error, in order to sharpen up linguistic awareness. The results of the study can ultimately help improve of the quality of NMT in terms of better lexical selection and attempt to give inputs as a contribution for a more adequate translation into Romanian by Google Machine Translation.
APA, Harvard, Vancouver, ISO, and other styles
8

Gutierrez, Fernando, Christian Soto, Bernardo Riffo, et al. "Literador: A Comprehensive Tutoring System for Spanish Writing." In Human Interaction and Emerging Technologies (IHIET-AI 2022) Artificial Intelligence and Future Applications. AHFE International, 2022. http://dx.doi.org/10.54941/ahfe100874.

Full text
Abstract:
In a professional setting and in adult education, a well-written text needs to convey meaning by presenting ideas in a coherent and cohesive form that facilitates readability, such as the balance in the use of coreferences, the abstraction of the language used, and the lexical diversity of the text. In this work, we proposed Literador, an Intelligent Tutoring System for Spanish writing. By incorporating different Natural language Processing tools, Literador can analyze and provide feedback on the different aspects of a text, beyond just content. These tools are the Spanish-trained language model BETO, the text complexity analyzer Trunajod, and regular expressions to identify the use of key lexical elements. By combining all of these sources of information, Literador can provide feedback following a strategy that prompts different types of messages depending on the student’s level and tries, which also intends to avoid information overflow.
APA, Harvard, Vancouver, ISO, and other styles
9

Russo, Maxim. "Representation of lexical polysemy in the database (semantic shift 'sun/day')." In INTERNATIONAL CONFERENCE on Computational Linguistics and Intellectual Technologies. RSUH, 2023. http://dx.doi.org/10.28995/2075-7182-2023-22-1108-1111.

Full text
Abstract:
This paper, based on the data from more than 600 languages collected during the work on the database of semantic shifts in the languages of the world, adresses the semantic transition 'sun'/'day'. We analyze the geographic and genealogical distribution of this semantic shift, the predominant direction of semantic development, and the patterns of morphological derivation associated with the shift.
APA, Harvard, Vancouver, ISO, and other styles
10

Kelious, Abdelhak, Mathieu Constant, and Christophe Coeur. "Investigating strategies for lexical complexity prediction in a multilingual setting using generative language models and supervised approaches." In 13th Workshop on Natural Language Processing for Computer Assisted Language Learning. Linköping University Electronic Press, 2024. http://dx.doi.org/10.3384/ecp211008.

Full text
Abstract:
This paper explores methods to automatically predict lexical complexity in a multilingual setting using advanced natural language processing models. More precisely, it investigates the use of transfer learning and data augmentation techniques in the context of supervised learning, showing the great interest of multilingual approaches. We also assess the potential of generative large language models for predicting lexical complexity. Through different prompting strategies (zero-shot, one-shot, and chain-of-thought prompts), we analyze model performance in diverse languages. Our findings reveal that while generative models achieve high correlation scores, their predictive quality varies. The comparative study illustrates that while generative large language models have potential, optimized task-specific models still outperform them in accuracy and reliability.
APA, Harvard, Vancouver, ISO, and other styles

Reports on the topic "Lexical analyze"

1

Guo, Libo, and Ruihua Zhang. Building a Singapore learner corpus of English writing for pedagogy. National Institute of Education, Nanyang Technological University, Singapore, 2020. https://doi.org/10.32658/10497/28997.

Full text
Abstract:
This project seeks to (1) build a learner corpus of English writing by collecting sample language data at Primary 6, Secondary 4, and Junior College 2 levels; (2) generate linguistic profiles for these stages and analyze/compare lexical, grammatical and discoursal features across these stages to track the development of English writing; (3) inform the learning, teaching and curriculum design of English writing; and (4) contribute to the theorization on the nature of writing development (Christie &amp; Derewianka, 2008). It attempts to answer the following research questions: 1. What are the lexical, grammatical, stylistic, and discoursal features of the English writing produced by students at Primary 6, Secondary 4, and Junior College 2? 2. Are there any differences in the language produced by students from different developmental stages? If yes, what are the differences and what is the nature of such differences? 3. Within each developmental stage, are there any differences in terms of lexical, grammatical, and discoursal features in the writing from students of different gender and ethnic groups? If yes, how are they different? 4. To what extent do these features reflect students’ development of English proficiency in general and their development of writing skills in particular? 5. Are there any differences in English writing processes, perception of difficulty in English writing and perceived use of writing strategies among the students from the three educational levels? Are there any gender differences in these variables among the students from the same level? How are these variables related to writing performance? 6. What pedagogic implications do these analyses have for the English teaching and learning in Singapore? What implications do these analyses have for curriculum development for English language education?
APA, Harvard, Vancouver, ISO, and other styles
2

Hernández, Ana, Magaly Lavadenz, and JESSEA YOUNG. Mapping Writing Development in Young Bilingual Learners. CEEL, 2012. http://dx.doi.org/10.15365/ceel.article.2012.2.

Full text
Abstract:
A growing interest in Two-Way Bilingual Immersion (TWBI) programs has led to increased attention to bilingualism, biliteracy, and biculturalism. This article describes the writing development in Spanish and English for 49 kindergarten students in a 50/50 Two-Way Bilingual Immersion program. Over the course of an academic year, the authors collected writing samples to analyze evidence of cross-linguistic resource sharing using a grounded theoretical approach to compare and contrast writing samples to determine patterns of cross-linguistic resource sharing in English and Spanish. The authors identified four patterns: phonological, syntactic, lexical, and metalinguistic awareness. Findings indicated that emergent writers applied similar strategies as older bilingual students, including lexical level code-switching, applied phonological rules of L1 to their respective L2s, and used experiential and content knowledge to write in their second language. These findings have instructional implications for both English Learners and native English speakers as well as for learning from students for program improvement.
APA, Harvard, Vancouver, ISO, and other styles
3

Meeson, Reginald N. Ada Lexical Analyzer Generator. Defense Technical Information Center, 1989. http://dx.doi.org/10.21236/ada227595.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Meeson, Reginald N. Ada Lexical Analyzer Generator User's Guide. Defense Technical Information Center, 1989. http://dx.doi.org/10.21236/ada227596.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

OSIYANOVA, A., and D. SARKINA. LEXICAL AND GRAMMATICAL TRANSFORMATIONS IN THE SOCIO-POLITICAL TEXTS TRANSLATION (ON THE MATERIAL OF THE US PRESIDENT J. BIDEN’S SPEECHES). Science and Innovation Center Publishing House, 2022. http://dx.doi.org/10.12731/2077-1770-2022-14-2-3-14-22.

Full text
Abstract:
The given article analyzes lexical and grammatical transformations in the translation of socio-political texts in the example of speeches made by the US President Joseph Biden and its Russian translation. A comparative analysis of the original and the translation is presented, which makes it possible to compile statistics on the frequency of use of certain translation transformations.
APA, Harvard, Vancouver, ISO, and other styles
6

Balandina, Nadiya. STRUCTURE OF MEDIAENVIRONMENT THROUGH THE PRISM OF LEXICAL INNOVATIONS. Ivan Franko National University of Lviv, 2024. http://dx.doi.org/10.30970/vjo.2024.54-55.12167.

Full text
Abstract:
Using theoretical and empirical methods, this paper proves that the modern mediaenvironment is a complex configuration made by the material and the virtual components and is reflected in the language in various ways. Innovative lexis with the component media and its systematization has become the key to understanding the mediareality, in particular, detecting the constituent elements of the structure of the mediatized environment. In total, 455 lexemes chosen from the Ukrainian dictionaries, academic publications, and the results provided by Google search engine have been analyzed. The systematization of the lexical units have been done according to the principle of the lexical and semantic field, and as the result macro-, midi-, and mini-fields have been allocated with the subsequent ranging of the lexis into the core, the centre, and the periphery. Within the macro-field MEDIAENVIRONMENT, the midi-fields have been located – TECHNICAL, TECHNOLOGICAL, and SOCIO-COMMUNICATIVE. The conceptual core of the macro-field MEDIAENVIRONMENT has been shown to form lexemes signifying space specified from the point of view of temporality, structuredness, contextuality, and evaluation. TECHNICAL midi-field is represented by the names of media equipment; TECHNOLOGICAL midi-field – by the names of methods, processes, and ways of using media equipment. The structural and semantic framework of the SOCIO-COMMUNICATIVE concentre is represented by mini-fields: WHO – FOR WHAT PURPOSE – WHAT – CHANNEL – TO WHOM – EFFECTS. The senders of information are shown as social institutions, groups, and individuals. The nominations of the addressee reflect the transformation of the passive information consumer to the user. The informational object has been specified according to its form, meaning, and the virtual «shell». The wide array of names for communication channels shows the link between the social field with the technical and technological ones, with their relative autonomy. The effects of consuming the information are verbalized by a smaller number of lexemes, largely of behavioral and evaluative nature. To sum up, the specified groups of neolexemes with the component media create a qualitative-quantitative hierarchical structure, the most numerous parts of which serve as the indicators of the most media-determined spheres of socioenvironment. Key words: mediaenvironment, structure, verbalization, neolexemes with the component media, lexical-semantical field.
APA, Harvard, Vancouver, ISO, and other styles
7

Yatsymirska, Mariya. SOCIAL EXPRESSION IN MULTIMEDIA TEXTS. Ivan Franko National University of Lviv, 2021. http://dx.doi.org/10.30970/vjo.2021.49.11072.

Full text
Abstract:
The article investigates functional techniques of extralinguistic expression in multimedia texts; the effectiveness of figurative expressions as a reaction to modern events in Ukraine and their influence on the formation of public opinion is shown. Publications of journalists, broadcasts of media resonators, experts, public figures, politicians, readers are analyzed. The language of the media plays a key role in shaping the worldview of the young political elite in the first place. The essence of each statement is a focused thought that reacts to events in the world or in one’s own country. The most popular platform for mass information and social interaction is, first of all, network journalism, which is characterized by mobility and unlimited time and space. Authors have complete freedom to express their views in direct language, including their own word formation. Phonetic, lexical, phraseological and stylistic means of speech create expression of the text. A figurative word, a good aphorism or proverb, a paraphrased expression, etc. enhance the effectiveness of a multimedia text. This is especially important for headlines that simultaneously inform and influence the views of millions of readers. Given the wide range of issues raised by the Internet as a medium, research in this area is interdisciplinary. The science of information, combining language and social communication, is at the forefront of global interactions. The Internet is an effective source of knowledge and a forum for free thought. Nonlinear texts (hypertexts) – «branching texts or texts that perform actions on request», multimedia texts change the principles of information collection, storage and dissemination, involving billions of readers in the discussion of global issues. Mastering the word is not an easy task if the author of the publication is not well-read, is not deep in the topic, does not know the psychology of the audience for which he writes. Therefore, the study of media broadcasting is an important component of the professional training of future journalists. The functions of the language of the media require the authors to make the right statements and convincing arguments in the text. Journalism education is not only knowledge of imperative and dispositive norms, but also apodictic ones. In practice, this means that there are rules in media creativity that are based on logical necessity. Apodicticity is the first sign of impressive language on the platform of print or electronic media. Social expression is a combination of creative abilities and linguistic competencies that a journalist realizes in his activity. Creative self-expression is realized in a set of many important factors in the media: the choice of topic, convincing arguments, logical presentation of ideas and deep philological education. Linguistic art, in contrast to painting, music, sculpture, accumulates all visual, auditory, tactile and empathic sensations in a universal sign – the word. The choice of the word for the reproduction of sensory and semantic meanings, its competent use in the appropriate context distinguishes the journalist-intellectual from other participants in forums, round tables, analytical or entertainment programs. Expressive speech in the media is a product of the intellect (ability to think) of all those who write on socio-political or economic topics. In the same plane with him – intelligence (awareness, prudence), the first sign of which (according to Ivan Ogienko) is a good knowledge of the language. Intellectual language is an important means of organizing a journalistic text. It, on the one hand, logically conveys the author’s thoughts, and on the other – encourages the reader to reflect and comprehend what is read. The richness of language is accumulated through continuous self-education and interesting communication. Studies of social expression as an important factor influencing the formation of public consciousness should open up new facets of rational and emotional media broadcasting; to trace physical and psychological reactions to communicative mimicry in the media. Speech mimicry as one of the methods of disguise is increasingly becoming a dangerous factor in manipulating the media. Mimicry is an unprincipled adaptation to the surrounding social conditions; one of the most famous examples of an animal characterized by mimicry (change of protective color and shape) is a chameleon. In a figurative sense, chameleons are called adaptive journalists. Observations show that mimicry in politics is to some extent a kind of game that, like every game, is always conditional and artificial.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!