Dissertations / Theses on the topic 'Anglais (langue) – Informatique'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 25 dissertations / theses for your research on the topic 'Anglais (langue) – Informatique.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Bründl, Monika Elisabeth. "Lexikalische Dynamik : Kognitiv-linguistische Untersuchungen am englischen Computerwortschatz /." Tübingen : M. Niemeyer, 2001. http://catalogue.bnf.fr/ark:/12148/cb39918274d.
Full textLabrosse, Pierre. "L'aspect en anglais de l'informatique." Paris 4, 1999. http://www.theses.fr/1999PA040150.
Full textMemet, Monique. "Anglais de spécialité dans le domaine du génie électrique : aspects sociolinguistiques et applications didactiques résultant du traitement informatique des données." Bordeaux 2, 1994. http://www.theses.fr/1994BOR21010.
Full textBlanvillain, Odile. "Représentation formelle de l'emploi de should en anglais contemporain, en vue d'un traitement automatique." Paris 7, 1993. http://www.theses.fr/1993PA070074.
Full textThe aim of this thesis is to propose a formal (metalinguistic) representation of the different uses of should, in modern english, within an utterercentered approach, and to examine how this study can contribute to a possible automatic processing of the modal, in context. Its aim is also to bring out the particular status of should in comparison with shall, must or ought to, for example, and to understand the interplay of epistemic and root modalities that allows should, in its "ambiguous" uses. It appears that the difference between should and shall, must or ought to is essentially qualitative. Besides, the different possible ambiguities in the interpretation of should can be explained, in each case, by the specific characteristics of its fonctioning (positive valuation, particularities of its deontic origin, etc. . ) concerning the contribution to a possible automatic processing, the bringing out of these characteristics can be of importance within a work of automatic comprehension of texts. But, on the syntactic point of view, on the contrary, it appears extremely difficult to determine automatically the value of the modal
Baldo, Sabrina. "Représentation des valeurs sémantiques de l'auxiliaire modal would en anglais : étude linguistique et système d'exploration contextuelle en vue du traitement informatique de sa traduction automatique en français." Paris 4, 2000. http://www.theses.fr/2000PA040221.
Full textKübler, Natalie. "L'automatisation de la correction d'erreurs syntaxique : application aux verbes de transfert en anglais pour francophones." Paris 7, 1995. http://www.theses.fr/1995PA070007.
Full textCorrecting syntactic errors is a complex problem. Organizing the linguistic data represents the first important step, especially in the field of second-language error correction. The errors that native speakers fo french make in english must be taken into account when structuring this information. It is also necessary to have a wellformed description of the correct syntactic structures of a language in order to compare them with incorrect structures. Therefore, we describe the syntactic structures of transfer verbs - i. E. Verbs that express the transfer of an object or of words between to persons - in the theoretical and methodological frame of the lexicon-grammar. This description is compared with a typology of errors and with the corresponding structures of the french transfer verbs, which enables us to have a well-formed description of the incorrect structures produced by french-speakers in english. The specific mecanism used to detect the errors is based on a variation of finite state automata. This mecanism does not analyses whole sentences but is based on island processing. Looking for the specific incorrect structures that have previously been described
Depierre, Amélie. "La Composition en anglais scientifique : application au domaine de la physico-chimie." Littoral, 2001. http://www.theses.fr/2001DUNK0070.
Full textThis study stems from teaching experience and comes as a response to students' questions and errors. It has two main objectives : on the one hand a critical review of existing research on compounds and other complex nominals in English, and on the other hand the application of the theoretical data available in order to help French-speaking students and researchers, who need to communicate (understand, speak and write) in English, namely in the domain of physico-chemical sciences. After a survey of the characteristics of compounds, and an investigation of axamples taken from several corpora, a new morphological classification of binominal compounds is suggested, then the semantic relations between their elements are pointed out. The results are extended to longer nominals, with a particular attention paid to words built from Greek and Latin bases, and acronyms, which abound in the field of sciences. These competing structures : N₂ N₁, N₂'s N₁ and N₁ of N₂ are contrasted, as they are a major source of error among French speakers, especially at the encoding stage. Lastly, an electronic glossary in Excel format including some of the terms of the domain, together with examples of attested sentences, has been developed. One of the future objectives is to transfer the glossary in HTML format in order to make it more available for use in various domains
Zumstein, Franck. "Variation accentuelle, variation phonétique : étude systématique fondée sur des corpus lexico-phonétiques informatisés anglais." Poitiers, 2007. http://www.theses.fr/2007POIT5025.
Full textIn this piece of work is examined word-stress variation in contemporary English, such as in the word "controversy" (stressed /1000/ or /0100/). Our linguistic analyses are based on automatic data retrieval from a computer-readable version of John Well's "Longman Pronunciation Dictionary" published in 1990. Eighteenth and nineteenth century English dictionaries are also consulted so as to determine whether variants are old stressings or new patterns. The results of this study, which is a follow-up of Lionel Guierre's work on word stress in English, highlight the direction of the change of the alternating stress patterns and show that regular criteria, notably the meaning of words, account for such variations
Boughedaoui, Mourad. "Anglais spécialisé de l'informatique : contribution à l'étude de la complexification des noms composés par chaînage." Bordeaux 2, 1995. http://www.theses.fr/1995BOR2A001.
Full textThis work presents a study of the complexification process of string compound nouns (SCNS) through different registers of a corpus of computer science written texts. This process is characterized by the existence of various overlapping levels of the component parts within SCNS. In this study, these levels are identified using a formal segmentation based essentially on the location of collocational sequences in the linguistic environment of SCNS. This approach made it possible to highlight the active role of collocations in the binary structuring of SCNS as well as in the formation of compound adjectives. In order to establish any inter-relationship between the complexification phenomenon and the four registers studied, a statistical analysis of the frequencies and complexities of SCNS has been carried out. An SCN typology has been drawn up so as to put forward the complex relationship resulting from the multiple potential combinations of the various syntactic categories featuring in SCNS. The last section lays special emphasis on the pedagogic implications
Lutrand-Pezant, Brigitte. "Les propositions complétives en that en anglais contemporain." Paris 4, 2003. http://www.theses.fr/2003PA040212.
Full textThis study, based on a computer-based corpus of over 7 000 examples, falls into three parts : first a presentation of the linguistic knowledge so far on the subject of that clauses, then a description of the forms present in the corpus together with statistics and third an analysis of their lexical and syntactic environments. The issue of the choice between that and Ø has been carefully examined. The behaviour of these clauses has been studied through literary texts of the 19th and 20th centuries as well as in journalistic and scientific writings. Oral English has been compared to written English when necessary. This research did also try to show the characteristics of South African, Irish, British and American English regarding these clauses
Resche, Catherine. "La bourse des valeurs : Terminologie comparée du français et de l'anglais." Rouen, 1986. http://www.theses.fr/1986ROUEL017.
Full textThe aim of the author here is to study the terminology of the stock exchange through the contemporary press abd compare the french, british and american terms so as to establish equivalents which will later be computerized as part of a data bank. The study is composed of two parts: the first one concentrates on theory and the second one consists of "index cards" -two thousands of them with definitions and contexts, among which those starting with the letter "a" in french and the letter "s" in english have been chosen to appear here as an illustration of the whole. In the first part, the author tries to define comparative terminology by opposing it to other sciences (linguistics, lexicology, translation techniques. . . ). She analyses the problem raised by switching from one language to the other, underlining the differences that exist between the french, british and american civilisations and systems, and she explains her method. A brief review of the recent reforms helps her to illustrate the terms that have just been coined and comment upon their evolution in the language. The terminology of the stock exchange is more complex than it seems; apart from the technical terms, whose nature and formation are examined, it is also composed of images and adages which are worth stuying. A questionnaire sent to specialists throws some light on questions concerning synonyms, acronyms, terms borrowed from another language, etc. . . The issue of standardization is also raised by the observation of polysemic terms; finally the choice of the computerized formula is justified by the analysis of the advantages offered by a data bank and by the demands upon logic this system implies thus guaranteeing the accuracy of the whole
Ganeval-Béciri, Hélène. "La néologie dans le domaine microinformatique professionnel en français et en anglais : création lexicale et explicitation de contexte." Paris 7, 1999. http://www.theses.fr/1999PA070108.
Full textJoulia, Danielle. "Faciliter la lecture-compréhension de documentations informatiques en anglais : de l'analyse à l'expérimentation didactique assistée par ordinateur." Thesis, Toulouse 2, 2011. http://www.theses.fr/2011TOU20102.
Full textReading documentation is an essential skill to develop in computer science students, whose level of English is low and who are poorly motivated by English, a subject considered to be minor in a science curriculum. This research aimed at proposing learning and teaching solutions to make them proficient in reading professional texts. We selected a bottom-up approach, inspired by fieldwork, and a top-down approach from the theoretical foundations related to the language skills involved. The first part expounds the theoretical framework and clarifies the terms of our research problem (English for Specific Purposes, reading, comprehension, ICT/CALL), which allowed us to construct our object of research. ESP is thus defined as a teaching/learning and linguistic object, reading and understanding are characterized from cognitive and SLA points of view, and the contribution of ICT to reading comprehension is examined. The second part, which is more practical, contains a corpus analysis of specialized texts and presents a teaching approach implemented through a software program designed to enhance reading comprehension. The latter is then experimented to test our hypothesis of a better comprehension in student groups who used hypertext reading aids compared to control groups, thus evaluating the impact of our research tool. We conclude on the validity of this approach to improve the reading strategies of computer science students
Magno, e. Silva Walkyria. "Etude linguistique d'un corpus de manuels d'informatique : application à la lecture en langue étrangère." Toulouse 2, 2002. http://www.theses.fr/2002TOU20048.
Full textA corpus of 23 hardware and software computer manuals was studied in order to find out their main characteristics. Focusing on the "Installation and Use" section of these manuals, a linguistic analysis was conducted under the light of two main theoretical currents. On the one hand, the methodology derived from the theory of sublanguages (Z. Harris, 1968, 1982; L. Hirschman & N. Sager, 1982; J. Lehrberger, 1982) helped establish the common grounds of phrase stuctures found in this written genre. Techniques base on distributional analysis were employed to determine the information format of the six most frequent action verbs in the corpus (J. M. Adam, 1985,1992) proved the high incidence of the procedural discourse and showed the interrelation among this type of text and the descriptive and explanatory ones. Pedagogical actions intended to a public of university students who need to read texts in English in the domain of computer sciences were eventually proposed
Intakosum, Sasi. "Proposition d'une pédagogie d'enseignement de la traduction de textes techniques d'anglais et de français en thai͏̈ : application aux textes informatiques." Paris 3, 2000. http://www.theses.fr/2000PA030066.
Full textYahiaoui, Abdelghani. "Conception et développement d'un outil d'aide à la traduction anglais/arabe basé sur des corpus parallèles." Thesis, Lyon, 2017. http://www.theses.fr/2017LYSE2042.
Full textWe create an innovative English/Arabic translation aid tool to meet the growing need for online translation tools centered on the Arabic language. This tool combines dictionaries appropriate to the specificities of the Arabic language and a bilingual concordancer derived from parallel corpora. Given its agglutinative and unvoweled nature, Arabic words require specific treatment. For this reason, and to construct our dictionary resources, we base on Buckwalter's morphological analyzer which, on the one hand, allows a morphological analysis taking into account the complex composition of the Arabic word (proclitic, prefix, stem, suffix, enclitic), and on the other hand, provides translational resources enabling rehabilitation in a translation system. Furthermore, this morphological analyzer is compatible with the approach defined around the DIINAR database (DIctionnaire Informatisé de l’Arabe - Computerized Dictionary for Arabic), which was constructed, among others, by members of our research team. In response to the contextual issue in translation, a bilingual concordancer was developed from parallel corpora. The latter represent a novel linguistic resource with multiple uses, in this case aid for translation. We therefore closely analyse these corpora, their alignment methods, and we proposed a mixed approach that significantly improves the quality of sub-sentential alignment of English-Arabic corpora. Several technologies have been used for the implementation of this translation aid tool which have been made available online (tarjamaan.com) and which allow the user to search the translation of millions of words and expressions while visualizing their original contexts. An evaluation of this tool has been made with a view to its optimization and its enlargement to support other language pairs
Gaillat, Thomas. "Reference in interlanguage : the case of this and that : From linguistic annotation to corpus interoperability." Sorbonne Paris Cité, 2016. http://www.theses.fr/2016USPCC107.
Full textThis thesis describes unexpected constructions based on THIS and THAT by French and Spanish learners of English. Chapter 1 raises the issue of the study of THIS and THAT as markers in the two microsystems of pro-forms and deictics. Chapter 2 covers different types of analyses of reference with THIS and THAT in native English and refers to different theoretical frameworks (Cornish, Cotte, Halliday & Hasan, Kleiber, Fraser & Joly, Lapaire & Rotgé). It cross-references representations (anaphora/deixis; endophoricity/exophoricity) with an analysis of functional realisations. Chapter 3 broaches the issue of interlanguage analysis, and it shows that a dynamic systemic approach grounded in the functional distinction of the forms is necessary. Chapter 4 gives details about existing annotation tagsets for English corpora (Penn Treebank, Claws7, ICE-GB). It shows the need for a finer-grained annotation relying on functional tags and for semantic information on the positions (subject v. Oblique). Chapter 5 describes the multilayer annotation structure which is implemented for the analysis of different corpora. It also covers the methods used to automatically annotate functional categories (as well as their evaluation), and it justifies the choices made to support corpus interoperability. Chapter 6 offers a regression analysis which provides evidence on the tendencies of the operationalised variables (the L1, the written or spoken mode of the corpora and the type of reference). Chapter 7 examines the role of the previously coded linguistic properties of the analysis. With the use of classifiers, it describes a system for automatic error analysis. Chapter 8 concludes on the methodologies used in the thesis and their implications in linguistic analysis
Zellal, Nassim. "Proposition d'un modèle de définition terminologique : approche du domaine de l'informatique en langues française, arabe et anglaise." Paris 3, 2008. http://www.theses.fr/2008PA030086.
Full textOur study seeks to examine the structure of terminological definition by systematizing its writing. We start from the general hypothesis that the terminological definition would respect a systemic general model. This one reflects the organization of constants which are extracted from a french, english and arabic corpus about computer science. These constants are generated by an abstraction and generalization operation which enables to identify partitive concepts and functional characteristics. Thus, each definition reflects a system of concepts and its relationships. In order to refine this model, we have identified specific functional schemas. This is the subject of our second hypothesis validation. The first part of the thesis describes, through our readings, a summary of the fundamental principals of the traditional theory of terminology which is also called the classical theory. In the end of this summary, we synthetically set out the main theoretical criticism expressed by the new terminological approaches. We then present the three constitutive aspects of the theoretical terminological triangle and we try to distinguish them from its linguistics counterparts. These distinctions are indeed essential because they allow putting the emphasis on the scientific status and autonomy of terminology. This part comes after the presentation of the various relationships between the concepts, within a system and the multidimensionality of the concept. We end this first part of the thesis by defining the “definition” with a proposal of a definition typology. We also present in this chapter fundamental recommendations about the definition writing, in order to rigorously define the subject of our study which is the terminological definition. In the second part, we present the domain of computer science, the reasons of its choice and its conceptual representation, in order to obtain a panoramic view. We then mark the boundary of its subdomains and the sample of concepts we have chosen by respecting a qualitative perspective. We then make a study of terms that designate the peripheral equipments and its components. The metaphor and synonymy are also set out and analyzed in this chapter in their various forms of realization. This linguistic study turns out to be fundamental because we have to take into account some linguistic constraints which concern the specialized vocabulary as well. The next chapter introduces the study of terminological definition
Brouat, Thérèse. "Le dire et l'induireContribution à la didactique de l'anglais des spécialités scientifiques et techniques." Chambéry, 1997. http://www.theses.fr/1997CHAML004.
Full textThis research aims at studying advertisements for computer products published in british and american computer magazines in order to identify their potential with a view to teaching English to computing students. The analysis deals with the way argumentation is built on the close association of pictures and discourse. It emphasizes the effects of their dynamic interaction on the receiver and the role of the latter in construing meaning. Ranging from a descriptive pole where referential function prevails and a poetic pole open to connotations, advertisements for computer products are hybrid and multidimensional messages. Our approach consists in showing how scientific reference is embedded in other discourse types such as description, explanation, dialogue, story-telling, illustration, poetry, etc. , to appeal to the reader's emotional and rational sides. When the reader is an ESP student, these multidimensional messages can be adapted into teaching material in order to meet specific language needs that include, but are not limited to, technical terminology. For ESP teachers, advertisements provide a mine of easily updatable information in an ever-evolving domain. The last part of this research gives suggestions and guidelines on how to integrate advertisements for computer products in a course for ESP students. It focuses on activities that build on the cognitive, cultural and affective energy of learners and involve them in the process of mastering comprehension and production skills
Bendinelli, Marion. "Etude des auxiliaires modaux et des semi-modaux dans les débats présidentiels américains (1960-2008) : analyse qualitative et quantitative." Nice, 2012. http://www.theses.fr/2012NICE2017.
Full textThis dissertation consists in the study of modals and semi-modals in a corpus of American presidential debates (1960-2008). Our aim is to characterize their use with respect to issues of power and influence. On the one hand, the theoretical framework borrows from various fields of linguistic analysis, namely enunciation, pragmatics and discourse analysis; the latter helps define notions such as persuasive communication, manipulation and influence strategies. Besides, we suggest that the modals and some semi-modals can activate dialogic echoes and serve as evidential markers. On the other hand, the study relies on a computer-based discourse analysis methodology, namely logometry. Considered as a valuable heuristic approach, it runs statistical measures on linguistic data – be it tokens, lemmas or parts of speech – so as to provide the analyst with frequency lists and cooccurrence networks, among other functionalities. Logometry also offers a bottom-up as well as a top-down approach to corpus scrutiny. The results help identify three enunciative strategies depending on the level of commitment of the utterer with his epistemic or non-epistemic judgements. Also, they determine the existence of several types of discourses that are considered in the light of various issues: speech practice evolution, ideological positioning, rhetorical identity of political parties, candidates’ campaign strategies. As far as the modals and semi-modals CAN, MUST, HAVE TO, HAVE GOT TO, NEED TO and OUGHT TO are concerned, a close examination of their contexts of use shows that they reflect different strategies of presentation of the self, of others, of the audience and of the factual context
Braud, Chloé. "Identification automatique des relations discursives implicites à partir de corpus annotés et de données brutes." Sorbonne Paris Cité, 2015. https://hal.inria.fr/tel-01256884.
Full textBuilding discourse parsers is currently a major challenge in Natural Language Processing. The identification of the relations (such as Explanation, Contrast. . . ) linking spans of text in the document is the main difficulty. Especially, identifying the so-called implicit relations, that is the relations that lack a discourse connective (such as but, because. . . ), is known as an hard tank sine it requires to take into account varions factors, and because it leads to specific difficulties in a classification system. In this thesis, we use raw data to improve automatic identification of implicit relations. First, we propose to use discourse markers in order to automatically annotate new data. We use domain adaptation methods to deal with the distributional differences between automatically and manually annotated data : we report improvements for systems built on the French corpus ANNODIS and on the English corpus Penn Discourse Treebank. Then, we propose to use word representations built from raw data, which may be automatically annotated with discourse markers, in order to feed a representation of the data based on the words found in the spans of text to be linked. We report improvements on the English corpus Penn Discourse Treebank, and especially we show that this method alleviates the need for rich resources, available but for a few languages
Malik, Mohamed Mahdi. "Structures prédicatives nominales en Anglais : acquisition de données lexicales pour l'analyse automatique de textes." Thesis, Aix-Marseille 2, 2010. http://www.theses.fr/2010AIX22017/document.
Full textIn this thesis, we focus on the relation that may exist between verbal predicates (e.g., regulate)and nominal predicates (e.g., regulation) whose argument structures involve common information.We make a formalization of the conditions in which equivalent relations between verbal and nominalconstructions are carried out. Bringing out the equivalence of argument structures between these twotypes of constructions is fundamental for achieving, for example, very efficient Information Extractionsystems. Based on data from the Specialist Lexicon, we propose a reasonable prediction of the syntacticbehavior of nominal arguments, which belong to different predicate noun phrases (PNPs), when theyare in postmodifier position. This study has led us to design a set of algorithms and develop a platform,PredicateDB, to produce a lexicon of nominalizations. For each entry belonging to this lexicon, we havedefined its argument structures and achievements in PNPs whose arguments are marked by specificprepositions
Piccato, Mariangela. "Création et exploitation d'un corpus trilingue du tourisme (italien/français/anglais) en vue de la réalisation d'une base de données lexicale informatisée." Thesis, Lyon 2, 2012. http://www.theses.fr/2012LYO20051.
Full textOur study concerns the language of tourism from a lexicographical perspective.Exploiting the web we realized a corpus ad hoc. This corpus is composed by about 10.000 texts in three languages (French, Italian and English), aligned using “Alinea”.Starting from terminological extraction, we analysed some collocations at the aim to create a trilingual and tri-directional glossary.We chose this subject according to the increasing importance taken from tourism economy in the world.Our study fields are thematic terminology, corpus linguistics and automatic language treatment.The first chapter presents the study field of our research. First of all, we introduced to corpus linguistics presenting the different categories of corpus and pointing out our attention on two main notions: representativeness and context.Therefore, we explained the link between Language for Special Purposes and tourism discourse as a Specialized Discourse.In the second chapter, we showed the trilingual thematic corpus we created during our researches. We described the main steps to create a corpus: collection of texts, cleaning and annotation.In this chapter, we gave a particular attention to the presentation of “Alinea”.Finally, the third chapter is a study of frequent collocations with the term “town” (ville).The annexes present the glossary as well as the methodological principals we followed in the redaction
Frini, Marouane. "Diagnostic des engrenages à base des indicateurs géométriques des signaux électriques triphasés." Thesis, Lyon, 2018. http://www.theses.fr/2018LYSES052.
Full textAlthough they are widely used, classical vibration measurements have several limitations. Vibration analysis can only identify about 60% of the defects that may occur in mechanical systems. However, the main drawbacks of vibration measurements are the difficult access to the transmission system in order to place the sensor as well as the consequent cost of implementation. This results in sensitivity problems relative to the position of the installation and the difficulty to distinguish the source of vibration because of the diversity of mechanical excitations that exist in the industrial environment.Hence, the Motor Current Signatures Analysis (M.C.S.A.) represents a promising alternative to the vibration analysis and has therefore been the subject of increasing attention in recent years. Indeed, the analysis of electrical signatures has the advantage of being a technically accessible method as well as inexpensive and non-intrusive to the system. Techniques based on currents and voltages only require the motor’s electrical measurements which are often already supervised for the purposes of the control and the protection of the electrical machines. This process was mainly used for the detection of motors faults such as rotor bars breakage and eccentricity faults as well as bearings defects. On the other hand, very little research has been focused on gear faults detection using the current analysis. In addition, three-phase electrical signals are characterized by specific geometric representations related to their waveforms and they can serve as different indicators providing additional information. Among these geometric indicators, the Park and Concordia transforms model the electrical components in a two-dimensional coordinate system and any deviation from the original representation indicates the apparition of a malfunction. Moreover, the differential equations of Frenet-Serret represent the trajectory of the signal in a three-dimensional euclidean space and thus indicate any changes in the state of the system. Although they have been previously used for bearing defects, these indicators have not been applied in the detection of gear defects using the analysis of electrical current signatures. Hence, the innovative idea of combining these indicators with signal processing techniques, as well as classification techniques for gears diagnosis using the three-phase motor’s electrical current signatures analysis is established.Hence, in this work, a new approach is proposed for gear faults diagnosis using the motor currents analysis, based on a set of geometric indicators (Park and Concordia transforms as well as the properties of the Frenet-Serret frame). These indicators are part of a specifically built fault signatures library and which also includes the classical indicators used for a wide range of faults. Thus, a proposed estimation algorithm combines experimental measurements of electrical signals with advanced signal processing methods (Empirical Mode Decomposition, ...). Next, it selects the most relevant indicators within the library based on feature selection algorithms (Sequential Backward Selection and Principal Component Analysis). Finally, this selection is combined with non-supervised classification (K-means) for the distinction between the healthy state and faulty states. It was finally validated with a an additional experimental configuration in different cases with gear faults, bearing faults and combined faults with various load levels
Lafrance, Angélique. "Étude de l’évolution dans la terminologie de l’informatique en anglais avant et après 2006." Thèse, 2016. http://hdl.handle.net/1866/18768.
Full textIn this study, we propose a method to observe lexical changes (neology and necrology) in English in the field of computer science in short-period diachrony. Since computer science evolves quickly, we believe that a short-period diachronic approach (over a period of 10 years) lends itself to studying the terminology of that field. For this purpose, we built a corpus in English with articles taken from computer science magazines for the general public, PC Magazine and PC World, covering the years 2001 to 2010. The corpus was divided into two subcorpora: 2001-2005 and 2006-2010. We chose year 2006 as a pivot, because Facebook (the most popular social network) has been open to the public since that year, and we believed that would cause quick changes in computer science terminology. For each of the magazines, we selected one issue per year from 2001 to 2010, for a total of about 540,000 words for the 2001-2005 subcorpus and about 390,000 words for the 2006-2010 subcorpus. Each subcorpus was submitted to term extractor TermoStat to extract nominal, verbal and adjectival term candidates. We proceeded to three experiment groups, according to the reference corpus used. In the first experiment group (Exp1), we compared each subcorpus to the default reference corpus in TermoStat for English, a British National Corpus (BNC) extract. In the second experiment group (Exp2), we compared each subcorpus to the whole computer science corpus we created. In the third experiment group (Exp3), we compared the two subcorpora with each other. After cleaning up the term candidates lists thus obtained to retain only the terms in the field of computer science, and generating data about relative frequency and relative specificity of the terms between subcorpora, we proceeded to the validation of novelty and obsolescence of the first terms of each list to determine whether the proposed method works better with a particular type of lexical change (novelty or obsolescence), part of speech (nominal, verbal or adjectival term), or experiment group. The validation results show that the method seems to work better with neology extraction than with necrology extraction. Also, we had better results with nominal and adjectival terms than with verbal terms. Finally, we had much more results with Exp1 than with Exp2 and Exp3.