Journal articles on the topic 'Natural language processing (Computer science) – Research'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 journal articles for your research on the topic 'Natural language processing (Computer science) – Research.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse journal articles on a wide variety of disciplines and organise your bibliography correctly.

1

Zhao, Liping, Waad Alhoshan, Alessio Ferrari, et al. "Natural Language Processing for Requirements Engineering." ACM Computing Surveys 54, no. 3 (2021): 1–41. http://dx.doi.org/10.1145/3444689.

Full text
Abstract:
Natural Language Processing for Requirements Engineering (NLP4RE) is an area of research and development that seeks to apply natural language processing (NLP) techniques, tools, and resources to the requirements engineering (RE) process, to support human analysts to carry out various linguistic analysis tasks on textual requirements documents, such as detecting language issues, identifying key domain concepts, and establishing requirements traceability links. This article reports on a mapping study that surveys the landscape of NLP4RE research to provide a holistic understanding of the field. Following the guidance of systematic review, the mapping study is directed by five research questions, cutting across five aspects of NLP4RE research, concerning the state of the literature, the state of empirical research, the research focus, the state of tool development, and the usage of NLP technologies. Our main results are as follows: (i) we identify a total of 404 primary studies relevant to NLP4RE, which were published over the past 36 years and from 170 different venues; (ii) most of these studies (67.08%) are solution proposals, assessed by a laboratory experiment or an example application, while only a small percentage (7%) are assessed in industrial settings; (iii) a large proportion of the studies (42.70%) focus on the requirements analysis phase, with quality defect detection as their central task and requirements specification as their commonly processed document type; (iv) 130 NLP4RE tools (i.e., RE specific NLP tools) are extracted from these studies, but only 17 of them (13.08%) are available for download; (v) 231 different NLP technologies are also identified, comprising 140 NLP techniques, 66 NLP tools, and 25 NLP resources, but most of them—particularly those novel NLP techniques and specialized tools—are used infrequently; by contrast, commonly used NLP technologies are traditional analysis techniques (e.g., POS tagging and tokenization), general-purpose tools (e.g., Stanford CoreNLP and GATE) and generic language lexicons (WordNet and British National Corpus). The mapping study not only provides a collection of the literature in NLP4RE but also, more importantly, establishes a structure to frame the existing literature through categorization, synthesis and conceptualization of the main theoretical concepts and relationships that encompass both RE and NLP aspects. Our work thus produces a conceptual framework of NLP4RE. The framework is used to identify research gaps and directions, highlight technology transfer needs, and encourage more synergies between the RE community, the NLP one, and the software and systems practitioners. Our results can be used as a starting point to frame future studies according to a well-defined terminology and can be expanded as new technologies and novel solutions emerge.
APA, Harvard, Vancouver, ISO, and other styles
2

Heidorn, P. Bryan. "Natural language processing." Information Processing & Management 32, no. 1 (1996): 122–23. http://dx.doi.org/10.1016/s0306-4573(96)90089-8.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Li, Yong, Xiaojun Yang, Min Zuo, Qingyu Jin, Haisheng Li, and Qian Cao. "Deep Structured Learning for Natural Language Processing." ACM Transactions on Asian and Low-Resource Language Information Processing 20, no. 3 (2021): 1–14. http://dx.doi.org/10.1145/3433538.

Full text
Abstract:
The real-time and dissemination characteristics of network information make net-mediated public opinion become more and more important food safety early warning resources, but the data of petabyte (PB) scale growth also bring great difficulties to the research and judgment of network public opinion, especially how to extract the event role of network public opinion from these data and analyze the sentiment tendency of public opinion comment. First, this article takes the public opinion of food safety network as the research point, and a BLSTM-CRF model for automatically marking the role of event is proposed by combining BLSTM and conditional random field organically. Second, the Attention mechanism based on vocabulary in the field of food safety is introduced, the distance-related sequence semantic features are extracted by BLSTM, and the emotional classification of sequence semantic features is realized by using CNN. A kind of Att-BLSTM-CNN model for the analysis of public opinion and emotional tendency in the field of food safety is proposed. Finally, based on the time series, this article combines the role extraction of food safety events and the analysis of emotional tendency and constructs a net-mediated public opinion early warning model in the field of food safety according to the heat of the event and the emotional intensity of the public to food safety public opinion events.
APA, Harvard, Vancouver, ISO, and other styles
4

Mazzei, Daniele, Filippo Chiarello, and Gualtiero Fantoni. "Analyzing Social Robotics Research with Natural Language Processing Techniques." Cognitive Computation 13, no. 2 (2021): 308–21. http://dx.doi.org/10.1007/s12559-020-09799-1.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Taskin, Zehra, and Umut Al. "Natural language processing applications in library and information science." Online Information Review 43, no. 4 (2019): 676–90. http://dx.doi.org/10.1108/oir-07-2018-0217.

Full text
Abstract:
Purpose With the recent developments in information technologies, natural language processing (NLP) practices have made tasks in many areas easier and more practical. Nowadays, especially when big data are used in most research, NLP provides fast and easy methods for processing these data. The purpose of this paper is to identify subfields of library and information science (LIS) where NLP can be used and to provide a guide based on bibliometrics and social network analyses for researchers who intend to study this subject. Design/methodology/approach Within the scope of this study, 6,607 publications, including NLP methods published in the field of LIS, are examined and visualized by social network analysis methods. Findings After evaluating the obtained results, the subject categories of publications, frequently used keywords in these publications and the relationships between these words are revealed. Finally, the core journals and articles are classified thematically for researchers working in the field of LIS and planning to apply NLP in their research. Originality/value The results of this paper draw a general framework for LIS field and guides researchers on new techniques that may be useful in the field.
APA, Harvard, Vancouver, ISO, and other styles
6

Chen, Xieling, Ruoyao Ding, Kai Xu, Shan Wang, Tianyong Hao, and Yi Zhou. "A Bibliometric Review of Natural Language Processing Empowered Mobile Computing." Wireless Communications and Mobile Computing 2018 (June 28, 2018): 1–21. http://dx.doi.org/10.1155/2018/1827074.

Full text
Abstract:
Natural Language Processing (NLP) empowered mobile computing is the use of NLP techniques in the context of mobile environment. Research in this field has drawn much attention given the continually increasing number of publications in the last five years. This study presents the status and development trend of the research field through an objective, systematic, and comprehensive review of relevant publications available from Web of Science. Analysis techniques including a descriptive statistics method, a geographic visualization method, a social network analysis method, a latent dirichlet allocation method, and an affinity propagation clustering method are used. We quantitatively analyze the publications in terms of statistical characteristics, geographical distribution, cooperation relationship, and topic discovery and distribution. This systematic analysis of the field illustrates the publications evolution over time and identifies current research interests and potential directions for future research. Our work can potentially assist researchers in keeping abreast of the research status. It can also help monitoring new scientific and technological development in the research field.
APA, Harvard, Vancouver, ISO, and other styles
7

Cambria, Erik, and Bebo White. "Jumping NLP Curves: A Review of Natural Language Processing Research [Review Article]." IEEE Computational Intelligence Magazine 9, no. 2 (2014): 48–57. http://dx.doi.org/10.1109/mci.2014.2307227.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Leeson, William, Adam Resnick, Daniel Alexander, and John Rovers. "Natural Language Processing (NLP) in Qualitative Public Health Research: A Proof of Concept Study." International Journal of Qualitative Methods 18 (January 1, 2019): 160940691988702. http://dx.doi.org/10.1177/1609406919887021.

Full text
Abstract:
Qualitative data-analysis methods provide thick, rich descriptions of subjects’ thoughts, feelings, and lived experiences but may be time-consuming, labor-intensive, or prone to bias. Natural language processing (NLP) is a machine learning technique from computer science that uses algorithms to analyze textual data. NLP allows processing of large amounts of data almost instantaneously. As researchers become conversant with NLP, it is becoming more frequently employed outside of computer science and shows promise as a tool to analyze qualitative data in public health. This is a proof of concept paper to evaluate the potential of NLP to analyze qualitative data. Specifically, we ask if NLP can support conventional qualitative analysis, and if so, what its role is. We compared a qualitative method of open coding with two forms of NLP, Topic Modeling, and Word2Vec to analyze transcripts from interviews conducted in rural Belize querying men about their health needs. All three methods returned a series of terms that captured ideas and concepts in subjects’ responses to interview questions. Open coding returned 5–10 words or short phrases for each question. Topic Modeling returned a series of word-probability pairs that quantified how well a word captured the topic of a response. Word2Vec returned a list of words for each interview question ordered by which words were predicted to best capture the meaning of the passage. For most interview questions, all three methods returned conceptually similar results. NLP may be a useful adjunct to qualitative analysis. NLP may be performed after data have undergone open coding as a check on the accuracy of the codes. Alternatively, researchers can perform NLP prior to open coding and use the results to guide their creation of their codebook.
APA, Harvard, Vancouver, ISO, and other styles
9

Smith, Glenn Gordon, Robert Haworth, and Slavko Žitnik. "Computer Science Meets Education: Natural Language Processing for Automatic Grading of Open-Ended Questions in eBooks." Journal of Educational Computing Research 58, no. 7 (2020): 1227–55. http://dx.doi.org/10.1177/0735633120927486.

Full text
Abstract:
We investigated how Natural Language Processing (NLP) algorithms could automatically grade answers to open-ended inference questions in web-based eBooks. This is a component of research on making reading more motivating to children and to increasing their comprehension. We obtained and graded a set of answers to open-ended questions embedded in a fiction novel written in English. Computer science students used a subset of the graded answers to develop algorithms designed to grade new answers to the questions. The algorithms utilized the story text, existing graded answers for a given question and publicly accessible databases in grading new responses. A computer science professor used another subset of the graded answers to evaluate the students’ NLP algorithms and to select the best algorithm. The results showed that the best algorithm correctly graded approximately 85% of the real-world answers as correct, partly correct, or wrong. The best NLP algorithm was trained with questions and graded answers from a series of new text narratives in another language, Slovenian. The resulting NLP algorithm model was successfully used in fourth-grade language arts classes for providing feedback to student answers on open-ended questions in eBooks.
APA, Harvard, Vancouver, ISO, and other styles
10

Suzuki, Kenji. "AI: A New Open Access Journal for Artificial Intelligence." AI 1, no. 2 (2020): 141–42. http://dx.doi.org/10.3390/ai1020007.

Full text
Abstract:
As a branch of computer science, artificial intelligence (AI) attempts to understand the essence of intelligence, and produce new kinds of intelligent machines that can respond in a similar way to human intelligence, with broad research areas of machine and deep learning, data science, reinforcement learning, data mining, knowledge discovery, knowledge reasoning, speech recognition, natural language processing, language recognition, image recognition, computer vision, planning, robotics, gaming, and so on [...]
APA, Harvard, Vancouver, ISO, and other styles
11

Argamon, Shlomo Engelson. "Register in computational language research." Register Studies 1, no. 1 (2019): 100–135. http://dx.doi.org/10.1075/rs.18015.arg.

Full text
Abstract:
Abstract Shlomo Argamon is Professor of Computer Science and Director of the Master of Data Science Program at the Illinois Institute of Technology (USA). In this article, he reflects on the current and potential relationship between register and the field of computational linguistics. He applies his expertise in computational linguistics and machine learning to a variety of problems in natural language processing. These include stylistic variation, forensic linguistics, authorship attribution, and biomedical informatics. He is particularly interested in the linguistic structures used by speakers and writers, including linguistic choices that are influenced by social variables such as age, gender, and register, as well as linguistic choices that are unique or distinctive to the style of individual authors. Argamon has been a pioneer in computational linguistics and NLP research in his efforts to account for and explore register variation. His computational linguistic research on register draws inspiration from Systemic Functional Linguistics, Biber’s multi-dimensional approach to register variation, as well as his own extensive experience accounting for variation within and across text types and authors. Argamon has applied computational methods to text classification and description across registers – including blogs, academic disciplines, and news writing – as well as the interaction between register and other social variables, such as age and gender. His cutting-edge research in these areas is certain to have a lasting impact on the future of computational linguistics and NLP.
APA, Harvard, Vancouver, ISO, and other styles
12

Moon, Nazmun Nessa, Imrus Salehin, Masuma Parvin, et al. "Natural language processing based advanced method of unnecessary video detection." International Journal of Electrical and Computer Engineering (IJECE) 11, no. 6 (2021): 5411. http://dx.doi.org/10.11591/ijece.v11i6.pp5411-5419.

Full text
Abstract:
<span>In this study we have described the process of identifying unnecessary video using an advanced combined method of natural language processing and machine learning. The system also includes a framework that contains analytics databases and which helps to find statistical accuracy and can detect, accept or reject unnecessary and unethical video content. In our video detection system, we extract text data from video content in two steps, first from video to MPEG-1 audio layer 3 (MP3) and then from MP3 to WAV format. We have used the text part of natural language processing to analyze and prepare the data set. We use both Naive Bayes and logistic regression classification algorithms in this detection system to determine the best accuracy for our system. In our research, our video MP4 data has converted to plain text data using the python advance library function. This brief study discusses the identification of unauthorized, unsocial, unnecessary, unfinished, and malicious videos when using oral video record data. By analyzing our data sets through this advanced model, we can decide which videos should be accepted or rejected for the further actions.</span>
APA, Harvard, Vancouver, ISO, and other styles
13

Adam, Nabil R., Aryya Gangopadhyay, and James Clifford. "A Form-Based Approach to Natural Language Query Processing." Journal of Management Information Systems 11, no. 2 (1994): 109–35. http://dx.doi.org/10.1080/07421222.1994.11518042.

Full text
APA, Harvard, Vancouver, ISO, and other styles
14

Mati, Diellza Nagavci, Mentor Hamiti, Arsim Susuri, Besnik Selimi, and Jaumin Ajdari. "Building Dictionaries for Low Resource Languages: Challenges of Unsupervised Learning." Annals of Emerging Technologies in Computing 5, no. 3 (2021): 52–58. http://dx.doi.org/10.33166/aetic.2021.03.005.

Full text
Abstract:
The development of natural language processing resources for Albanian has grown steadily in recent years. This paper presents research conducted on unsupervised learning-the challenges associated with building a dictionary for the Albanian language and creating part-of-speech tagging models. The majority of languages have their own dictionary, but languages with low resources suffer from a lack of resources. It facilitates the sharing of information and services for users and whole communities through natural language processing. The experimentation corpora for the Albanian language includes 250K sentences from different disciplines, with a proposal for a part-of-speech tagging tag set that can adequately represent the underlying linguistic phenomena. Contributing to the development of Albanian is the purpose of this paper. The results of experiments with the Albanian language corpus revealed that its use of articles and pronouns resembles that of more high-resource languages. According to this study, the total expected frequency as a means for correctly tagging words has been proven effective for populating the Albanian language dictionary.
APA, Harvard, Vancouver, ISO, and other styles
15

Nguyen, Thi Tuyet Hai, Adam Jatowt, Mickael Coustaty, and Antoine Doucet. "Survey of Post-OCR Processing Approaches." ACM Computing Surveys 54, no. 6 (2021): 1–37. http://dx.doi.org/10.1145/3453476.

Full text
Abstract:
Optical character recognition (OCR) is one of the most popular techniques used for converting printed documents into machine-readable ones. While OCR engines can do well with modern text, their performance is unfortunately significantly reduced on historical materials. Additionally, many texts have already been processed by various out-of-date digitisation techniques. As a consequence, digitised texts are noisy and need to be post-corrected. This article clarifies the importance of enhancing quality of OCR results by studying their effects on information retrieval and natural language processing applications. We then define the post-OCR processing problem, illustrate its typical pipeline, and review the state-of-the-art post-OCR processing approaches. Evaluation metrics, accessible datasets, language resources, and useful toolkits are also reported. Furthermore, the work identifies the current trend and outlines some research directions of this field.
APA, Harvard, Vancouver, ISO, and other styles
16

Ayoub, Jackie, X. Jessie Yang, and Feng Zhou. "Combat COVID-19 infodemic using explainable natural language processing models." Information Processing & Management 58, no. 4 (2021): 102569. http://dx.doi.org/10.1016/j.ipm.2021.102569.

Full text
APA, Harvard, Vancouver, ISO, and other styles
17

Chen, Jieh-Haur, Mu-Chun Su, Vidya Trisandini Azzizi, Ting-Kwei Wang, and Wei-Jen Lin. "Smart Project Management: Interactive Platform Using Natural Language Processing Technology." Applied Sciences 11, no. 4 (2021): 1597. http://dx.doi.org/10.3390/app11041597.

Full text
Abstract:
Technological developments have made the construction industry efficient. The aim of this research is to solve communication interaction problems to build a project management platform using the interactive concept of natural language processing technology. A comprehensive literature review and expert interviews associated with techniques dealing with natural languages suggests the proposed system containing the Progressive Scale Expansion Network (PSENet), Convolutional Recurrent Neural Network (CRNN), and Bi-directional Recurrent Neutral Networks Convolutional Recurrent Neural Network (BRNN-CNN) toolboxes to extract the key words for construction projects contracts. The results show that a fully automatic platform facilitating contract management is achieved. For academic domains, the Contract Keyword Detection (CKD) mechanism integrating PSENet, CRNN, and BRNN-CNN approaches to cope with real-time massive document flows is novel in the construction industry. For practice, the proposed approach brings significant reduction for manpower and human error, an alternative for settling down misunderstanding or disputes due to real-time and precise communication, and a solution for efficient documentary management. It connects all contract stakeholders proficiently.
APA, Harvard, Vancouver, ISO, and other styles
18

Tafazoli, Dara, Elena Gómez María, and Cristina A. Huertas Abril. "Intelligent Language Tutoring System." International Journal of Information and Communication Technology Education 15, no. 3 (2019): 60–74. http://dx.doi.org/10.4018/ijicte.2019070105.

Full text
Abstract:
Intelligent computer-assisted language learning (ICALL) is a multidisciplinary area of research that combines natural language processing (NLP), intelligent tutoring system (ITS), second language acquisition (SLA), and foreign language teaching and learning (FLTL). Intelligent tutoring systems (ITS) are able to provide a personalized approach to learning by assuming the role of a real teacher/expert who adapts and steers the learning process according to the specific needs of each learner. This article reviews and discusses the issues surrounding the development and use of ITSs for language learning and teaching. First, the authors look at ICALL history: its evolution from CALL. Second, issues in ICALL research and integration will be discussed. Third, they will explain how artificial intelligence (AI) techniques are being implemented in language education as ITS and intelligent language tutoring systems (ITLS). Finally, the successful integration and development of ITLS will be explained in detail.
APA, Harvard, Vancouver, ISO, and other styles
19

Rista, Amarildo, and Arbana Kadriu. "Automatic Speech Recognition: A Comprehensive Survey." SEEU Review 15, no. 2 (2020): 86–112. http://dx.doi.org/10.2478/seeur-2020-0019.

Full text
Abstract:
Abstract Speech recognition is an interdisciplinary subfield of natural language processing (NLP) that facilitates the recognition and translation of spoken language into text by machine. Speech recognition plays an important role in digital transformation. It is widely used in different areas such as education, industry, and healthcare and has recently been used in many Internet of Things and Machine Learning applications. The process of speech recognition is one of the most difficult processes in computer science. Despite numerous searches in this domain, an optimal method for speech recognition has not yet been found. This is due to the fact that there are many attributes that characterize natural languages and every language has its particular highlights. The aim of this research is to provide a comprehensive understanding of the various techniques within the domain of Speech Recognition through a systematic literature review of the existing work. We will introduce the most significant and relevant techniques that may provide some directions in the future research.
APA, Harvard, Vancouver, ISO, and other styles
20

Chao, Min-Hua, Amy J. C. Trappey, and Chun-Ting Wu. "Emerging Technologies of Natural Language-Enabled Chatbots: A Review and Trend Forecast Using Intelligent Ontology Extraction and Patent Analytics." Complexity 2021 (May 24, 2021): 1–26. http://dx.doi.org/10.1155/2021/5511866.

Full text
Abstract:
Natural language processing (NLP) is a critical part of the digital transformation. NLP enables user-friendly interactions between machine and human by making computers understand human languages. Intelligent chatbot is an essential application of NLP to allow understanding of users’ utterance and responding in understandable sentences for specific applications simulating human-to-human conversations and interactions for problem solving or Q&As. This research studies emerging technologies for NLP-enabled intelligent chatbot development using a systematic patent analytic approach. Some intelligent text-mining techniques are applied, including document term frequency analysis for key terminology extractions, clustering method for identifying the subdomains, and Latent Dirichlet Allocation for finding the key topics of patent set. This research utilizes the Derwent Innovation database as the main source for global intelligent chatbot patent retrievals.
APA, Harvard, Vancouver, ISO, and other styles
21

Nikiforos, Maria Nefeli, Yorghos Voutos, Anthi Drougani, Phivos Mylonas, and Katia Lida Kermanidis. "The Modern Greek Language on the Social Web: A Survey of Data Sets and Mining Applications." Data 6, no. 5 (2021): 52. http://dx.doi.org/10.3390/data6050052.

Full text
Abstract:
Mining social web text has been at the heart of the Natural Language Processing and Data Mining research community in the last 15 years. Though most of the reported work is on widely spoken languages, such as English, the significance of approaches that deal with less commonly spoken languages, such as Greek, is evident for reasons of preserving and documenting minority languages, cultural and ethnic diversity, and identifying intercultural similarities and differences. The present work aims at identifying, documenting and comparing social text data sets, as well as mining techniques and applications on social web text that target Modern Greek, focusing on the arising challenges and the potential for future research in the specific less widely spoken language.
APA, Harvard, Vancouver, ISO, and other styles
22

Yang, Li, Yuhui Wang, and Haixia Li. "Research on the Disease Intelligent Diagnosis Model Based on Linguistic Truth-Valued Concept Lattice." Complexity 2021 (May 13, 2021): 1–11. http://dx.doi.org/10.1155/2021/6630077.

Full text
Abstract:
Uncertainty natural language processing has always been a research focus in the artificial intelligence field. In this paper, we continue to study the linguistic truth-valued concept lattice and apply it to the disease intelligent diagnosis by building an intelligent model to directly handle natural language. The theoretical bases of this model are the classical concept lattice and the lattice implication algebra with natural language. The model includes the case library formed by patients, attributes matching, and the matching degree calculation about the new patient. According to the characteristics of the patients, the disease attributes are firstly divided into intrinsic invariant attributes and extrinsic variable attributes. The calculation algorithm of the linguistic truth-valued formal concepts and the constructing algorithm of the linguistic truth-valued concept lattice based on the extrinsic attributes are proposed. And the disease bases of the different treatments for different patients with the same disease are established. Secondly, the matching algorithms of intrinsic attributes and extrinsic attributes are given, and all the linguistic truth-valued formal concepts that match the new patient’s extrinsic attributes are found. Lastly, by comparing the similarity between the new patients and the matching formal concepts, we calculate the best treatment options to realize the intelligent diagnosis of the disease.
APA, Harvard, Vancouver, ISO, and other styles
23

Lin, Nankai, Boyu Chen, Xiaotian Lin, Kanoksak Wattanachote, and Shengyi Jiang. "A Framework for Indonesian Grammar Error Correction." ACM Transactions on Asian and Low-Resource Language Information Processing 20, no. 4 (2021): 1–12. http://dx.doi.org/10.1145/3440993.

Full text
Abstract:
Grammatical Error Correction (GEC) is a challenge in Natural Language Processing research. Although many researchers have been focusing on GEC in universal languages such as English or Chinese, few studies focus on Indonesian, which is a low-resource language. In this article, we proposed a GEC framework that has the potential to be a baseline method for Indonesian GEC tasks. This framework treats GEC as a multi-classification task. It integrates different language embedding models and deep learning models to correct 10 types of Part of Speech (POS) error in Indonesian text. In addition, we constructed an Indonesian corpus that can be utilized as an evaluation dataset for Indonesian GEC research. Our framework was evaluated on this dataset. Results showed that the Long Short-Term Memory model based on word-embedding achieved the best performance. Its overall macro-average F 0.5 in correcting 10 POS error types reached 0.551. Results also showed that the framework can be trained on a low-resource dataset.
APA, Harvard, Vancouver, ISO, and other styles
24

Reiter, Ehud, and Somayajulu Sripada. "Human Variation and Lexical Choice." Computational Linguistics 28, no. 4 (2002): 545–53. http://dx.doi.org/10.1162/089120102762671981.

Full text
Abstract:
Much natural language processing research implicitly assumes that word meanings are fixed in a language community, but in fact there is good evidence that different people probably associate slightly different meanings with words. We summarize some evidence for this claim from the literature and from an ongoing research project, and discuss its implications for natural language generation, especially for lexical choice, that is, choosing appropriate words for a generated text.
APA, Harvard, Vancouver, ISO, and other styles
25

Chandrasekaran, Dhivya, and Vijay Mago. "Evolution of Semantic Similarity—A Survey." ACM Computing Surveys 54, no. 2 (2021): 1–37. http://dx.doi.org/10.1145/3440755.

Full text
Abstract:
Estimating the semantic similarity between text data is one of the challenging and open research problems in the field of Natural Language Processing (NLP). The versatility of natural language makes it difficult to define rule-based methods for determining semantic similarity measures. To address this issue, various semantic similarity methods have been proposed over the years. This survey article traces the evolution of such methods beginning from traditional NLP techniques such as kernel-based methods to the most recent research work on transformer-based models, categorizing them based on their underlying principles as knowledge-based, corpus-based, deep neural network–based methods, and hybrid methods. Discussing the strengths and weaknesses of each method, this survey provides a comprehensive view of existing systems in place for new researchers to experiment and develop innovative ideas to address the issue of semantic similarity.
APA, Harvard, Vancouver, ISO, and other styles
26

Shivachi, Casper Shikali, Refuoe Mokhosi, Zhou Shijie, and Liu Qihe. "Learning Syllables Using Conv-LSTM Model for Swahili Word Representation and Part-of-speech Tagging." ACM Transactions on Asian and Low-Resource Language Information Processing 20, no. 4 (2021): 1–25. http://dx.doi.org/10.1145/3445975.

Full text
Abstract:
The need to capture intra-word information in natural language processing (NLP) tasks has inspired research in learning various word representations at word, character, or morpheme levels, but little attention has been given to syllables from a syllabic alphabet. Motivated by the success of compositional models in morphological languages, we present a Convolutional-long short term memory (Conv-LSTM) model for constructing Swahili word representation vectors from syllables. The unified architecture addresses the word agglutination and polysemous nature of Swahili by extracting high-level syllable features using a convolutional neural network (CNN) and then composes quality word embeddings with a long short term memory (LSTM). The word embeddings are then validated using a syllable-aware language model ( 31.267 ) and a part-of-speech (POS) tagging task ( 98.78 ), both yielding very competitive results to the state-of-art models in their respective domains. We further validate the language model using Xhosa and Shona, which are syllabic-based languages. The novelty of the study is in its capability to construct quality word embeddings from syllables using a hybrid model that does not use max-over-pool common in CNN and then the exploitation of these embeddings in POS tagging. Therefore, the study plays a crucial role in the processing of agglutinative and syllabic-based languages by contributing quality word embeddings from syllable embeddings, a robust Conv–LSTM model that learns syllables for not only language modeling and POS tagging, but also for other downstream NLP tasks.
APA, Harvard, Vancouver, ISO, and other styles
27

Mahankali, Ranjeeth, Brian R. Johnson, and Alex T. Anderson. "Deep learning in design workflows: The elusive design pixel." International Journal of Architectural Computing 16, no. 4 (2018): 328–40. http://dx.doi.org/10.1177/1478077118800888.

Full text
Abstract:
The recent wave of developments and research in the field of deep learning and artificial intelligence is causing the border between the intuitive and deterministic domains to be redrawn, especially in computer vision and natural language processing. As designers frequently invoke vision and language in the context of design, this article takes a step back to ask if deep learning’s capabilities might be applied to design workflows, especially in architecture. In addition to addressing this general question, the article discusses one of several prototypes, BIMToVec, developed to examine the use of deep learning in design. It employs techniques like those used in natural language processing to interpret building information models. The article also proposes a homogeneous data format, provisionally called a design pixel, which can store design information as spatial-semantic maps. This would make designers’ intuitive thoughts more accessible to deep learning algorithms while also allowing designers to communicate abstractly with design software.
APA, Harvard, Vancouver, ISO, and other styles
28

Szakács, Béla Benedek, and Tamás Mészáros. "Hybrid Distance-based, CNN and Bi-LSTM System for Dictionary Expansion." Infocommunications journal, no. 4 (2020): 6–13. http://dx.doi.org/10.36244/icj.2020.4.2.

Full text
Abstract:
Dictionaries like Wordnet can help in a variety of Natural Language Processing applications by providing additional morphological data. They can be used in Digital Humanities research, building knowledge graphs and other applications. Creating dictionaries from large corpora of texts written in a natural language is a task that has not been a primary focus of research, as other tasks have dominated the field (such as chat-bots), but it can be a very useful tool in analysing texts. Even in the case of contemporary texts, categorizing the words according to their dictionary entry is a complex task, and for less conventional texts (in old or less researched languages) it is even harder to solve this problem automatically. Our task was to create a software that helps in expanding a dictionary containing word forms and tagging unprocessed text. We used a manually created corpus for training and testing the model. We created a combination of Bidirectional Long-Short Term Memory networks, convolutional networks and a distancebased solution that outperformed other existing solutions. While manual post-processing for the tagged text is still needed, it significantly reduces the amount of it.
APA, Harvard, Vancouver, ISO, and other styles
29

Santana Suárez, Octavio, Francisco J. Carreras Riudavets, Zenón Hernández Figueroa, and Antonio C. González Cabrera. "Integration of an XML electronic dictionary with linguistic tools for natural language processing." Information Processing & Management 43, no. 4 (2007): 946–57. http://dx.doi.org/10.1016/j.ipm.2006.08.005.

Full text
APA, Harvard, Vancouver, ISO, and other styles
30

Ram Kinkar, Chhayarani, and Yogendra Kumar Jain. "An optimal method for enhancing the generation of machine code from natural language data set." International Journal of Engineering & Technology 8, no. 4 (2019): 590. http://dx.doi.org/10.14419/ijet.v8i4.26593.

Full text
Abstract:
Natural language processing is a very active area of research and development, there is not a single agreed upon a method that would satisfy everyone for the use of natural language to operate electronic devices or other practical applications. But there are some aspects used from many years in the formulation and solution of computational problem arising in natural language processing. This paper describes a model in which numerical values are assigned to word of natural language speech data set to convert the information present in natural language speech data set into an intermediate numeric form as a structured data set. The intermediated numerical values of each word will be used for generation of machine code which will be easily understand by electronic devices to draw inferences from data set. The designed model is useful for a number of practical applications and very simple to implement.
APA, Harvard, Vancouver, ISO, and other styles
31

Hsu, Chun-Nan, Anita E. Bandrowski, Thomas H. Gillespie, et al. "Comparing the Use of Research Resource Identifiers and Natural Language Processing for Citation of Databases, Software, and Other Digital Artifacts." Computing in Science & Engineering 22, no. 2 (2020): 22–32. http://dx.doi.org/10.1109/mcse.2019.2952838.

Full text
APA, Harvard, Vancouver, ISO, and other styles
32

Sprugnoli, Rachele, and Sara Tonelli. "Novel Event Detection and Classification for Historical Texts." Computational Linguistics 45, no. 2 (2019): 229–65. http://dx.doi.org/10.1162/coli_a_00347.

Full text
Abstract:
Event processing is an active area of research in the Natural Language Processing community, but resources and automatic systems developed so far have mainly addressed contemporary texts. However, the recognition and elaboration of events is a crucial step when dealing with historical texts Particularly in the current era of massive digitization of historical sources: Research in this domain can lead to the development of methodologies and tools that can assist historians in enhancing their work, while having an impact also on the field of Natural Language Processing. Our work aims at shedding light on the complex concept of events when dealing with historical texts. More specifically, we introduce new annotation guidelines for event mentions and types, categorized into 22 classes. Then, we annotate a historical corpus accordingly, and compare two approaches for automatic event detection and classification following this novel scheme. We believe that this work can foster research in a field of inquiry as yet underestimated in the area of Temporal Information Processing. To this end, we release new annotation guidelines, a corpus, and new models for automatic annotation.
APA, Harvard, Vancouver, ISO, and other styles
33

Arnarsson, Ivar Örn, Otto Frost, Emil Gustavsson, Mats Jirstrand, and Johan Malmqvist. "Natural language processing methods for knowledge management—Applying document clustering for fast search and grouping of engineering documents." Concurrent Engineering 29, no. 2 (2021): 142–52. http://dx.doi.org/10.1177/1063293x20982973.

Full text
Abstract:
Product development companies collect data in form of Engineering Change Requests for logged design issues, tests, and product iterations. These documents are rich in unstructured data (e.g. free text). Previous research affirms that product developers find that current IT systems lack capabilities to accurately retrieve relevant documents with unstructured data. In this research, we demonstrate a method using Natural Language Processing and document clustering algorithms to find structurally or contextually related documents from databases containing Engineering Change Request documents. The aim is to radically decrease the time needed to effectively search for related engineering documents, organize search results, and create labeled clusters from these documents by utilizing Natural Language Processing algorithms. A domain knowledge expert at the case company evaluated the results and confirmed that the algorithms we applied managed to find relevant document clusters given the queries tested.
APA, Harvard, Vancouver, ISO, and other styles
34

Palade, Vasile, and J. Gerard Wolff. "A Roadmap for the Development of the ‘SP Machine’ for Artificial Intelligence." Computer Journal 62, no. 11 (2019): 1584–604. http://dx.doi.org/10.1093/comjnl/bxy126.

Full text
Abstract:
AbstractThis paper describes a roadmap for the development of the SP Machine, based on the SP Theory of Intelligence and its realization in the SP Computer Model. The SP Machine will be developed initially as a software virtual machine with high levels of parallel processing, hosted on a high-performance computer. The system should help users visualize knowledge structures and processing. Research is needed into how the system may discover low-level features in speech and in images. Strengths of the SP System in the processing of natural language may be augmented, in conjunction with the further development of the SP System’s strengths in unsupervised learning. Strengths of the SP System in pattern recognition may be developed for computer vision. Work is needed on the representation of numbers and the performance of arithmetic processes. A computer model is needed of SP-Neural, the version of the SP Theory expressed in terms of neurons and their interconnections. The SP Machine has potential in many areas of application, several of which may be realized on short-to-medium timescales.
APA, Harvard, Vancouver, ISO, and other styles
35

Losee, Robert M. "Natural language processing in support of decision-making: phrases and part-of-speech tagging." Information Processing & Management 37, no. 6 (2001): 769–87. http://dx.doi.org/10.1016/s0306-4573(00)00061-3.

Full text
APA, Harvard, Vancouver, ISO, and other styles
36

Jararweh, Yaser, Mahmoud Al-Ayyoub, and Elhadj Benkhelifa. "Advanced Arabic Natural Language Processing (ANLP) and its applications: Introduction to the special issue." Information Processing & Management 56, no. 2 (2019): 259–61. http://dx.doi.org/10.1016/j.ipm.2018.09.003.

Full text
APA, Harvard, Vancouver, ISO, and other styles
37

Mumcuoğlu, Emre, Ceyhun E. Öztürk, Haldun M. Ozaktas, and Aykut Koç. "Natural language processing in law: Prediction of outcomes in the higher courts of Turkey." Information Processing & Management 58, no. 5 (2021): 102684. http://dx.doi.org/10.1016/j.ipm.2021.102684.

Full text
APA, Harvard, Vancouver, ISO, and other styles
38

Sun, Yuan, Andong Chen, Chaofan Chen, Tianci Xia, and Xiaobing Zhao. "A Joint Model for Representation Learning of Tibetan Knowledge Graph Based on Encyclopedia." ACM Transactions on Asian and Low-Resource Language Information Processing 20, no. 2 (2021): 1–17. http://dx.doi.org/10.1145/3447248.

Full text
Abstract:
Learning the representation of a knowledge graph is critical to the field of natural language processing. There is a lot of research for English knowledge graph representation. However, for the low-resource languages, such as Tibetan, how to represent sparse knowledge graphs is a key problem. In this article, aiming at scarcity of Tibetan knowledge graphs, we extend the Tibetan knowledge graph by using the triples of the high-resource language knowledge graphs and Point of Information map information. To improve the representation learning of the Tibetan knowledge graph, we propose a joint model to merge structure and entity description information based on the Translating Embeddings and Convolution Neural Networks models. In addition, to solve the segmentation errors, we use character and word embedding to learn more complex information in Tibetan. Finally, the experimental results show that our model can make a better representation of the Tibetan knowledge graph than the baseline.
APA, Harvard, Vancouver, ISO, and other styles
39

Shutova, Ekaterina, Lin Sun, Elkin Darío Gutiérrez, Patricia Lichtenstein, and Srini Narayanan. "Multilingual Metaphor Processing: Experiments with Semi-Supervised and Unsupervised Learning." Computational Linguistics 43, no. 1 (2017): 71–123. http://dx.doi.org/10.1162/coli_a_00275.

Full text
Abstract:
Highly frequent in language and communication, metaphor represents a significant challenge for Natural Language Processing (NLP) applications. Computational work on metaphor has traditionally evolved around the use of hand-coded knowledge, making the systems hard to scale. Recent years have witnessed a rise in statistical approaches to metaphor processing. However, these approaches often require extensive human annotation effort and are predominantly evaluated within a limited domain. In contrast, we experiment with weakly supervised and unsupervised techniques—with little or no annotation—to generalize higher-level mechanisms of metaphor from distributional properties of concepts. We investigate different levels and types of supervision (learning from linguistic examples vs. learning from a given set of metaphorical mappings vs. learning without annotation) in flat and hierarchical, unconstrained and constrained clustering settings. Our aim is to identify the optimal type of supervision for a learning algorithm that discovers patterns of metaphorical association from text. In order to investigate the scalability and adaptability of our models, we applied them to data in three languages from different language groups—English, Spanish, and Russian—achieving state-of-the-art results with little supervision. Finally, we demonstrate that statistical methods can facilitate and scale up cross-linguistic research on metaphor.
APA, Harvard, Vancouver, ISO, and other styles
40

Kim, Meen Chul, Seojin Nam, Fei Wang, and Yongjun Zhu. "Mapping scientific landscapes in UMLS research: a scientometric review." Journal of the American Medical Informatics Association 27, no. 10 (2020): 1612–24. http://dx.doi.org/10.1093/jamia/ocaa107.

Full text
Abstract:
Abstract Objective The Unified Medical Language System (UMLS) is 1 of the most successful, collaborative efforts of terminology resource development in biomedicine. The present study aims to 1) survey historical footprints, emerging technologies, and the existing challenges in the use of UMLS resources and tools, and 2) present potential future directions. Materials and Methods We collected 10 469 bibliographic records published between 1986 and 2019, using a Web of Science database. graph analysis, data visualization, and text mining to analyze domain-level citations, subject categories, keyword co-occurrence and bursts, document co-citation networks, and landmark papers. Results The findings show that the development of UMLS resources and tools have been led by interdisciplinary collaboration among medicine, biology, and computer science. Efforts encompassing multiple disciplines, such as medical informatics, biochemical sciences, and genetics, were the driving forces behind the domain’s growth. The following topics were found to be the dominant research themes from the early phases to mid-phases: 1) development and extension of ontologies and 2) enhancing the integrity and accessibility of these resources. Knowledge discovery using machine learning and natural language processing and applications in broader contexts such as drug safety surveillance have recently been receiving increasing attention. Discussion Our analysis confirms that while reaching its scientific maturity, UMLS research aims to boundary-span to more variety in the biomedical context. We also made some recommendations for editorship and authorship in the domain. Conclusion The present study provides a systematic approach to map the intellectual growth of science, as well as a self-explanatory bibliometric profile of the published UMLS literature. It also suggests potential future directions. Using the findings of this study, the scientific community can better align the studies within the emerging agenda and current challenges.
APA, Harvard, Vancouver, ISO, and other styles
41

Cao, Zehong. "A review of artificial intelligence for EEG‐based brain−computer interfaces and applications." Brain Science Advances 6, no. 3 (2020): 162–70. http://dx.doi.org/10.26599/bsa.2020.9050017.

Full text
Abstract:
The advancement in neuroscience and computer science promotes the ability of the human brain to communicate and interact with the environment, making brain–computer interface (BCI) top interdisciplinary research. Furthermore, with the modern technology advancement in artificial intelligence (AI), including machine learning (ML) and deep learning (DL) methods, there is vast growing interest in the electroencephalogram (EEG)‐based BCIs for AI‐related visual, literal, and motion applications. In this review study, the literature on mainstreams of AI for the EEG‐based BCI applications is investigated to fill gaps in the interdisciplinary BCI field. Specifically, the EEG signals and their main applications in BCI are first briefly introduced. Next, the latest AI technologies, including the ML and DL models, are presented to monitor and feedback human cognitive states. Finally, some BCI‐inspired AI applications, including computer vision, natural language processing, and robotic control applications, are presented. The future research directions of the EEG‐based BCI are highlighted in line with the AI technologies and applications.
APA, Harvard, Vancouver, ISO, and other styles
42

Vedula, Nikhita. "Modeling knowledge and functional intent for context-aware pragmatic analysis." ACM SIGWEB Newsletter, Winter (January 2021): 1–4. http://dx.doi.org/10.1145/3447879.3447882.

Full text
Abstract:
Nikhita Vedula is an Applied Scientist at Amazon Alexa Science. She obtained her PhD in Computer Science and Engineering from the Ohio State University in August 2020, advised by Professor Srinivasan Parthasarathy. She received her bachelor's degree from the National Institute of Technology, Nagpur, India in 2015. Her research interests are at the intersection of data mining, natural language processing and social computing. Over the course of her PhD, her research involved designing efficient and novel machine learning and computational linguistic techniques that extract, interpret and transform the vast, unstructured digital content into structured knowledge representations in diverse contexts. She has worked with researchers from interdisciplinary fields such as emergency response, marketing, sociology and psychology. She performed research internships at Nokia Bell Laboratories, Adobe Research and Amazon Alexa AI. Her work has been published at several top data mining conferences such as the Web Conference, SIGIR, WSDM and ICDM. Her work on detecting user intentions from their natural language interactions won the Best paper award at the Web Conference 2020. She was a recipient of a Graduate Research Award (2020), a Presidential Fellowship (2019) and a University Graduate Fellowship (2015) at the Ohio State University. She was also selected as a Rising Star in EECS (2019).
APA, Harvard, Vancouver, ISO, and other styles
43

Shi, Shumin, Dan Luo, Xing Wu, Congjun Long, and Heyan Huang. "Multi-level Chunk-based Constituent-to-Dependency Treebank Transformation for Tibetan Dependency Parsing." ACM Transactions on Asian and Low-Resource Language Information Processing 20, no. 2 (2021): 1–12. http://dx.doi.org/10.1145/3424247.

Full text
Abstract:
Dependency parsing is an important task for Natural Language Processing (NLP). However, a mature parser requires a large treebank for training, which is still extremely costly to create. Tibetan is a kind of extremely low-resource language for NLP, there is no available Tibetan dependency treebank, which is currently obtained by manual annotation. Furthermore, there are few related kinds of research on the construction of treebank. We propose a novel method of multi-level chunk-based syntactic parsing to complete constituent-to-dependency treebank conversion for Tibetan under scarce conditions. Our method mines more dependencies of Tibetan sentences, builds a high-quality Tibetan dependency tree corpus, and makes fuller use of the inherent laws of the language itself. We train the dependency parsing models on the dependency treebank obtained by the preliminary transformation. The model achieves 86.5% accuracy, 96% LAS, and 97.85% UAS, which exceeds the optimal results of existing conversion methods. The experimental results show that our method has the potential to use a low-resource setting, which means we not only solve the problem of scarce Tibetan dependency treebank but also avoid needless manual annotation. The method embodies the regularity of strong knowledge-guided linguistic analysis methods, which is of great significance to promote the research of Tibetan information processing.
APA, Harvard, Vancouver, ISO, and other styles
44

Loucky, John Paul, and Frank Tuzi. "Comparing Foreign Language Learners’ Use of Online Glossing Programs." International Journal of Virtual and Personal Learning Environments 1, no. 4 (2010): 31–51. http://dx.doi.org/10.4018/jvple.2010100103.

Full text
Abstract:
This study furthers research in three crucial related areas: 1) comparing various online glossing and vocabulary learning tools; 2) language teaching and learning using a more natural bilingualized approach to developing online reading skills in a second or foreign language; and 3) comparing the relative level of enjoyment and effectiveness students experience when using various CALL programs. This paper applies recent insights into vocabulary learning behaviors and functions online and investigates whether teachers can help learners increase their use of online glosses to improve their vocabulary learning by giving them automatic mouse-over instant glosses versus optional, clickable, mechanical access. The authors compare Japanese college students’ actual use of three types of glossing when reading similar texts online. The findings suggest that an expanded glossing system that helps encourage deeper lexical processing by providing automatic, archivable glosses would be superior for digital vocabulary learning because it can simultaneously offer better monitoring and more motivation vis-à-vis online word learning.
APA, Harvard, Vancouver, ISO, and other styles
45

Sarma, Jhumpa. "Role of Artificial Intelligence in Medicine and Clinical Research." International Journal for Research in Applied Science and Engineering Technology 9, no. 8 (2021): 1512–18. http://dx.doi.org/10.22214/ijraset.2021.37617.

Full text
Abstract:
Abstract: Artificial Intelligence is a branch of computer science that enables to analyse complex medical data. The proficiency of artificial intelligence techniques has been explored to a great extent in the field of medicine. Most of the medications go to the business sector after a long tedious process of drug development. It can take a period of 10-15 years or more to convey a medication from its introductory revelation to the hands of the patients. Artificial Intelligence can significantly reduce the time required and can also cut down the expenses by half. Among the methods, artificial neural network is the most widely used analytical tool while other techniques like fuzzy expert systems, natural language processing, robotic process automation and evolutionary computation have been used in different clinical settings. The aim of this paper is to discuss the different artificial intelligence techniques and provide a perspective on the benefits, future opportunities and risks of established artificial intelligence applications in clinical practice on medical education, physicians, healthcare institutions and bioethics. Keywords: Artificial intelligence, clinical trials, medical technologies, artificial neural networks, diagnosis.
APA, Harvard, Vancouver, ISO, and other styles
46

Jang, Hyejin, and Byungun Yoon. "TechWordNet: Development of semantic relation for technology information analysis using F-term and natural language processing." Information Processing & Management 58, no. 6 (2021): 102752. http://dx.doi.org/10.1016/j.ipm.2021.102752.

Full text
APA, Harvard, Vancouver, ISO, and other styles
47

Almela, Ángela. "A Corpus-Based Study of Linguistic Deception in Spanish." Applied Sciences 11, no. 19 (2021): 8817. http://dx.doi.org/10.3390/app11198817.

Full text
Abstract:
In the last decade, fields such as psychology and natural language processing have devoted considerable attention to the automatization of the process of deception detection, developing and employing a wide array of automated and computer-assisted methods for this purpose. Similarly, another emerging research area is focusing on computer-assisted deception detection using linguistics, with promising results. Accordingly, in the present article, the reader is firstly provided with an overall review of the state of the art of corpus-based research exploring linguistic cues to deception as well as an overview on several approaches to the study of deception and on previous research into its linguistic detection. In an effort to promote corpus-based research in this context, this study explores linguistic cues to deception in the Spanish written language with the aid of an automatic text classification tool, by means of an ad hoc corpus containing ground truth data. Interestingly, the key findings reveal that, although there is a set of linguistic cues which contributes to the global statistical classification model, there are some discursive differences across the subcorpora, yielding better classification results on the analysis conducted on the subcorpus containing emotionally loaded language.
APA, Harvard, Vancouver, ISO, and other styles
48

Wang, Jing, Huan Deng, Bangtao Liu, et al. "Systematic Evaluation of Research Progress on Natural Language Processing in Medicine Over the Past 20 Years: Bibliometric Study on PubMed." Journal of Medical Internet Research 22, no. 1 (2020): e16816. http://dx.doi.org/10.2196/16816.

Full text
Abstract:
Background Natural language processing (NLP) is an important traditional field in computer science, but its application in medical research has faced many challenges. With the extensive digitalization of medical information globally and increasing importance of understanding and mining big data in the medical field, NLP is becoming more crucial. Objective The goal of the research was to perform a systematic review on the use of NLP in medical research with the aim of understanding the global progress on NLP research outcomes, content, methods, and study groups involved. Methods A systematic review was conducted using the PubMed database as a search platform. All published studies on the application of NLP in medicine (except biomedicine) during the 20 years between 1999 and 2018 were retrieved. The data obtained from these published studies were cleaned and structured. Excel (Microsoft Corp) and VOSviewer (Nees Jan van Eck and Ludo Waltman) were used to perform bibliometric analysis of publication trends, author orders, countries, institutions, collaboration relationships, research hot spots, diseases studied, and research methods. Results A total of 3498 articles were obtained during initial screening, and 2336 articles were found to meet the study criteria after manual screening. The number of publications increased every year, with a significant growth after 2012 (number of publications ranged from 148 to a maximum of 302 annually). The United States has occupied the leading position since the inception of the field, with the largest number of articles published. The United States contributed to 63.01% (1472/2336) of all publications, followed by France (5.44%, 127/2336) and the United Kingdom (3.51%, 82/2336). The author with the largest number of articles published was Hongfang Liu (70), while Stéphane Meystre (17) and Hua Xu (33) published the largest number of articles as the first and corresponding authors. Among the first author’s affiliation institution, Columbia University published the largest number of articles, accounting for 4.54% (106/2336) of the total. Specifically, approximately one-fifth (17.68%, 413/2336) of the articles involved research on specific diseases, and the subject areas primarily focused on mental illness (16.46%, 68/413), breast cancer (5.81%, 24/413), and pneumonia (4.12%, 17/413). Conclusions NLP is in a period of robust development in the medical field, with an average of approximately 100 publications annually. Electronic medical records were the most used research materials, but social media such as Twitter have become important research materials since 2015. Cancer (24.94%, 103/413) was the most common subject area in NLP-assisted medical research on diseases, with breast cancers (23.30%, 24/103) and lung cancers (14.56%, 15/103) accounting for the highest proportions of studies. Columbia University and the talents trained therein were the most active and prolific research forces on NLP in the medical field.
APA, Harvard, Vancouver, ISO, and other styles
49

Nishida, Toyoaki. "Augmenting Conversational Environment." International Journal of Cognitive Informatics and Natural Intelligence 6, no. 4 (2012): 103–24. http://dx.doi.org/10.4018/jcini.2012100105.

Full text
Abstract:
People are proficient in collaboratively forming and maintaining gatherings thereby shaping and cultivating collective thoughts through fluent conversational interactions. A big challenge is to develop a technology for augmenting the conversational environment so that people can conduct even better conversational interactions for collective intelligence and creation. Conversational informatics is a field of research that focuses on investigating conversational interactions and designing intelligent artifacts that can augment conversational interactions. The field draws on a foundation provided by artificial intelligence, natural language processing, speech and image processing, cognitive science, and conversation analysis. In this article, the author overviews a methodology for developing augmented conversational environment and major achievements. The author also discusses issues for making agents empathic so that they can induce sustained and constructive engagement with people.
APA, Harvard, Vancouver, ISO, and other styles
50

Ruiz-Dolz, Ramon, Montserrat Nofre, Mariona Taulé, Stella Heras, and Ana García-Fornes. "VivesDebate: A New Annotated Multilingual Corpus of Argumentation in a Debate Tournament." Applied Sciences 11, no. 15 (2021): 7160. http://dx.doi.org/10.3390/app11157160.

Full text
Abstract:
The application of the latest Natural Language Processing breakthroughs in computational argumentation has shown promising results, which have raised the interest in this area of research. However, the available corpora with argumentative annotations are often limited to a very specific purpose or are not of adequate size to take advantage of state-of-the-art deep learning techniques (e.g., deep neural networks). In this paper, we present VivesDebate, a large, richly annotated and versatile professional debate corpus for computational argumentation research. The corpus has been created from 29 transcripts of a debate tournament in Catalan and has been machine-translated into Spanish and English. The annotation contains argumentative propositions, argumentative relations, debate interactions and professional evaluations of the arguments and argumentation. The presented corpus can be useful for research on a heterogeneous set of computational argumentation underlying tasks such as Argument Mining, Argument Analysis, Argument Evaluation or Argument Generation, among others. All this makes VivesDebate a valuable resource for computational argumentation research within the context of massive corpora aimed at Natural Language Processing tasks.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography