To see the other types of publications on this topic, follow the link: Working with a text document.

Journal articles on the topic 'Working with a text document'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 journal articles for your research on the topic 'Working with a text document.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse journal articles on a wide variety of disciplines and organise your bibliography correctly.

1

Patil, Harsha, and Ramjeevan Singh Thakur. "A semantic approach for text document clustering using frequent itemsets and WordNet." International Journal of Engineering & Technology 7, no. 2.18 (2018): 102. http://dx.doi.org/10.14419/ijet.v7i2.9.10220.

Full text
Abstract:
Document Clustering is an unsupervised method for classified documents in clusters on the basis of their similarity. Any document get it place in any specific cluster, on the basis of membership score, which calculated through membership function. But many of the traditional clustering algorithms are generally based on only BOW (Bag of Words), which ignores the semantic similarity between document and Cluster. In this research we consider the semantic association between cluster and text document during the calculation of membership score of any document for any specific cluster. Several researchers are working on semantic aspects of document clustering to develop clustering performance. Many external knowledge bases like WordNet, Wikipedia, Lucene etc. are utilized for this purpose. The proposed approach exploits WordNet to improve cluster member ship function. The experimental result shows that clustering quality improved significantly by using proposed framework of semantic approach.
APA, Harvard, Vancouver, ISO, and other styles
2

Ülker, Mehtap, and A. Bedri Özer. "The Bart-based Model for Scientific Articles Summarization." JUCS - Journal of Universal Computer Science 30, no. (13) (2024): 1807–28. https://doi.org/10.3897/jucs.115121.

Full text
Abstract:
With the development of deep learning techniques, many models have been proposed for abstractive text summarization. However, the problem of summarizing source documents while preserving their integrity persists due to token restrictions and the inability to adequately extract semantic word relations between different sentences. To overcome this problem, a fine-tuning BART-based model was proposed, which generates a scientific summary by selecting important words contained in the text in the input document. The input text consists of terminology and keywords from the source document. The proposed model is based on the working principle of graph-based methods. Thus, the proposed model can summarize the source document with as few words as possible that are relevant to the content. The proposed model was compared with baseline models and the results of human evaluation. The experimental results demonstrate that the proposed model outperforms the baseline methods with a 37.60 ROUGE-L score.
APA, Harvard, Vancouver, ISO, and other styles
3

Zagorodnikov, Mikhail Viktorovich, and Andrey Anatolevich Mikhaylov. "Recovering Text Layer from PDF Documents with Complex Background." Proceedings of the Institute for System Programming of the RAS 36, no. 3 (2024): 189–202. http://dx.doi.org/10.15514/ispras-2024-36(3)-13.

Full text
Abstract:
The article considers PDF as a tool for storing and transferring documents. Special attention is paid to the problem of converting data from PDF back to its original format. The relevance of the study is due to the widespread use of PDF in electronic document management of modern organizations. However, despite the convenience of using PDF, extracting information from such documents can be difficult due to the peculiarities of information storage in the format and the lack of effective tools for reverse conversion. The paper proposes a solution based on the analysis of the text information from the output stream of the PDF format. This allows automatic recognition of text in PDF documents, even if they contain non-standard fonts, complex backgrounds, or damaged encoding. The research is of interest to specialists in the field of electronic document management, as well as software developers involved in creating tools for working with PDF.
APA, Harvard, Vancouver, ISO, and other styles
4

Hasan, Ismael, Javier Parapar, and Álvaro Barreiro. "Improving the Extraction of Text in PDFs by Simulating the Human Reading Order." JUCS - Journal of Universal Computer Science 18, no. (5) (2012): 623–49. https://doi.org/10.3217/jucs-018-05-0623.

Full text
Abstract:
Text preprocessing and segmentation are critical tasks in search and text mining applications. Due to the huge amount of documents that are exclusively presented in PDF format, most of the Data Mining (DM) and Information Retrieval (IR) systems must extract content from the PDF files. In some occasions this is a difficult task: the result of the extraction process from a PDF file is plain text, and it should be returned in the same order as a human would read the original PDF file. However, current tools for PDF text extraction fail in this objective when working with complex documents with multiple columns. For instance, this is the case of official government bulletins with legal information. In this task, it is mandatory to get correct and ordered text as a result of the application of the PDF extractor. It is very usual that a legal article in a document refers to a previous article and they should be offered in the right sequential order. To overcome these difficulties we have designed a new method for extraction of text in PDFs that simulates the human reading order. We evaluated our method and compared it against other PDF extraction tools and algorithms. Evaluation of our approach shows that it significantly outperforms the results of the existing tools and algorithms.
APA, Harvard, Vancouver, ISO, and other styles
5

Bakar, Abu, Raheem Sarwar, Saeed-Ul Hassan, and Raheel Nawaz. "Extracting Algorithmic Complexity in Scientific Literature for Advance Searching." Journal of Computational and Applied Linguistics 1 (July 18, 2023): 39–65. https://doi.org/10.33919/jcal.23.1.2.

Full text
Abstract:
Non-textual document elements such as charts, diagrams, algorithms and tables play an important role to present key information in scientific documents. Recent advances in information retrieval systems tap this information to answer more complex user queries by mining text pertaining to non-textual document elements from full text. Algorithms are critically important in computer science. Researchers are working on existing algorithms to improve them for critical application. Moreover, new algorithms for unsolved and newly faced problems are under development. These enhanced and new algorithms are mostly published in scholarly documents. The complexity of these algorithms is also discussed in the same document by the authors. Complexity of an algorithm is also an important factor for information retrieval (IR) systems. In this paper, we mine the relevant complexities of algorithms from full text document by comparing the metadata of the algorithm, such as caption and function name, with the context of the paragraph in which complexity related discussion is made by the authors. Using the dataset of 256 documents downloaded from CiteSeerX repository, we manually annotate 417 links between algorithms and their complexities. Further, we apply our novel rule-based approach that identifies the desired links with 81% precision, 75% recall, 78% F1-score and 65% accuracy. Overall, our method of identifying the links has potential to improve information retrieval systems that tap the advancements of full text and more specifically non-textual document elements.
APA, Harvard, Vancouver, ISO, and other styles
6

Ülker, Mehtap, and A. Bedri Özer. "The Bart-based Model for Scientific Articles Summarization." JUCS - Journal of Universal Computer Science 30, no. 13 (2024): 1807–28. https://doi.org/10.3897/jucs.115121.

Full text
Abstract:
With the development of deep learning techniques, many models have been proposed for abstractive text summarization. However, the problem of summarizing source documents while preserving their integrity persists due to token restrictions and the inability to adequately extract semantic word relations between different sentences. To overcome this problem, a fine-tuning BART-based model was proposed, which generates a scientific summary by selecting important words contained in the text in the input document. The input text consists of terminology and keywords from the source document. The proposed model is based on the working principle of graph-based methods. Thus, the proposed model can summarize the source document with as few words as possible that are relevant to the content. The proposed model was compared with baseline models and the results of human evaluation. The experimental results demonstrate that the proposed model outperforms the baseline methods with a 37.60 ROUGE-L score.
APA, Harvard, Vancouver, ISO, and other styles
7

Hahnel, Carolin, Cornelia Schoor, Ulf Kroehne, Frank Goldhammer, Nina Mahlow, and Cordula Artelt. "The role of cognitive load in university students' comprehension of multiple documents." Zeitschrift für Pädagogische Psychologie 33, no. 2 (2019): 105–18. http://dx.doi.org/10.1024/1010-0652/a000238.

Full text
Abstract:
Abstract. The study investigates the cognitive load of students working on tasks that require the comprehension of multiple documents (Multiple Document Comprehension, MDC). In a sample of 310 students, perceived task difficulty (PD) and mental effort (ME) were examined in terms of task characteristics, individual characteristics, and students' processing behavior. Moreover, it was investigated if PD and ME can still contribute to MDC while controlling for these variables. The perceived difficulty of the task was shown to be related to the number of documents, text length, study level, and sourcing. Mental effort was predicted by text length, study level, and processing time. When including these variables as covariates, cognitive load was incrementally predictive of MDC. The results are discussed in terms of how working memory resources can shape the process of comprehending multiple documents.
APA, Harvard, Vancouver, ISO, and other styles
8

Javed, Hira, Nadeem Akhtar, and M. M. Sufyan Beg. "Multimodal news document summarization." Journal of Information and Optimization Sciences 45, no. 4 (2024): 959–68. http://dx.doi.org/10.47974/jios-1619.

Full text
Abstract:
With the increase in multimedia content, the domain of multimodal processing is experiencing constant growth. The question of whether combining these modalities is beneficial may come up. In this work, we investigate this by working on multi-modal content for obtaining quality summaries. We have conducted several experiments on the extractive summarization process employing asynchronous text, audio, image,and video. Information present in the multimedia content has been leveraged to bridge the semantic gaps between different modes. Vision Transformers and BERT have been used for the imagematching and similarity-checking tasks. Furthermore, audio transcriptions have been used for incorporating the audio information in the summaries. The obtained news summaries have been evaluated with Rouge Score and a comparative analysis has been done.
APA, Harvard, Vancouver, ISO, and other styles
9

K. Adi Narayana Reddy. "Multi-Document Summarization using Discrete Bat Optimization." Journal of Electrical Systems 20, no. 7s (2024): 831–42. http://dx.doi.org/10.52783/jes.3457.

Full text
Abstract:
With the World Wide Web, we now have a wide range of data that was previously unavailable. Therefore, it has become a complex problem to find useful information in large datasets. In recent years, text summarization has emerged as a viable option for mining relevant data from massive collections of texts. We may classify summarizing as either "single document" or "multi document" depending on how many source documents we are working with. Finding an accurate summary from a collection of documents is more difficult for researchers than doing it from a single document. For this reason, this research proposes a Discrete Bat algorithm Optimization based multi document summarizer (DBAT-MDS) to tackle the issue of multi document summarizing. Comparisons are made between the proposed DBAT-MDS based model and three different summarization algorithms that take their inspiration from the natural world. All methods are evaluated in relation to the benchmark Document Understanding Conference (DUC) datasets using a variety of criteria, such as the ROUGE score and the F score. Compared to the other summarizers used in the experiment, the suggested method performs much better.
APA, Harvard, Vancouver, ISO, and other styles
10

Ovchinnikova, Irina Germanovna. "Working on Сomputer-Assisted Translation platforms: New advantages and new mistakes". Russian Journal of Linguistics 23, № 2 (2019): 544–61. http://dx.doi.org/10.22363/2312-9182-2019-23-2-544-561.

Full text
Abstract:
The paper presents analysis of errors in translation on the CAT platform Smartcat, which accumulates all tools for computer-assisted translation (CAT) including a machine translation (MT) system and translation memory (TM). The research is conducted on the material of the translation on Smartcat platform (a joint project of a tourist guide translation (35,000 words) from Hebrew to Russian, English, and French). The errors on the CAT platform disclose difficulties in mastering text semantic coherence and stylistic features. The influence of English as lingua franca appears in peculiar orthographic and punctuation errors in the target text in Russian. Peculiar errors in translation on the CAT platform reveal the necessity of advanced technological competence in translators. The peculiar errors uncover problems associated with a source text segmentation into sentences. The segmentation can trigger a translator to preserve the sentence boundaries and use a Russian complicated compound sentence that provoke punctuation errors. Difficulties of the anaphora resolution in distant semantically coherent segments are also associated with the source text segmentation and working window formatting. A joint project presupposes different translators to translate different files of the source document. To generate the coherence, contiguity and integrity of the whole document, the files have to be revised by a third-party editor to avoid conflict of interest. The editor-reviser is also responsible for improving the target text pragmatic and genre characteristics while applying top-down strategy to target text analysis. Thus, the translator’s errors while applying CAT tools reveal the effect of bottom-up text processing alongside with cross-language interference.
APA, Harvard, Vancouver, ISO, and other styles
11

VAN DEEMTER, KEES, and RICHARD POWER. "High-level authoring of illustrated documents." Natural Language Engineering 9, no. 2 (2003): 101–26. http://dx.doi.org/10.1017/s1351324902003054.

Full text
Abstract:
This paper starts by introducing a class of future document authoring systems that will allow authors to specify the content and form of a text+pictures document at a high level of abstraction, while leaving responsibility for linguistic and graphical details to the system. Next, we describe two working prototypes that implement parts of this functionality, based on semantic modeling of the pictures and the text of the document; one of these two, the ILLUSTRATE prototype, is a multimedia extension of previous text authoring systems in the What You See Is What You Meant (WYSIWYM) tradition. The paper concludes with an exploration of the ways in which Multimedia WYSIWYM can be further enhanced, allowing it to approximate the ‘ideal’ systems that were sketched earlier in the paper. Applications of Multimedia WYSIWYM to general-purpose picture retrieval (in the context of the Semantic Web, for example) are also discussed.
APA, Harvard, Vancouver, ISO, and other styles
12

Zaman, Zainab, Saad Bin Ahmed, and Muhammad Imran Malik. "Analysis of Hyperspectral Data to Develop an Approach for Document Images." Sensors 23, no. 15 (2023): 6845. http://dx.doi.org/10.3390/s23156845.

Full text
Abstract:
Hyperspectral data analysis is being utilized as an effective and compelling tool for image processing, providing unprecedented levels of information and insights for various applications. In this manuscript, we have compiled and presented a comprehensive overview of recent advances in hyperspectral data analysis that can provide assistance for the development of customized techniques for hyperspectral document images. We review the fundamental concepts of hyperspectral imaging, discuss various techniques for data acquisition, and examine state-of-the-art approaches to the preprocessing, feature extraction, and classification of hyperspectral data by taking into consideration the complexities of document images. We also explore the possibility of utilizing hyperspectral imaging for addressing critical challenges in document analysis, including document forgery, ink age estimation, and text extraction from degraded or damaged documents. Finally, we discuss the current limitations of hyperspectral imaging and identify future research directions in this rapidly evolving field. Our review provides a valuable resource for researchers and practitioners working on document image processing and highlights the potential of hyperspectral imaging for addressing complex challenges in this domain.
APA, Harvard, Vancouver, ISO, and other styles
13

Kargren, Mati, John April, Gina Clark, Jonathan Mackinnon, Aliza Nathoo, and Elizabeth Theron. "Unlocking new efficiencies: How structured content authoring is streamlining the production of clinical documents for the pharmaceutical industry." Medical Writing 32, no. 3 (2023): 32–37. http://dx.doi.org/10.56012/xafs6978.

Full text
Abstract:
Current practice requires clinical and regulatory documents to be created and updated manually by medical writers throughout a product’s development. Conventionally, document content is unstructured, with freeform text, figures, and tables that the medical writer can arrange in any configuration. By structuring and standardising clinical and regulatory content, the pharmaceutical industry can shift from a document-based to a content-based approach. This transition will require adopting structured content management tools and common structures, and standardising content. In tandem, medical writers must evolve their skillset and ways of working, primarily through planning and producing content and adopting structured content authoring practices to facilitate content creation and reuse. This article introduces structured content authoring and outlines how the medical writing role in the pharmaceutical industry may soon evolve.
APA, Harvard, Vancouver, ISO, and other styles
14

Christina, Sherly, Enny Dwi Oktaviyani, and Buyung Famungkas. "Mendeteksi Plagiarism Pada Dokumen Proposal Skripsi Menggunakan Algoritma Jaro Winkler Distance." Jurnal SAINTEKOM 8, no. 2 (2018): 143. http://dx.doi.org/10.33020/saintekom.v8i2.68.

Full text
Abstract:
The practice of plagiarism among students in working on their thesis might happen. Plagiarism is done considering to the limited workmanship and lack of motivation to try with their own ability. It is necessary to provide a tool to prevent the act of plagiarism among the students. In this research, we built an application to detect the plagiarism on the thesis proposal. 
 The application applies the Jaro Winkler Distance algorithm to detect the similarity from several documents. The first phase to detect the similarity of the thesis proposals is The text preprocessing phase of the document such as case folding, tokenizing, stopwords removal and stemming. Precision and Recall formula will be used to analyze the performance of the method in application. 
 The test result of the data set shows the application could find the 80% relevant data that indicated the similarity. It means the application could contribute to detect the plagiarism on the thesis proposal document.
APA, Harvard, Vancouver, ISO, and other styles
15

SHAKHOVSKA, Nataliya, and Andrii SHEBEKO. "DEVELOPMENT OF THE ARCHITECTURE OF DOCUMENT OPTICAL CHARACTER RECOGNITION SYSTEM." Herald of Khmelnytskyi National University. Technical sciences 309, no. 3 (2022): 50–54. http://dx.doi.org/10.31891/2307-5732-2022-309-3-50-54.

Full text
Abstract:
This paper aims to develop information technology for document optical character recognition systems. The difficulty of processing images, which are a set of pixels, causes inconvenience in working with such data. This problem can be solved in different ways: usual mathematical approaches, a single complicated neural network, and a set of problem-specific deep neural networks. Usual mathematical approaches perform poor with unstructured data like images. A single neural network is like a swiss knife: it can do many tasks, but none with the best quality. So we will use two different deep neural networks, each for the appropriate part of the problem. The critical elements of this technology are the module for text detection and segmentation of the image, the module for text recognition in Ukrainian and English languages, the module for parsing multiple keywords, and the module for searching for the final data. The first and second modules consist of several machine learning models with specific architecture, depending on their task. All trained models are tested for accuracy and noise resistance and will be used in the future for searching required data from different document images. Output data of the developed system provide speedup, automation processing images and scans of the documents, reduce the number of mistakes caused by human factor. All data is converted from image pixels into a structured text set represented in the document, which the machine can easily use. We can use such technology in banking and insurance, where we can send images of documents and they will be automatically processed and converted into user name, surname, date of birth, serial number, and required fields for specific services.
APA, Harvard, Vancouver, ISO, and other styles
16

Gazul, Stanislav M., and Ekaterina S. Kireeva. "COMPARATIVE ANALYSIS OF ARTIFICIAL INTELLIGENCE LANGUAGE MODELS AS PART OF SOLVING THE PROBLEM OF TEXT CHECKING." SOFT MEASUREMENTS AND COMPUTING 3, no. 88 (2025): 26–36. https://doi.org/10.36871/2618-9976.2025.03.003.

Full text
Abstract:
The article provides an overview of web services and applications with builtin generative models of artificial intelligence (AI) available to users for text generation purposes. The authors conducted a series of experiments to compare the results of using current generative AI models to solve practical problems of working with technical documentation. The task of checking the document "business requirement" was chosen as an example for comparison; such documents are usually formed during projects on the development of information systems (IS); therefore, the results of the study will be relevant for the widest range of organizations. Based on the analysis of the results of the experiments conducted by the authors, the best use scenarios of existing web services and applications using generative AI models were highlighted.
APA, Harvard, Vancouver, ISO, and other styles
17

Van Gelder, Hilde. "Photography Today: Between Tableau and Document." Recherches sémiotiques 28, no. 1-2 (2010): 69–86. http://dx.doi.org/10.7202/044589ar.

Full text
Abstract:
This essay seeks to examine the position of photography in contemporary art or, more specifically, the way photography now hovers between tableau and the ‘document’ Three methodological levels are considered: first, a co-textual reading of select images by Jeff Wall and Allan Sekula in relation to their titles; secondly, an examination of their various treatment of pictorial elements, remnants of a long-standing artistic tradition; thirdly, at the level of the meta-text, the same images are confronted with a much broader contextual relationship. Here, the differences between two modes of working in contemporary photography — singular tableau and (pseudo-) documentary montage — become clear. A historical flashback traces this current photographic paragone back to 16th Century Southern Netherlandish art. Finally, the author raises a question pertaining to a fourth, practical level: that of the art market.
APA, Harvard, Vancouver, ISO, and other styles
18

Willumsen, Liv Helene. "A Narratological Approach to Witchcraft Trial: A Scottish Case." Journal of Early Modern History 15, no. 6 (2011): 531–60. http://dx.doi.org/10.1163/157006511x600837.

Full text
Abstract:
Abstract This article is a microstudy of a Scottish witchcraft document from 1662, focusing of the case of a woman accused of witchcraft in Bute. It emphasizes the various “voices” that are possible to “hear” in the material—for example, the voice of the scribe, the witnesses, or the accused person. It argues on linguistic grounds that the way the story was told by the scribe influences the interpretation of this document, since the scribe had the authority over the contents of the text. This narratological analysis is finally put in a broader historical context, adding factual information about the woman accused of witchcraft and her final fate. Methodologically, this article crosses the border between literature and history, working in a new way with regard to interpretation of historical documents and is an empirical example of the fruitfulness of cross-cultural studies.
APA, Harvard, Vancouver, ISO, and other styles
19

Kaledina, E. A., O. E. Kaledin, and A. V. Artamonov. "Implementation of a Level of Access to a Language Model Using the Example of an Intelligent Applicant Support System." Programmnaya Ingeneria 16, no. 5 (2025): 260–68. https://doi.org/10.17587/prin.16.260-268.

Full text
Abstract:
The paper proposes models for segmenting text documents to implement fine-tuning of large language models using the RAG (Retrival-Augmented Generation) method. In the first model, the document is divided into a large number of fragments equal in number of characters. In this case, the partition is carried out in such a way that for any fragment there are either one or two fragments that coincide exactly by half. In the second model, the source document is divided into fragments according to the semantic content. In turn, the fragments are divided into segments, resulting in a two-level partition of the document. Based on the concept of two-level partitioning, a "Key-Value" structure is created, where the keys are text segments, and the value is the corresponding text fragment. Using the example of implementing an intelligent information system to support university applicants, a comparison and analysis of these models was carried out. The information system includes a database, web interface and chat bot. Giga chat from the Sber company was used as a large language model. Fine-tuning is carried out on the basis of data provided by the National Research Ogarev Mordovia State University. The work describes the presentation model and document structure for further work with the chatbot, as well as a vector storage that provides a convenient interface for working with vectorized data from the outside. The developed language model access layer provides an abstraction over the connected chatbot, while providing the user with a fairly low-level degree of interaction with the generative model, which can be useful when constructing complex queries.
APA, Harvard, Vancouver, ISO, and other styles
20

Letychevskyi, Oleksandr O., Maryna K. Morokhovets, and Natalia M. Shchogoleva. "Some Means of Processing Electronic Text Documents." Control Systems and Computers, no. 4 (294) (November 2021): 13–18. http://dx.doi.org/10.15407/csc.2021.04.013.

Full text
Abstract:
Introduction. Digitization of legislation is an important area today, which is identified by the government as a priority. Creating digital legal documents and verifying them for compliance with the law is a necessary task in all areas of jurisprudence. This sets the task of automatic formalizing a legal document created as an arbitrary text in natural language. Purpose. Preparing a document for storage in digital format for further processing may require prior work with an original text. When using automatic means of linguistic analysis of the texts submitted in natural language, in particular, legal, which processes the text in sentences (working up the text sequentially sentence by sentence), problems of local and global nature arise. The problem of local nature is created, in particular, by the presence in the text of the sentences, which due to their considerable length are difficult to process (with the help of one or another tool of text analysis). The problem of a global nature arises when the semantic connection between the components of different sentences should be taken into account during the automatic processing of the text. The purpose of this work is to develop means for overcoming these problems. Results. A model for structuring long sentences containing enumerations as well as a method for eliminating the synonymy of object names referred to in the text, which is intended for automatic analysis, has been developed. Conclusion. Marking up sentences containing enumerations is useful, especially when the text is intended for analysis using a procedure that processes the text sentence by sentence. Structuring a sentence with an enumeration enables, on the one hand, to prepare the sentence for processing in parts, and on the other hand, not to lose the integrity of the sentence when processing in parts. In the method of eliminating the synonymy of names proposed in this paper, both the step of identifying the names of objects and the step of revealing the identity of names requires semantic analysis. To control the correctness of these steps, Oracle was introduced to improve the reliability of the result.
APA, Harvard, Vancouver, ISO, and other styles
21

Wen, Ximing. "Language Model Meets Prototypes: Towards Interpretable Text Classification Models through Prototypical Networks." Proceedings of the AAAI Conference on Artificial Intelligence 39, no. 28 (2025): 29307–8. https://doi.org/10.1609/aaai.v39i28.35231.

Full text
Abstract:
Pretrained transformer-based Language Models (LMs) are well-known for their ability to achieve significant improvement on NLP tasks, but their black-box nature, which leads to a lack of interpretability, has been a major concern. My dissertation focuses on developing intrinsically interpretable models when using LMs as encoders while maintaining their superior performance via prototypical networks. I initiated my research by investigating enhancements in performance for interpretable models of sarcasm detection. My proposed approach focuses on capturing sentiment incongruity to enhance accuracy while offering instance-based explanations for the classification decisions. Later, we develop a novel white-box multi-head graph attention-based prototypical framework designed to explain the decisions of text classification models without sacrificing the accuracy of the original black-box LMs. In addition, I am working on extending the attention-based prototypical framework with contrastive learning to redesign an interpretable graph neural network for document classification, aiming to enhance both the interpretability and performance of the model in document classification.
APA, Harvard, Vancouver, ISO, and other styles
22

Wu, Ying-Han, Mei-Xin Huang, and Heiu-Jou Shaw. "Research the Virtual Assembly Multimedia Ship Construction Information System." Marine Technology and SNAME News 44, no. 04 (2007): 203–11. http://dx.doi.org/10.5957/mt1.2007.44.4.203.

Full text
Abstract:
This research proposes to build a "Multimedia Ship Construction Information System" to integrate the information of ship design and construction with the implementation document. The system exploits the characteristics of XML to present the electronic implementation document, and thus the information can be used repetitiously. It will then further develop into an extensible and reconstructible script language for the process script of ship construction. It integrates the virtual reality technology and establishes a mechanism that transforms text description into virtual assembly. A three-dimensional (3D) solid model of ship component built previously can be reused, and through the virtual assembly the visualization of implementation process simulation and planning can be conducted. Human intuition is used to present the construction information of the ship so that documents can be transformed from two-dimensional to three-dimensional form to accelerate the information flow on shipbuilding. With relational database technology, a multimedia (text and image) database on ship design and shipbuilding will be established through product data management (PDM), whose function is to control and manage documents to raise the value of information utilization. Moreover, the establishment of a wireless network not only can solve the communication problem between the internal shipyard and the construction site, but also assists the constructors to build the ship correctly, significantly elevate the quality of construction, and increase the working efficiency.
APA, Harvard, Vancouver, ISO, and other styles
23

Mahor, Urmila, and Aarti Kumar. "REVEALING THE TRUE AUTHORSHIP: A CASE STUDY." International Journal of Engineering Applied Sciences and Technology 7, no. 1 (2022): 197–201. http://dx.doi.org/10.33564/ijeast.2022.v07i01.031.

Full text
Abstract:
As digital media is growing, the problem of text proliferation is becoming a big problem. Therefore, the identification of true ownership of a document has become a cumbersome task. In the digital era, it is very easy to copy someone's document and publish it in their name. So it is very necessary to check the true authorship. Authorship attribution becomes difficult when we conduct it manually. However, this process needs automation, when the document size becomes large. AA is a mixture of art, science and technology that helps to discover the genuine authorship of an unknown text/document, based on its specific writing features. These specific features can reflect the author's mood, education, gender, age, ideology, religion, or motivation. Many kinds of characteristics, like lexical, character, structural, syntactic, and semantic are used in authorship recognition. In this experiment, we used approximately 120 different kinds of feature sets. In our experiment, we examined that the logistic classifier was working well and gave good results in the form of accuracy.
APA, Harvard, Vancouver, ISO, and other styles
24

Mahor, Urmila, and Aarti Kumar. "REVEALING THE TRUE AUTHORSHIP: A CASE STUDY." International Journal of Engineering Applied Sciences and Technology 7, no. 3 (2022): 212–16. http://dx.doi.org/10.33564/ijeast.2022.v07i03.033.

Full text
Abstract:
As digital media is growing, the problem of text proliferation is becoming a big problem. Therefore, the identification of true ownership of a document has become a cumbersome task. In the digital era, it is very easy to copy someone's document and publish it in their name. So it is very necessary to check the true authorship. Authorship attribution becomes difficult when we conduct it manually. However, this process needs automation, when the document size becomes large. AA is a mixture of art, science and technology that helps to discover the genuine authorship of an unknown text/document, based on its specific writing features. These specific features can reflect the author's mood, education, gender, age, ideology, religion, or motivation. Many kinds of characteristics, like lexical, character, structural, syntactic, and semantic are used in authorship recognition. In this experiment, we used approximately 120 different kinds of feature sets. In our experiment, we examined that the logistic classifier was working well and gave good results in the form of accuracy.
APA, Harvard, Vancouver, ISO, and other styles
25

Iovescu, Daniel, and Cătălin Tudose. "Real-Time Document Collaboration—System Architecture and Design." Applied Sciences 14, no. 18 (2024): 8356. http://dx.doi.org/10.3390/app14188356.

Full text
Abstract:
This article explores the world of dependable systems, specifically focusing on system design, software solutions, and architectural decisions that facilitate collaborative work on shared text documents across multiple users in near real time. It aims to dive into the intricacies of designing robust and effective document collaboration software focusing on understanding the requirements of such a system, the working principle of collaborative text editing, software architecture, technology stack selection, and tooling that can sustain such a system. To examine the pros and cons of the proposed system, the paper will detail how collaborative text editing software can benefit from such an architecture regarding availability, elasticity, and scaling. The intricate nature of this system renders this paper a valuable resource for prospective investigations within the domain of dependable systems and distributed systems. This research first examines the requirements of a real-time collaboration system and the necessary core features. Then, it analyzes the design, the application structure, and the system organization while also considering key architectural requirements as the necessity of scaling, the usage of microservices, cross-service communications, and client–server communication. For the technology stack of the implementation, this research considers the alternatives at each layer, from client to server. Once these decisions are made, it follows system development while examining possible improvements for the issues previously encountered. To validate the architecture, a testing strategy is developed, to examine the key capabilities of the system, such as resource consumption and throughput. The conclusions review the combination of modern and conventional application development principles needed to address the challenges of conflict-free document replication, decoupled and stateless event-driven architecture, idempotency, and data consistency. This paper not only showcases the design and implementation process but also sets a foundation for future research and innovation in dependable systems, collaborative technologies, sustainable solutions, and distributed system architecture.
APA, Harvard, Vancouver, ISO, and other styles
26

Bjørkvold, Tuva. "Arbeidstekster i utforskende arbeidsmetoder." Nordic Studies in Science Education 16, no. 1 (2020): 67–83. http://dx.doi.org/10.5617/nordina.6703.

Full text
Abstract:
Inquiry learning depends on a variety of writing in order to plan, collect data and keep track of the information. In school science, however, writing to document teacher initiated experiments, following a setup, is widespread. This article follows a class of students during inquiry learning. Most of the student texts are so called working texts, texts used to gather and process their data. Thus the students writing practice is closer to that of scientists, as opposed to the praxis characterizing scientific literacy in school, as reproducing and concerning the schooled text culture, such as text books. This article argues that the term working texts can pose a useful contribution to the view of writing in school science.
APA, Harvard, Vancouver, ISO, and other styles
27

Halytska, Olena. "ALGORITHM OF EXCERPTING OF PUBLICISTIC TEXT." Naukovì zapiski Nacìonalʹnogo unìversitetu «Ostrozʹka akademìâ». Serìâ «Fìlologìâ» 1, no. 16(84) (2022): 120–23. http://dx.doi.org/10.25264/2519-2558-2022-16(84)-120-123.

Full text
Abstract:
The proposed linguistic didactic study is devoted to the problem of forming the skills and abilities of philology bachelors in the analytical and synthetic processing of journalistic Ukrainian-language text and the creation of their own German-language excerpt. The article provides methodological recommendations for teaching bilingual excerpting of a Ukrainian-language journalistic text in German for students of higher education of the first bachelor's level in the field of knowledge 03 Humanities, specialty 035 Philology, under the educational and professional program «Language and Literature (German / English). Translation». A four-year-tested algorithm for bilingual abstracting of a magazine or newspaper article is proposed, which practically helps students of the Faculty of Foreign Philology master the technology of working with modern journalistic text at an advanced stage, i.e. level B2 according to the All-European Recommendations on language education. The correct automated observance of all steps of the algorithm (familiarization with the journalistic text, careful reading of the article, translation recoding of the excerpt, design of the written excerpt, self-control of the written excerpt, editing of the text of the excerpt) contributes to the correct semantic folding and adequate linguistic compression of the source text, which contributes to the optimization of the educational process under time to study German as a primary and second foreign language. The peculiarities of the structure of the excerpt as a specific secondary document are given: the exordium (bibliographic description of the primary document), the abstract part itself (the basis of the excerpt, that is, the information obtained by analytically and synthetically processing the content of the primary text) and the conclusion (addition). Creating an excerpt according to the proposed effective algorithm significantly saves time and effort of higher education students and, accordingly, increases their motivation to study. The difficulties that accompanied higher education applicants during the compilation of an excerpt were analyzed.
APA, Harvard, Vancouver, ISO, and other styles
28

Lifshits, Ilya, and Anastasiya Shatalova. "Three is none? Third-party funding in investment disputes: The UNCITRAL working document." Meždunarodnoe pravosudie 12, no. 2 (2022): 113–31. https://doi.org/10.21128/2226-2059-2022-2-113-131.

Full text
Abstract:
Since 2017 the UNCITRAL Working Group III has been working on reforming the investor-state dispute settlement system. One of the initiatives in this project is the development of draft provisions on third-party funding. Such funding constitutes a challenge to the investment dispute resolution system since third parties get an opportunity to use the system to earn a substantial profit and such funding may increase the number of frivolous claims against states. Rules of many arbitral institutions have been amended to require disclosure of a third-party funder or the provisions of a funding agreement. However, disclosure alone may not be sufficient to remedy the adverse effects of funding a case by a third party who does not participate in the proceedings. The UNCITRAL Secretariat's Draft provisions offer both an outright prohibition of this funding and various models for limiting it. The authors on the base of doctrinal sources and general scientific and specialized research methods critically assess the models proposed by the Secretariat and suggest their own vision of the regulation of third-party funding in investment disputes. The authors argue that third party finding should be allowed only in two situations: non-profit funding and funding of small and medium-sized enterprises. This approach would arguably help to avoid vague wording as well as varying interpretations by arbitrators of the provisions of the current UNCITRAL Secretariat's draft provisions on the allowed types of third-party funding of financing. At the same time this approach would not violate the right to access to justice of certain categories of investors. The authors believe that the provisions developed by states’ delegations under the auspices of UNCITRAL could become a universal model for arbitral institutions engaged in ISDS and for inclusion in the text of international investment treaties.
APA, Harvard, Vancouver, ISO, and other styles
29

Klychova, Guzaliya, Alsu Zakirova, Angelina Dyatlova, Ayrat Valiev, and Nuriya Nigmatullina. "IMPROVEMENT OF METHODOLOGICAL SUPPORT OF INTERNAL CONTROL OF THE ENTERPRISE MANAGEMENT SYSTEM." Vestnik of Kazan State Agrarian University 18, no. 2 (2023): 150–59. http://dx.doi.org/10.12737/2073-0462-2023-150-159.

Full text
Abstract:
In modern conditions, the issues of developing internal control of the management system of an economic entity are of particular relevance. In the process of internal control, management activity is evaluated in terms of its compliance with standards containing the main provisions for the effective functioning of social and economic systems. Proposed to improve the effectiveness of control procedures at the planning stage, the working document of internal control “Program of internal control of the management system” allows for a rational distribution of responsibilities between controllers by defining the control procedures to be performed. The “Test for assessing the effectiveness of the organizational structure of an economic entity” proposed in the course of the working document study allows clarifying the structure of an economic entity and its compliance with strategic development goals, assessing the effectiveness of managing structural divisions, and establishing a mechanism for interaction between management and divisions. For documenting the results of the control of the personnel management system, namely the results of the assessment of personnel potential, it is recommended to use the working document of internal control “Analysis of personnel composition”. On the basis of this document, the inspector, having assessed the qualitative composition of personnel, can determine the advantages and disadvantages in the existing personnel structure, develop recommendations for improving the personnel policy in the personnel management system. When organizing internal control of the production management system, it is important to determine the areas of control, specific tools for control measures, users of information and the results of control measures. In this regard, it is proposed to use the document “Internal control of the production management system” in the verification process. The proposed working documents of internal control make it possible to systematize the results of control procedures and provide information to the management of an economic entity for making managerial decisions.
APA, Harvard, Vancouver, ISO, and other styles
30

Medne, Antra. "Aleksandra Čaka poēmu cikla „Mūžības skartie” tapšanas galvenie iedvesmas avoti." Aktuālās problēmas literatūras un kultūras pētniecībā: rakstu krājums, no. 26/1 (March 1, 2021): 67–76. http://dx.doi.org/10.37384/aplkp.2021.26-1.067.

Full text
Abstract:
In this research, we looked at the poet Aleksandrs Čaks’s work in the historical document and memory collection “Latvian Riflemen” (Latviešu Strēlnieki), issued by Latvian former riflemen society. The main tasks of the society were to preserve the recollections of Latvian riflemen and retrieve the documents of riflemen troops and battalion battle journals, which were very important sources for historians. This information was stored in the archives of the Soviet Union. The society asked civilians and former soldiers to share knowledge and information about riflemen history. In 1935, Čaks started working on the journal “Latvian Riflemen”. That was an affirmation of his suitability and professionality after working in such journals as “Lyra of the new” (Jauno Līra), “Green Crow” (Zaļā Vārna), “Thoughts” (Domas). His duties were to edit and correct texts that were approved by the Editor-in-chief and put together the text and pictures for the issue. The issue’s main emphasis was on memories and documents about formation and preparation for the fight of former riflemen battalions. A wide variety of illustrative material was used to make the representation of the historical situation more attractive. The time Čaks spent working on the poem cycle “Touched by Eternity” (Mūžības skartie) is also studied. The big success of the cycle, received prizes and praises, events that worshiped the poet and his literary monument for Latvian riflemen are examined.
APA, Harvard, Vancouver, ISO, and other styles
31

Böschen, Ingmar. "Software review: The JATSdecoder package—extract metadata, abstract and sectioned text from NISO-JATS coded XML documents; Insights to PubMed central’s open access database." Scientometrics 126, no. 12 (2021): 9585–601. http://dx.doi.org/10.1007/s11192-021-04162-z.

Full text
Abstract:
AbstractJATSdecoder is a general toolbox which facilitates text extraction and analytical tasks on NISO-JATS coded XML documents. Its function JATSdecoder() outputs metadata, the abstract, the sectioned text and reference list as easy selectable elements. One of the biggest repositories for open access full texts covering biology and the medical and health sciences is PubMed Central (PMC), with more than 3.2 million files. This report provides an overview of the PMC document collection processed with JATSdecoder(). The development of extracted tags is displayed for the full corpus over time and in greater detail for some meta tags. Possibilities and limitations for text miners working with scientific literature are outlined. The NISO-JATS-tags are used quite consistently nowadays and allow a reliable extraction of metadata and text elements. International collaborations are more present than ever. There are obvious errors in the date stamps of some documents. Only about half of all articles from 2020 contain at least one author listed with an author identification code. Since many authors share the same name, the identification of person-related content is problematic, especially for authors with Asian names. JATSdecoder() reliably extracts key metadata and text elements from NISO-JATS coded XML files. When combined with the rich, publicly available content within PMCs database, new monitoring and text mining approaches can be carried out easily. Any selection of article subsets should be carefully performed with in- and exclusion criteria on several NISO-JATS tags, as both the subject and keyword tags are used quite inconsistently.
APA, Harvard, Vancouver, ISO, and other styles
32

Burkard, Benjamin, Georg Vogeler, and Stefan Gruner. "Informatics for Historians: Tools for Medieval Document XML Markup, and their Impact on the History-Sciences." JUCS - Journal of Universal Computer Science 14, no. (2) (2008): 193–210. https://doi.org/10.3217/jucs-014-02-0193.

Full text
Abstract:
This article is a revised and extended version of [VBG, 07]. We conjecture that the digitalization of historical text documents as a basis of data mining and information retrieval for the purpose of progress in the history sciences is urgently needed. We present a novel, specialist XML tool-suite supporting the working historian in the transcription of original medieval charters into a machine-readable form, and we also address some latest developments which can be found in the field since the publication of [VBG, 07].
APA, Harvard, Vancouver, ISO, and other styles
33

Klychova, Guzaliya, Alsu Zakirova, Amina Husainova, Augul Klychova, and Regina Nurieva. "DEVELOPMENT OF METHODOLOGICAL TOOLS FOR INTERNAL CONTROL OF PAYMENTS WITH PERSONNEL FOR WAGES." Vestnik of Kazan State Agrarian University 18, no. 3 (2023): 170–79. http://dx.doi.org/10.12737/2073-0462-2023-170-179.

Full text
Abstract:
The relevance of the study is due to the need to improve the methodological support of control procedures carried out in the process of internal control of compliance with labor legislation and accounting for settlements with personnel for wages. The internal control program developed to improve control planning serves as instructions and a means of monitoring the timely and effective implementation of control procedures. The program focuses on important areas of inspection, which contributes to the optimal organization of control activities. The internal control working document “Test for assessing the accounting system of settlements with personnel for wages”, proposed for assessing the accounting system of settlements with personnel for wages, can be used by controllers to determine areas of verification that require additional control procedures. When monitoring compliance with labor legislation, it is necessary to check the documentation of the movement of enterprise personnel. The working document “Checking the documentation of the movement of enterprise personnel” recommended for documenting the results of documentary control should be used to systematize the violations identified during the formal inspection. This allows them to be promptly eliminated and general recommendations to be formulated to prevent these violations in the future. The working document “Checking wage accruals and deductions from it” systematically reflects the operations for which violations and deviations were identified, the essence of the identified violations and recommendations for their elimination. This document allows you to identify areas in the accounting of settlements with personnel for wages, in terms of wage calculations and deductions from it, which require systematic control by the accounting service. The proposed working documents of internal control make it possible to systematize the results of control procedures and provide information to the management of an economic entity for making management decisions.
APA, Harvard, Vancouver, ISO, and other styles
34

Proskurina, N. M. "The Audit of Transactions with Customer-Owned Raw Materials in Foreign Economic Activities." Statistics of Ukraine 80, no. 1 (2018): 77–84. http://dx.doi.org/10.31767/su.1(80).2018.01.10.

Full text
Abstract:
The specifics of audit of the transactions with customer-owned raw materials in foreign economic activities (FEA) are shown. It is determined that FEA, being risky ones, need special attention from an external auditor. When scrutinizing such transactions, the latter need to be aware that when a FEA transaction is signed, including one for processing of customer-owned raw materials, the requirements of legal and regulatory acts being in force in Ukraine need to be duly considered. It is demonstrated the Custom Code of Ukraine regulates the work arrangements with customer-owned raw materials in FEA and specifies custom regimes for FEA transactions, with implications for their taxation. The taxation, in turn, is regulated by the Tax Code of Ukraine and has specifics in assessment and payment of value added tax and profit tax. The eligibility for tax credit on the commodities purchased and thereafter exported as raw materials in the custom regime of processing occur on general grounds specified in Section 5 “Value Added Tax” of the Tax Code of Ukraine, and the tax liabilities occur on the date when a customs cargo declaration is submitted for execution. According to Article 39 of the Tax Code of Ukraine, taxation of profits from transactions with related nonresident entities has some specifics. An example of the authors’ review of transactions with customer-owned raw materials in FEA, recording of the results of auditor procedures and the obtained evidences in working documents is given. The format of a working document is subject to professional opinion of an auditor. The test of recording transactions on processing of customer-owned raw materials beyond the boundaries of custom territory of Ukraine is proposed as a working document. The working document, intended to confirm the fact of collecting auditor evidences in order to form the auditor opinion on financial reporting, covers the content of transactions on processing of customer-owned raw materials, information sources for an auditor (primary documents, accounts records), audit procedures and their results, with the auditor’s opinion about the authenticity of recorded information (its confirmation). The working documents enable for regulation and quality assurance of audit procedures, and for unification of the auditor’s actions.
APA, Harvard, Vancouver, ISO, and other styles
35

Phillips, Thomas E., and Mark Glen Bilby. "Web Review: Working Virtually on the Text and Manuscripts Behind the Document: Doing New Testament Textual Criticism on the Web." Theological Librarianship 8, no. 1 (2015): 7–9. http://dx.doi.org/10.31046/tl.v8i1.374.

Full text
APA, Harvard, Vancouver, ISO, and other styles
36

Острянська, Н. В. "ТЕКСТ ДОКУМЕНТА – ВАЖЛИВА КОМУНІКАТИВНА СКЛАДОВА ПИСЕМНОГО ДІЛОВОГО МОВЛЕННЯ". Humanities journal, № 3 (3 жовтня 2019): 99–105. http://dx.doi.org/10.32620/gch.2019.3.10.

Full text
Abstract:
The purpose of the article is to comprehend and analyze the peculiarities of the text of the document as a communicative unit of written business language.Important communicative qualities of written business language are concentrated in the main element of the document, in the unit of written language - the text. The defining function of the text is to be a means of social communication in all spheres of public life.The text and its actual implementation is the problem of developing a culture of writing, the solution of which involves knowing the laws of logical and lexical word compatibility, the ability to use verbal formulas, distinguish the stylistic features of texts, and select appropriate language material for them.The text as the main element of writing has such communicative qualities: brevity, consistency of presentation of the material, conciseness, semantic accuracy, consistency, objectivity, simplicity, expressiveness of language and style and accuracy of thought. They are expressed in language constructs and verbal formulas of texts of different levels (reports, business letters, contracts, orders), in logical elements (introduction, evidence, conclusions).In the practice of modern office work, general requirements have been developed for the creation of texts of documents, some certain basic rules that ensure the realization of communicative qualities of business language and compliance with lexical, grammatical and stylistic norms.The text of the document is the basis for the use of didactic materials for various educational work.In our opinion, an important form of educational work on the formation of communication skills is the editing of texts of documents.Tasks and exercises on the correct choice of synonyms, grammatical forms for compiling the text of the document form the ability of students to distinguish the meanings of words, develop a sense of language and intuition.In the course of the study, the author of the article comes to the conclusion that it is necessary to be creative in studying business language, taking into account the level of students' knowledge, and not to forget that the text, as a real communicative unit of the document, also performs an important educational function. When flawlessly edited and formatted, the text is a part of the reputation, additional advertising, a part of the image of the enterprise, institution, organization where students will work, a kind of business card, and a sign of a high management culture.
APA, Harvard, Vancouver, ISO, and other styles
37

Uronen, L., H. Moen, S. Teperi, K.-P. Martimo, J. Hartiala, and S. Salanterä. "Towards automated detection of psychosocial risk factors with text mining." Occupational Medicine 70, no. 3 (2020): 203–6. http://dx.doi.org/10.1093/occmed/kqaa022.

Full text
Abstract:
Abstract Background Psychosocial risk factors influence early retirement and absence from work. Health checks by occupational health nurses (OHNs) may prevent deterioration of work ability. Health checks are documented electronically mostly as free text, and therefore the effect of psychological risk factors on working capacity is difficult to detect. Aims To evaluate the potential of text mining for automated early detection of psychosocial risk factors by examining health check free-text documentation, which may indicate medical statements recommending early retirement, prolonged sick leave or rehabilitation. Psychosocial risk factors were extracted from OHN documentation in a nationwide occupational health care registry. Methods Analysis of health check documentation and medical statements regarding pension, sick leave and rehabilitation. Annotations of 13 psychosocial factors based on the Prima-EF standard (PAS 1010) were used with a combination of unsupervised machine learning, a document search engine and manual filtering. Results Health check documentation was analysed for 7078 employees. In 83% of their health checks, psychosocial risk factors were mentioned. All of these occurred more frequently in the group that received medical statements for pension, rehabilitation or sick leave than the group that did not receive medical statement. Documentation of career development and work control indicated future loss of work ability. Conclusions This study showed that it was possible to detect risk factors for sick leave, rehabilitation and pension from free-text documentation of health checks. It is suggested to develop a text mining tool to automate the detection of psychosocial risk factors at an early stage.
APA, Harvard, Vancouver, ISO, and other styles
38

ARSLAN, Hüsnü Çağdaş. "SOME THOUGHTS ON THE OLD UIGHUR DOCUMENT U 5933 (CONTRIBUTIONS TO MORIYASU 2019)." Turkology 109, no. 1 (2022): 9–19. http://dx.doi.org/10.47526/2022-1/2664-3162.01.

Full text
Abstract:
There were private letters among the Uighur manuscripts in the Thousand Buddha Caves in Dunhuang. It is seen that these letters have different contents from daily human relations to working life, from commerce to state administration. There are more than two hundred letters belonging to the Old Uighurs from different times. This study consists of presenting a document which is only given the text and translation of it in the doctoral thesis I have prepared on the Old Uighur letters and their vocabulary, and announcing updated information and comments about the document. The document mentioned, read and interpreted in this study is a letter and it is preserved in the Turfan Collection of the Berlin-Brandenburgische Akademie der Wissenschaften as U 5933 (T M 84 = COUL no. 172). This document was first examined by P. Zieme in 1975 and later by J. Wilkens in 2000. Finally, this document was discussed in T. Moriyasu’s Corpus, which brought together Old Uighur private letters in 2019. Considering the view that Moriyasu expresses as “I cannot understand the reason why Zieme and Wilkens regard it as a writing by a Manichaean”, some justifications defending this view are suggested here. In addition, some new reading and interpretation suggestions are made regarding this document.
APA, Harvard, Vancouver, ISO, and other styles
39

Grosskopf, Curtis. "JEDEC's Generation of Wire Bond Pull Test Methods to Address Pulling of Copper Wire Bonds." International Symposium on Microelectronics 2021, no. 1 (2021): 000249–55. http://dx.doi.org/10.4071/1085-8024-2021.1.000249.

Full text
Abstract:
Abstract More than 50 years ago when the wire pull test method was initially added to Mil-Std 883, in Condition D of Method 2011, Bond Strength (Destructive Bond Pull Test), the test procedure and minimum pull force values were based on pull testing of mostly ultrasonic wedge bonded aluminum and gold wires of just a few different diameters. The minimum pull force values from that original data were extrapolated to cover a much wider range of wire diameters for both gold and aluminum wires. Since the release of this test method the electronics industry has manufactured copper ultrasonic wedge bonds, widely adopted copper thermosonic ball bonding roughly 15 years ago, and even developed a niche market for silver thermosonic ball bonding. The industry also developed specialty bonds such as security bonds, reverse bonds also called "stitch on ball", and even multi-loop wires and ribbons. In all that time neither the test procedure nor the minimum pull force values in Method 2011 were reviewed to determine their appropriateness for these new materials or new types of bonds, even though the industry widely referenced the test method for all of them and thus, by default, accepted its use for all of them. In late 2013, I led a working group within JEDEC's JC14.1 subcommittee, Reliability Test Methods for Packaged Devices, to update JEDEC JESD22-B116, Ball Bond Shear Test Method, to expand its scope to include the shearing of Cu ball bonds. It took the working group three years to address the necessary technical issues to ensure that the revised test method adequately addressed the shearing of copper ball bonds and propose minimum acceptable shear values. The working group produced a greatly improved document with drawings and images depicting the different shear fail modes of both gold and copper bonds and added several informative annexes to aid in the performing of the test method. By 2018 it was apparent that none of the most commonly referenced wire pull test methods in the electronics industry had made any significant progress in updating their documents to include Cu wire bonds. Therefore, the JC14.1 working group agreed to work jointly with the JC-13.7 Subcommittee, New Electronic Device Technology, to create a new, wire pull test method document under JC14.1 that would be a companion to the JESD22-B116. This new document will use Method 2011, Conditions C and D as its basis, but expand on its scope to cover copper wire bonds, both ultrasonic wedge and thermosonic ball bonds. The new test method will describe the process for a ball pull test and a stitch pull test that are referenced for copper bonds by AEC Q006, Qualification Requirement for Component Using Copper (Cu) Wire Interconnection. The test method will also provide guidance on how to perform pull testing on several different bond types used today including reverse bonds, multi-loop bonds, and stacked die. The working group plans to propose minimum pull values for copper wire bonds which JC14.1 will reference in JESD47, Stress-Test-Driven Qualification of Integrated Circuits. After the joint working group completes its work, which is targeted for some time in 2022, JC13.7 would then be able to use the output of this working group to update Method 2011 Conditions C & D. This paper will first briefly discuss the updates made to B116 to cover Cu wire bonds, but mainly focus on the work that has so far been completed by the joint working group, including a general outline of the proposed new document, JESD22-B120, Wire Bond Pull Test Methods.
APA, Harvard, Vancouver, ISO, and other styles
40

Imawan, Arga Pribadi. "Working through Boundaries." PCD Journal 10, no. 1 (2022): 31–68. http://dx.doi.org/10.22146/pcd.v10i1.3665.

Full text
Abstract:
It is commonly assumed that multidisciplinary and interdisciplinary research collaborations involve various values, knowledge, and practices, thereby existing between science and policy. This study argues, oppositely, that multidisciplinary and interdisciplinary research collaborations are socially constructed and not to be taken for granted. To support its argument, this article uses the concept of boundary work to see how the interaction between science and policy is constructed. Taking as its case study the Ground Up consortium, a collaborative water management research programme involving the Netherlands and Indonesia, this study finds that boundary work generated and formed boundaries between science and policy through a joint call for proposal documents, research proposals, and three people operating at boundaries. Furthermore, this article shows that the collaborative research in the Ground Up consortium was a social process evidenced through three mediums: text, object, and person. This qualitative research thus uses a single-case study to explore boundary work in a consortium setting. Data were collected through a review of documents (meeting notes, research proposals, and calls for proposals) as well as in-depth interviews with three members of the Ground Up consortium.
APA, Harvard, Vancouver, ISO, and other styles
41

Tolmacheva, Ekaterina B. "Anthropological photography of the 1920s – early 1930s in the context of the development of national territories by the Soviet authorities." Etnograficheskoe obozrenie, no. 1 (February 17, 2024): 133–50. http://dx.doi.org/10.31857/s0869541524010073.

Full text
Abstract:
The article is devoted to various aspects of the creation of anthropological photography in the 1920s – early 1930s. Taking the case of the history and development of a visual document, it presents the problems of scientific development of national territories. The peculiarities of formation of the complex of photographic images in the period of nation-building are considered in the context of priorities for choosing territories and ethnic groups for research, interaction with the authorities, and approach to working with the local population. The text shows the various situations in which anthropological photographic documents were created, and the reports that reflected scientific tasks of researchers. The history of the formation of the MAE RAS photographic collection is exemplified by the main stages and directions of research, the technical specificities of creating photographic documents and archiving them in the museum collection. The article further raises questions of author’s methodology for creating and collecting images, the problems related to the ethics of interaction with subjects and the publication of research materials.
APA, Harvard, Vancouver, ISO, and other styles
42

Shynkarenko, V. I., and O. S. Kuropiatnyk. "Constructive-synthesizing model of text graph representation." PROBLEMS IN PROGRAMMING, no. 2-3 (June 2016): 063–72. http://dx.doi.org/10.15407/pp2016.02-03.063.

Full text
Abstract:
The article describes the graph model of the text, allowing speeds up processing. This model allows us to identify the same fragments in the documents with the change in the order of sentences and other parts. Using constructive-synthesizing structure to formalize this model is a promising approach to further automate the process of working with the model and the text accordingly.
APA, Harvard, Vancouver, ISO, and other styles
43

Pasichna, Olena. "FEATURES OF STUDYING DOCUMENTS REGARDING PERSONNEL IN THE COURSE «THE UKRAINIAN LANGUAGE (FOR PROFESSIONAL PURPOSES)»." Scientific Bulletin of Uzhhorod University. Series: «Pedagogy. Social Work» 1, no. 56 (2025): 172–76. https://doi.org/10.24144/2524-0609.2025.56.172-176.

Full text
Abstract:
Language training in institutions of pre-higher vocational education includes the development of skills and abilities of students to draft documents for various purposes, including personnel-related documents. These are very important for mastering, as they relate to many aspects of personnel management: hiring, transferring from one position to another, dismissal, granting leave, and employee incentives. The course «The Ukrainian Language (for Professional Purposes)» as a mandatory component of the educational and professional program provides the necessary set of knowledge and skills for students of pre-higher vocational education on the topic «Drafting Personnel-related Documents». The relevance of the issue is determined by the need to implement effective forms of work in the process of studying the topic «Drafting Personnel-related Documents». The article aims to describe the system of working with personnel-related documents in order to develop practical skills and abilities of students of pre-higher vocational education to draft business papers of this group. Research methods: analysis and synthesis of scientific literature (to determine the state of development of the issue), theoretical generalization (to formulate the final provisions and conclusions). The system of working with personnel-related documents is aimed at helping students of pre-higher vocational education acquire theoretical knowledge about the specifics of these business papers and develop practical skills and abilities in drafting them. The article reveals the general features (purpose, structure, and set of requisites) of resumes, applications, personnel-related orders, autobiographies, and references. To develop the skills and abilities of students of pre-higher vocational education to draft these business papers, a system of various tasks is proposed, namely: revealing the content of concepts based on the given definitions; selecting the normative variant of word usage in personnel-related documents; editing fixed constructions used in business papers of this group; completing given sentences; correcting mistakes caused by violations of stylistic and orthographic norms; determining the type of document based on the given features; creating sample forms of documents; drafting the text of a document based on specified data. The use of the proposed system of tasks in the educational process allows students of pre-higher vocational education to acquire the necessary skills and abilities defined by the syllabus of the course «The Ukrainian Language (for Professional Purposes)». The prospect of further research is seen in the development of methodological support for studying the specifics of employment agreements and contracts as mandatory documents in personnel and contractual matters.
APA, Harvard, Vancouver, ISO, and other styles
44

Петрова, Елена Юрьевна, and Татьяна Владимировна Ершова. "Graphic techniques of working with text in geography lessons." Pedagogical Review, no. 6(52) (December 22, 2023): 21–31. http://dx.doi.org/10.23951/2307-6127-2023-6-21-31.

Full text
Abstract:
Рассматривается проблема использования на уроках географии графических приемов работы с текстом. Описана история развития приемов работы с текстом. Изучены нормативно-правовые документы, в которых определена важность формирования у обучающихся навыков работы с текстом, в том числе его графическое представление. Проанализировано содержание школьных учебников географии на наличие в них заданий по работе с графическим материалом. Описаны графические приемы работы с текстом («Резюме парафраз», «Основные мысли. Термины и понятия. Вопросы по теме», «Денотатный граф», «Кольца Венна», «Фишбоун», «Кластер», «Логические опорные конспекты», «Ментальная карта») на материале разных курсов школьной географии, приведены примеры выполненных школьниками заданий. Проанализированы результаты анкетирования обучающихся на усвоение учебного материала через графическое отображение текста. Сделан вывод, что применение на уроках географии графических приемов работы с текстом способствует более глубокому его пониманию, развитию воображения, умению систематизировать и классифицировать, выявлять причинно-следственные связи. Цель статьи состоит в описании графических приемов работы с текстом на уроках географии. Проведенное анкетирование школьников показывает интерес к данным приемам. Результаты выражаются в обосновании необходимости развивать у школьников навыки графического отображения учебного текста и внедрять в образовательную практику общеобразовательных организаций. The problem of using graphic techniques of working with text in geography lessons is considered. The history of the development of techniques of working with text is described. The legal documents have been studied, which determine the importance of the formation of students’ skills to work with the text, including its graphical representation. The content of school geography textbooks is analyzed for the presence of tasks on working with graphic material in them. The graphic techniques of working with text (“Summary of paraphrases”, “Main thoughts. Terms and concepts. Questions on the topic”, “Denotational graph”, “Venn rings”, “Fishbone”, “Cluster”, “Logical reference notes”, “Mental map”) are described based on the material of different courses of school geography, with examples of tasks performed by pupils. The results of the survey of students on the assimilation of educational material through the graphical representation of the text are analyzed. It is concluded that the use of graphic techniques of working with text in geography lessons contributes to a deeper understanding of it, the development of imagination, the ability to systematize and classify, and to identify cause-and-effect relationships. The purpose of the article is to describe graphic techniques of working with text in geography lessons. The conducted survey of pupils shows interest in these techniques. The results are expressed in the justification of the need to develop students’ skills of the graphic representation of the educational text and introduce it into the educational practice of educational organizations.
APA, Harvard, Vancouver, ISO, and other styles
45

Hermawan, Latius, and Maria Bellaniar Ismiati. "Aplikasi Pengecekan Dokumen Digital Tugas Mahasiswa Berbasis Website." Jurnal Buana Informatika 11, no. 2 (2020): 93. http://dx.doi.org/10.24002/jbi.v11i2.3706.

Full text
Abstract:
Abstract. Website-Based Application for Checking Students’ Digital Assignment. Nowadays, technology is not only about computers as it has advanced to smartphones and other things. In UKMC, technology has certainly helped the job. However, in this university, there is no application for checking the plagiarism of the students’ digital assignments, whereas plagiarism is sometimes done by students when working on assignments from online sources. Students’ assignments can be easily done by doing copy and paste without mentioning its reference because students tend to think practically when working on assignments. Plagiarism is strictly prohibited in education because it is not permitted. Therefore, a plagiarism detection application should be created. It applies a string-matching algorithm in text documents to search the common words between documents. By applying the string-matching method in document that match with other documents, an output that will provide information on how similar the text documents are can be generated. After testing, it is obtained that this application can help lecturers and students to reduce the level of plagiarism.Keywords: Application, Plagiarism, Digital, Assignment Abstrak. Sekarang teknologi tidak hanya tentang computer karena kemajuannya telah merambah pada smartphone, dan hal- hal lainnya. Di UKMC, teknologi yang digunakan sudah sangat membantu pekerjaan. Namun di universitas ini, belum ada aplikasi yang dapat memeriksa plagiarisme dari tugas digital mahasiswa padahal plagiarisme terkadang dilakukan oleh mahasiswa saat mengerjakan tugas dari sumber online. Tugas mahasiswa dapat dengan mudah dibuat dengan cara copy-paste tanpa menyebutkan referensi, karena siswa cenderung berpikir praktis ketika mengerjakan tugas. Plagiarisme sangat dilarang dalam pendidikan karena tidak diizinkan. Oleh karena itu aplikasi pendeteksi plagiarisme perlu dibuat. Aplikasi ini menerapkan algoritma pencocokan string dalam dokumen teks untuk mencari kata-kata umum antar dokumen. Dengan metode pencocokan string pada dokumen yang cocok dengan beberapa dokumen lainnya dapat dihasilkan suatu keluaran yang akan memberikan informasi seberapa dekat antar dokumen teks tersebut. Setelah dilakukan pengujian, didapat hasil bahwa aplikasi ini dapat membantu dosen dan mahasiswa untuk mengurangi tingkat plagiarisme.Kata Kunci: aplikasi, plagiarisme, tugas kuliah.
APA, Harvard, Vancouver, ISO, and other styles
46

Tkachenko, Kostiantyn. "Semantic Analysis of Natural Language Texts: Ontological Approach." Digital Platform: Information Technologies in Sociocultural Sphere 7, no. 2 (2024): 211–23. https://doi.org/10.31866/2617-796x.7.2.2024.317726.

Full text
Abstract:
The development of information (intelligent) learning systems, electronic document management systems, web-oriented systems working with text information in natural language has led to an increase in the volume of educational content and/or arrays of processed full-text documents. All this requires new means of organizing access to information, many of which should be classified as intelligent systems for knowledge processing. One of the effective approaches to identifying and processing the meaning of educational content (and/or text documents) is the use of ontologies. The purpose of the article is research, analysis of various approaches to determining the semantic content of texts in natural language, consideration of existing concepts of text analysis and prospects for using the proposed ontological approach to semantic analysis of texts in natural language. Research methods are methods of semantic analysis of the main concepts of the analyzed subject area (semantic analysis of texts in natural language). The article considers an approach to the linguistic analysis of texts based on ontological modeling. The novelty of the research is the application of the proposed ontological approach to the semantic analysis of texts in natural language to determine the meaning (semantics) of text information, which is used in intelligent systems of various classes. The conclusion of the research carried out in the article is as follows: an ontological approach to the semantic analysis of natural language text, its tasks and methods is proposed. The use of the proposed approach to text analysis leads to the understanding of semantic analysis as a single triad: <Ontology – Text – Meaning>. For effective and correct extraction of knowledge, it is suggested to use a multi-level ontology. The result of the interaction of a specific natural language text with an ontology is an ontological meaning – a set of interconnected subgraphs of the ontograph. The ontological content is extracted from the ontograph using a semantic analyzer. The dialogue processor examines the syntactic tree of sentence parsing (a connected element of the natural language text) and, based on the given question, finds a fragment in the text that is the answer to the question. Computer understanding (in information or intellectual systems) of natural language text is achieved, in particular, by: immersion of the text in a single environment of knowledge – ontology; formal presentation of meaning (semantics) in the knowledge base of the corresponding system; the possibility of operations on the ontological content. The proposed approach can be used to create intelligent information repositories that work in a single knowledge environment. The proposed approach to the semantic analysis of texts in natural language is focused on the automatic extraction of metadata from texts of various nature (for example, a text document, Internet content, educational content of relevant online courses, description of computer and board games). With further development of the proposed approach, it can be used in systems of automatic referencing of scientific publications, meaningful interpretation of multimedia content, training and testing (including elements of visual display of information and elements of gamification).
APA, Harvard, Vancouver, ISO, and other styles
47

Miah, M. Saef Ullah, Junaida Sulaiman, Talha Bin Sarwar, Kamal Z. Zamli, and Rajan Jose. "Study of Keyword Extraction Techniques for Electric Double-Layer Capacitor Domain Using Text Similarity Indexes: An Experimental Analysis." Complexity 2021 (December 2, 2021): 1–12. http://dx.doi.org/10.1155/2021/8192320.

Full text
Abstract:
Keywords perform a significant role in selecting various topic-related documents quite easily. Topics or keywords assigned by humans or experts provide accurate information. However, this practice is quite expensive in terms of resources and time management. Hence, it is more satisfying to utilize automated keyword extraction techniques. Nevertheless, before beginning the automated process, it is necessary to check and confirm how similar expert-provided and algorithm-generated keywords are. This paper presents an experimental analysis of similarity scores of keywords generated by different supervised and unsupervised automated keyword extraction algorithms with expert-provided keywords from the electric double layer capacitor (EDLC) domain. The paper also analyses which texts provide better keywords such as positive sentences or all sentences of the document. From the unsupervised algorithms, YAKE, TopicRank, MultipartiteRank, and KPMiner are employed for keyword extraction. From the supervised algorithms, KEA and WINGNUS are employed for keyword extraction. To assess the similarity of the extracted keywords with expert-provided keywords, Jaccard, Cosine, and Cosine with word vector similarity indexes are employed in this study. The experiment shows that the MultipartiteRank keyword extraction technique measured with cosine with word vector similarity index produces the best result with 92% similarity with expert-provided keywords. This study can help the NLP researchers working with the EDLC domain or recommender systems to select more suitable keyword extraction and similarity index calculation techniques.
APA, Harvard, Vancouver, ISO, and other styles
48

Plotnikova, Maria M., Anna V. Ilyina, and Elena V. Dyatlova. "The Daily Life of the Cheremkhov Miners in the Year of the Great Turning Point: The Collective Agreement of 1930." Humanitarian Vector 19, no. 2 (2024): 45–54. http://dx.doi.org/10.21209/1996-7853-2024-19-2-45-54.

Full text
Abstract:
The article is devoted to the everyday life of the Cheremkhov workers through the content of the Collective Agreement of 1930 concluded between the Cheremkhov Union of Miners and the Mine Administration. In regional historiography, the topic of everyday life of coal industry workers has not been disclosed, which determines the scientific novelty and relevance of this article. The Contract determined the amount of wages, conditions of employment, production standards, provision of work clothes and housing, it recorded a list of working specialties with a description of the level labor mechanization. The purpose of the study to analyze the everyday life of Cheremkhov workers. The authors used the case study method, source and comparative analyses, a systematic approach, and the concept of Siberia industrialization by V. P. Zinoviev. The Document was adopted in the year of the Great Turning Point – the transition from the legal functioning of the labor market to the prohibition of market relations. The text of the Contract reflects both the inertia of the previous era and the emerging realities of the new era of labor relations. Its primary characteristic is the decisive role of party bodies, which prescriptively determine the main parameters of production standards. The priority was to increase labor productivity, increase output, and consolidate working hours. In conditions of predominance of manual labor and weak mechanization, this led to an increase in the exploitation of workers. At the same time, the Document protects some workers’ rights in the sphere of work and life, so the authors compare it with the Collective Agreement of 1904 between Baku workers and Caspian oil industrialist. There are no fundamental differences between these documents, and the working and living conditions have not changed since 1904. The authors concluded that the terms of the 1930 Agreement indicate that even such a highly paid and politically privileged part of the workers of Siberia as the Cheremkhov miners lived in difficult housing conditions and were forced to maintain vegetable gardens and livestock.
APA, Harvard, Vancouver, ISO, and other styles
49

Rohman, Abu Na'im, Mutmainnah Mustofa, and Dwi Fita Heriyawati. "The Use of Augmented Reality-Based Learning Media: Arloopa for English Descriptive Text in Secondary School." Journal of Foreign Language Teaching and Learning 9, no. 1 (2024): 1–18. http://dx.doi.org/10.18196/ftl.v9i1.18789.

Full text
Abstract:
This study aims to understand the English learning process using Arloopa as a digital medium. This study uses a descriptive qualitative design. To answer the problem of the study, the researcher observed the learning and teaching process in an Islamic secondary school in Malang. Two classes were observed by immersion in the classroom to get data. To verify the data obtained, the researcher checked the learning module document. The researcher revealed that using Arloopa as a digital learning media requires additional skill technology for the teacher if the students need to be equipped with a mobile phone, which Arloopa is only working with a mobile phone. The teacher should connect the mobile phone’s display to the LCD projector so that the students can view an object displayed by Arloopa. It would be different if each student in the class utilized an Android phone.
APA, Harvard, Vancouver, ISO, and other styles
50

Lubenchenko, О. Е. "Documenting the Results of Audit of the Information Disclosed in the Notes to the Financial Reporting Compiled by International Standards." Statistics of Ukraine 89, no. 2-3 (2020): 127–47. http://dx.doi.org/10.31767/su.2-3(89-90)2020.02-03.14.

Full text
Abstract:
Financial reporting releases the information requiring disclosure in accordance with national or international accounting and reporting standards. This information is usually released in the notes to the financial reporting, intended for a broader range of users. The notes are the most informative document, by which a user of the reporting can assess the company’s solvency, financial viability or business activity, can be informed about the terms of transactions with related parties, can make assumptions and receive data on management accounting, which are laid as the basis for the management strategy building. The released information is subject to obligatory disclosure, as its content is to be analyzed by national and international regulators and auditors in a way similar to the principal forms of financial reporting. An auditor assesses the received audit evidence in view of its sufficiency and reliability. Results of audit procedures can be summarized and documented. Because the International Standards on Auditing do not provide examples of documenting, the working document “The auditor’s testing of disclosure of the information attached to the financial reporting compiled by IFRS” has been formed, to support the process of documenting audit procedures with respect to the information given in the notes to the financial reporting. Using this working document, an auditor is able to obtain detailed information about a company and indicators of its financial reports, about its items and transactions that, not being subject to recognition in the reporting, are important for the management, about the accounting foundations and the opinions of management on which a company relied in compiling the reports. Depending on the operation specifics of a company, the working document can be supplemented by the following sections: IFRS 5 “Non-current Assets Held for Sale and Discontinued Operations”, IFRS 8 “Operational Segments”, IAS 19 “Employee Benefits”, IAS 37 “Provisions, Contingent Liabilities and Contingent Assets”, IAS 40 “Investment Property” etc. The test, which is a unified working document, can be used by any audit firm. The auditor’s opinion on the information disclosure in the notes to the financial reporting is vitally important bearing in mind that not all the users of financial reports have deep knowledge of the International Accounting and Reporting Standards, but all of them need unbiased and complete data about the company status and performance.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography