Academic literature on the topic 'Transactional databases'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Transactional databases.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Transactional databases"

1

Fouad, Mohammed M., Mostafa G. M. Mostafa, Abdulfattah S. Mashat, and Tarek F. Gharib. "IMIDB: An Algorithm for Indexed Mining of Incremental Databases." Journal of Intelligent Systems 26, no. 1 (January 1, 2017): 69–85. http://dx.doi.org/10.1515/jisys-2015-0107.

Full text
Abstract:
AbstractAssociation rules provide important knowledge that can be extracted from transactional databases. Owing to the massive exchange of information nowadays, databases become dynamic and change rapidly and periodically: new transactions are added to the database and/or old transactions are updated or removed from the database. Incremental mining was introduced to overcome the problem of maintaining previously generated association rules in dynamic databases. In this paper, we propose an efficient algorithm (IMIDB) for incremental itemset mining in large databases. The algorithm utilizes the trie data structure for indexing dynamic database transactions. Performance comparison of the proposed algorithm to recently cited algorithms shows that a significant improvement of about two orders of magnitude is achieved by our algorithm. Also, the proposed algorithm exhibits linear scalability with respect to database size.
APA, Harvard, Vancouver, ISO, and other styles
2

Szafrański, Bolesław, and Rafał Bałazy. "Data protection in transactional and statistical applications of databases." Computer Science and Mathematical Modelling, no. 10/2019 (September 30, 2020): 31–39. http://dx.doi.org/10.5604/01.3001.0014.4439.

Full text
Abstract:
The article describes a discussion on the issue of data protection in databases. The discussion attempts to answer the question about the possibility of using a transactional database system as a system capable of data protection in a statistical database. The discussion is preceded by a reminder of the basic issues related to data protection in databases, including reminder of flow control models, access control models and the inference. The key element of the article is the analysis, based on the example of the Oracle database management system, whether data protection mechanisms in transactional databases can be effective in case of data protection in statistical databases.
APA, Harvard, Vancouver, ISO, and other styles
3

AL-Khafaji, Hussein, and Noora Al-Saidi. "DWORM: A Novel Algorithm To Maintain Large Itemsets in Deleted Items and/or Transactions Databases Without Re-Mining." Journal of Al-Rafidain University College For Sciences ( Print ISSN: 1681-6870 ,Online ISSN: 2790-2293 ), no. 1 (October 23, 2021): 5–24. http://dx.doi.org/10.55562/jrucs.v26i1.418.

Full text
Abstract:
Transactional databases can be updated by three cases; the addition of new transactions, the deletion of set of transactions, andor increasing or decreasing of the support of the itemsets. The update process affects the previously mined itemsets, some of the large items will be small and vice versa. Therefore the updated database should be re-mined to discover the changes in the hidden itemsets. There are many algorithms to avoid the re-mining process in the case of updating a database by addition and there is one algorithm in case of changing the value of the support. But there is no actual algorithm to avoid the re-mining in the case of deletion. This research presents a novel algorithm to manipulate this case. The proposed algorithm manipulates the three possibilities of the deletion that are deletion of one or more items from a transaction or set of transactions, deletion of a set of transactions, and deletion of items from transactions and deletion of set of transactions at the same time. The experimental result shows that the updating algorithm outperforms the re-mining process in considerable amount of execution time.
APA, Harvard, Vancouver, ISO, and other styles
4

Chen, Hongzhi, Changji Li, Chenguang Zheng, Chenghuan Huang, Juncheng Fang, James Cheng, and Jian Zhang. "G-tran." Proceedings of the VLDB Endowment 15, no. 11 (July 2022): 2545–58. http://dx.doi.org/10.14778/3551793.3551813.

Full text
Abstract:
Graph transaction processing poses unique challenges such as random data access due to the irregularity of graph structures, low throughput and high abort rate due to the relatively large read/write sets in graph transactions. To address these challenges, we present G-Tran, a remote direct memory access (RDMA)-enabled distributed in-memory graph database with serializable and snapshot isolation support. First, we propose a graph-native data store to achieve good data locality and fast data access for transactional updates and queries. Second, G-Tran adopts a fully decentralized architecture that leverages RDMA to process distributed transactions with the massively parallel processing (MPP) model, which can achieve high performance by utilizing all computing resources. In addition, we propose a new multi-version optimistic concurrency control (MV-OCC) protocol with two optimizations to address the issue of large read/write sets in graph transactions. Extensive experiments show that G-Tran achieves competitive performance compared with other popular graph databases on benchmark workloads.
APA, Harvard, Vancouver, ISO, and other styles
5

Mazurova, Oksana, Artem Naboka, and Mariya Shirokopetleva. "RESEARCH OF ACID TRANSACTION IMPLEMENTATION METHODS FOR DISTRIBUTED DATABASES USING REPLICATION TECHNOLOGY." Innovative Technologies and Scientific Solutions for Industries, no. 2 (16) (July 6, 2021): 19–31. http://dx.doi.org/10.30837/itssi.2021.16.019.

Full text
Abstract:
Today, databases are an integral part of most modern applications designed to store large amounts of data and to request from many users. To solve business problems in such conditions, databases are scaled, often horizontally on several physical servers using replication technology. At the same time, many business operations require the implementation of transactional compliance with ACID properties. For relational databases that traditionally support ACID transactions, horizontal scaling is not always effective due to the limitations of the relational model itself. Therefore, there is an applied problem of efficient implementation of ACID transactions for horizontally distributed databases. The subject matter of the study is the methods of implementing ACID transactions in distributed databases, created by replication technology. The goal of the work is to increase the efficiency of ACID transaction implementation for horizontally distributed databases. The work is devoted to solving the following tasks: analysis and selection of the most relevant methods of implementation of distributed ACID transactions; planning and experimental research of methods for implementing ACID transactions by using of NoSQL DBMS MongoDB and NewSQL DBMS VoltDB as an example; measurements of metrics of productivity of use of these methods and formation of the recommendation concerning their effective use. The following methods are used: system analysis; relational databases design; methods for evaluating database performance. The following results were obtained: experimental measurements of the execution time of typical distributed transactions for the subject area of e-commerce, as well as measurements of the number of resources required for their execution; revealed trends in the performance of such transactions, formed recommendations for the methods studied. The obtained results allowed to make functions of dependence of the considered metrics on loading parameters. Conclusions: the strengths and weaknesses of the implementation of distributed ACID transactions using MongoDB and VoltDB were identified. Practical recommendations for the effective use of these systems for different types of applications, taking into account the resources consumed and the types of requests.
APA, Harvard, Vancouver, ISO, and other styles
6

Vijay Kumar, G., M. Sreedevi, K. Bhargav, and P. Mohan Krishna. "Incremental Mining of Popular Patterns from Transactional Databases." International Journal of Engineering & Technology 7, no. 2.7 (March 18, 2018): 636. http://dx.doi.org/10.14419/ijet.v7i2.7.10913.

Full text
Abstract:
From the day the mining of frequent pattern problem has been introduced the researchers have extended the frequent patterns to various helpful patterns like cyclic, periodic, regular patterns in emerging databases. In this paper, we get to know about popular pattern which gives the Popularity of every items between the incremental databases. The method that used for the mining of popular patterns is known as Incrpop-growth algorithm. Incrpop-tree structure is been applied in this algorithm. In incremental databases the event recurrence and the event conduct of the example changes at whatever point a little arrangement of new exchanges are added to the database. In this way proposes another calculation called Incrpop-tree to mine mainstream designs in incremental value-based database utilizing Incrpop-tree structure. At long last analyses have been done and comes about are indicated which gives data about conservativeness, time proficient and space productive.
APA, Harvard, Vancouver, ISO, and other styles
7

Avni, Hillel, and Trevor Brown. "Persistent hybrid transactional memory for databases." Proceedings of the VLDB Endowment 10, no. 4 (November 2016): 409–20. http://dx.doi.org/10.14778/3025111.3025122.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Xiang, Yang, Ruoming Jin, David Fuhry, and Feodor F. Dragan. "Summarizing transactional databases with overlapped hyperrectangles." Data Mining and Knowledge Discovery 23, no. 2 (October 24, 2010): 215–51. http://dx.doi.org/10.1007/s10618-010-0203-9.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Gowtham Srinivas, P., P. Krishna Reddy, A. V. Trinath, S. Bhargav, and R. Uday Kiran. "Mining coverage patterns from transactional databases." Journal of Intelligent Information Systems 45, no. 3 (May 30, 2014): 423–39. http://dx.doi.org/10.1007/s10844-014-0318-3.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

GKOULALAS-DIVANIS, ARIS, and VASSILIOS S. VERYKIOS. "EXACT KNOWLEDGE HIDING IN TRANSACTIONAL DATABASES." International Journal on Artificial Intelligence Tools 18, no. 01 (February 2009): 17–37. http://dx.doi.org/10.1142/s0218213009000020.

Full text
Abstract:
The hiding of sensitive knowledge in the form of frequent itemsets, has gained increasing attention over the past years. This paper highlights the process of border revision, which is essential for the identification of hiding solutions bearing no side-effects, and provides efficient algorithms for the computation of the revised positive and the revised negative borders. By utilizing border revision, we unify the theory behind two exact hiding algorithms that guarantee optimal solutions both in terms of database distortion and side-effects introduced by the hiding process. Following that, we propose a novel extension to one of the hiding algorithms that allows it to identify exact hiding solutions to a much wider range of problems (than its original counterpart). Through experimentation, we compare the exact hiding schemes against two state-of-the-art heuristic algorithms and demonstrate their ability to consistently provide solutions of higher quality to a wide variety of hiding problems.
APA, Harvard, Vancouver, ISO, and other styles
More sources

Dissertations / Theses on the topic "Transactional databases"

1

Navarro, Martín Joan. "From cluster databases to cloud storage: Providing transactional support on the cloud." Doctoral thesis, Universitat Ramon Llull, 2015. http://hdl.handle.net/10803/285655.

Full text
Abstract:
Durant les últimes tres dècades, les limitacions tecnològiques (com per exemple la capacitat dels dispositius d'emmagatzematge o l'ample de banda de les xarxes de comunicació) i les creixents demandes dels usuaris (estructures d'informació, volums de dades) han conduït l'evolució de les bases de dades distribuïdes. Des dels primers repositoris de dades per arxius plans que es van desenvolupar en la dècada dels vuitanta, s'han produït importants avenços en els algoritmes de control de concurrència, protocols de replicació i en la gestió de transaccions. No obstant això, els reptes moderns d'emmagatzematge de dades que plantegen el Big Data i el cloud computing—orientats a millorar la limitacions pel que fa a escalabilitat i elasticitat de les bases de dades estàtiques—estan empenyent als professionals a relaxar algunes propietats importants dels sistemes transaccionals clàssics, cosa que exclou a diverses aplicacions les quals no poden encaixar en aquesta estratègia degut a la seva alta dependència transaccional. El propòsit d'aquesta tesi és abordar dos reptes importants encara latents en el camp de les bases de dades distribuïdes: (1) les limitacions pel que fa a escalabilitat dels sistemes transaccionals i (2) el suport transaccional en repositoris d'emmagatzematge en el núvol. Analitzar les tècniques tradicionals de control de concurrència i de replicació, utilitzades per les bases de dades clàssiques per suportar transaccions, és fonamental per identificar les raons que fan que aquests sistemes degradin el seu rendiment quan el nombre de nodes i / o quantitat de dades creix. A més, aquest anàlisi està orientat a justificar el disseny dels repositoris en el núvol que deliberadament han deixat de banda el suport transaccional. Efectivament, apropar el paradigma de l'emmagatzematge en el núvol a les aplicacions que tenen una forta dependència en les transaccions és fonamental per a la seva adaptació als requeriments actuals pel que fa a volums de dades i models de negoci. Aquesta tesi comença amb la proposta d'un simulador de protocols per a bases de dades distribuïdes estàtiques, el qual serveix com a base per a la revisió i comparativa de rendiment dels protocols de control de concurrència i les tècniques de replicació existents. Pel que fa a la escalabilitat de les bases de dades i les transaccions, s'estudien els efectes que té executar diferents perfils de transacció sota diferents condicions. Aquesta anàlisi contínua amb una revisió dels repositoris d'emmagatzematge de dades en el núvol existents—que prometen encaixar en entorns dinàmics que requereixen alta escalabilitat i disponibilitat—, el qual permet avaluar els paràmetres i característiques que aquests sistemes han sacrificat per tal de complir les necessitats actuals pel que fa a emmagatzematge de dades a gran escala. Per explorar les possibilitats que ofereix el paradigma del cloud computing en un escenari real, es presenta el desenvolupament d'una arquitectura d'emmagatzematge de dades inspirada en el cloud computing la qual s’utilitza per emmagatzemar la informació generada en les Smart Grids. Concretament, es combinen les tècniques de replicació en bases de dades transaccionals i la propagació epidèmica amb els principis de disseny usats per construir els repositoris de dades en el núvol. Les lliçons recollides en l'estudi dels protocols de replicació i control de concurrència en el simulador de base de dades, juntament amb les experiències derivades del desenvolupament del repositori de dades per a les Smart Grids, desemboquen en el que hem batejat com Epidemia: una infraestructura d'emmagatzematge per Big Data concebuda per proporcionar suport transaccional en el núvol. A més d'heretar els beneficis dels repositoris en el núvol en quant a escalabilitat, Epidemia inclou una capa de gestió de transaccions que reenvia les transaccions dels clients a un conjunt jeràrquic de particions de dades, cosa que permet al sistema oferir diferents nivells de consistència i adaptar elàsticament la seva configuració a noves demandes de càrrega de treball. Finalment, els resultats experimentals posen de manifest la viabilitat de la nostra contribució i encoratgen als professionals a continuar treballant en aquesta àrea.
Durante las últimas tres décadas, las limitaciones tecnológicas (por ejemplo la capacidad de los dispositivos de almacenamiento o el ancho de banda de las redes de comunicación) y las crecientes demandas de los usuarios (estructuras de información, volúmenes de datos) han conducido la evolución de las bases de datos distribuidas. Desde los primeros repositorios de datos para archivos planos que se desarrollaron en la década de los ochenta, se han producido importantes avances en los algoritmos de control de concurrencia, protocolos de replicación y en la gestión de transacciones. Sin embargo, los retos modernos de almacenamiento de datos que plantean el Big Data y el cloud computing—orientados a mejorar la limitaciones en cuanto a escalabilidad y elasticidad de las bases de datos estáticas—están empujando a los profesionales a relajar algunas propiedades importantes de los sistemas transaccionales clásicos, lo que excluye a varias aplicaciones las cuales no pueden encajar en esta estrategia debido a su alta dependencia transaccional. El propósito de esta tesis es abordar dos retos importantes todavía latentes en el campo de las bases de datos distribuidas: (1) las limitaciones en cuanto a escalabilidad de los sistemas transaccionales y (2) el soporte transaccional en repositorios de almacenamiento en la nube. Analizar las técnicas tradicionales de control de concurrencia y de replicación, utilizadas por las bases de datos clásicas para soportar transacciones, es fundamental para identificar las razones que hacen que estos sistemas degraden su rendimiento cuando el número de nodos y/o cantidad de datos crece. Además, este análisis está orientado a justificar el diseño de los repositorios en la nube que deliberadamente han dejado de lado el soporte transaccional. Efectivamente, acercar el paradigma del almacenamiento en la nube a las aplicaciones que tienen una fuerte dependencia en las transacciones es crucial para su adaptación a los requerimientos actuales en cuanto a volúmenes de datos y modelos de negocio. Esta tesis empieza con la propuesta de un simulador de protocolos para bases de datos distribuidas estáticas, el cual sirve como base para la revisión y comparativa de rendimiento de los protocolos de control de concurrencia y las técnicas de replicación existentes. En cuanto a la escalabilidad de las bases de datos y las transacciones, se estudian los efectos que tiene ejecutar distintos perfiles de transacción bajo diferentes condiciones. Este análisis continua con una revisión de los repositorios de almacenamiento en la nube existentes—que prometen encajar en entornos dinámicos que requieren alta escalabilidad y disponibilidad—, el cual permite evaluar los parámetros y características que estos sistemas han sacrificado con el fin de cumplir las necesidades actuales en cuanto a almacenamiento de datos a gran escala. Para explorar las posibilidades que ofrece el paradigma del cloud computing en un escenario real, se presenta el desarrollo de una arquitectura de almacenamiento de datos inspirada en el cloud computing para almacenar la información generada en las Smart Grids. Concretamente, se combinan las técnicas de replicación en bases de datos transaccionales y la propagación epidémica con los principios de diseño usados para construir los repositorios de datos en la nube. Las lecciones recogidas en el estudio de los protocolos de replicación y control de concurrencia en el simulador de base de datos, junto con las experiencias derivadas del desarrollo del repositorio de datos para las Smart Grids, desembocan en lo que hemos acuñado como Epidemia: una infraestructura de almacenamiento para Big Data concebida para proporcionar soporte transaccional en la nube. Además de heredar los beneficios de los repositorios en la nube altamente en cuanto a escalabilidad, Epidemia incluye una capa de gestión de transacciones que reenvía las transacciones de los clientes a un conjunto jerárquico de particiones de datos, lo que permite al sistema ofrecer distintos niveles de consistencia y adaptar elásticamente su configuración a nuevas demandas cargas de trabajo. Por último, los resultados experimentales ponen de manifiesto la viabilidad de nuestra contribución y alientan a los profesionales a continuar trabajando en esta área.
Over the past three decades, technology constraints (e.g., capacity of storage devices, communication networks bandwidth) and an ever-increasing set of user demands (e.g., information structures, data volumes) have driven the evolution of distributed databases. Since flat-file data repositories developed in the early eighties, there have been important advances in concurrency control algorithms, replication protocols, and transactions management. However, modern concerns in data storage posed by Big Data and cloud computing—related to overcome the scalability and elasticity limitations of classic databases—are pushing practitioners to relax some important properties featured by transactions, which excludes several applications that are unable to fit in this strategy due to their intrinsic transactional nature. The purpose of this thesis is to address two important challenges still latent in distributed databases: (1) the scalability limitations of transactional databases and (2) providing transactional support on cloud-based storage repositories. Analyzing the traditional concurrency control and replication techniques, used by classic databases to support transactions, is critical to identify the reasons that make these systems degrade their throughput when the number of nodes and/or amount of data rockets. Besides, this analysis is devoted to justify the design rationale behind cloud repositories in which transactions have been generally neglected. Furthermore, enabling applications which are strongly dependent on transactions to take advantage of the cloud storage paradigm is crucial for their adaptation to current data demands and business models. This dissertation starts by proposing a custom protocol simulator for static distributed databases, which serves as a basis for revising and comparing the performance of existing concurrency control protocols and replication techniques. As this thesis is especially concerned with transactions, the effects on the database scalability of different transaction profiles under different conditions are studied. This analysis is followed by a review of existing cloud storage repositories—that claim to be highly dynamic, scalable, and available—, which leads to an evaluation of the parameters and features that these systems have sacrificed in order to meet current large-scale data storage demands. To further explore the possibilities of the cloud computing paradigm in a real-world scenario, a cloud-inspired approach to store data from Smart Grids is presented. More specifically, the proposed architecture combines classic database replication techniques and epidemic updates propagation with the design principles of cloud-based storage. The key insights collected when prototyping the replication and concurrency control protocols at the database simulator, together with the experiences derived from building a large-scale storage repository for Smart Grids, are wrapped up into what we have coined as Epidemia: a storage infrastructure conceived to provide transactional support on the cloud. In addition to inheriting the benefits of highly-scalable cloud repositories, Epidemia includes a transaction management layer that forwards client transactions to a hierarchical set of data partitions, which allows the system to offer different consistency levels and elastically adapt its configuration to incoming workloads. Finally, experimental results highlight the feasibility of our contribution and encourage practitioners to further research in this area.
APA, Harvard, Vancouver, ISO, and other styles
2

Araujo, Neto Afonso Comba de. "Security Benchmarking of Transactional Systems." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2012. http://hdl.handle.net/10183/143292.

Full text
Abstract:
A maioria das organizações depende atualmente de algum tipo de infraestrutura computacional para suportar as atividades críticas para o negócio. Esta dependência cresce com o aumento da capacidade dos sistemas informáticos e da confiança que se pode depositar nesses sistemas, ao mesmo tempo que aumenta também o seu tamanho e complexidade. Os sistemas transacionais, tipicamente centrados em bases de dados utilizadas para armazenar e gerir a informação de suporte às tarefas diárias, sofrem naturalmente deste mesmo problema. Assim, uma solução frequentemente utilizada para amenizar a dificuldade em lidar com a complexidade dos sistemas passa por delegar sob outras organizações o trabalho de desenvolvimento, ou mesmo por utilizar soluções já disponíveis no mercado (sejam elas proprietárias ou abertas). A diversidade de software e componentes alternativos disponíveis atualmente torna necessária a existência de testes padronizados que ajudem na seleção da opção mais adequada entre as alternativas existentes, considerando uma conjunto de diferentes características. No entanto, o sucesso da investigação em testes padronizados de desempenho e confiabilidade contrasta radicalmente com os avanços em testes padronizados de segurança, os quais têm sido pouco investigados, apesar da sua extrema relevância. Esta tese discute o problema da definição de testes padronizados de segurança, comparando-o com outras iniciativas de sucesso, como a definição de testes padronizados de desempenho e de confiabilidade. Com base nesta análise é proposta um modelo de base para a definição de testes padronizados de segurança. Este modelo, aplicável de forma genérica a diversos tipos de sistemas e domínios, define duas etapas principais: qualificação de segurança e teste padronizado de confiança. A qualificação de segurança é um processo que permite avaliar um sistema tendo em conta os aspectos e requisitos de segurança mais evidentes num determinado domínio de aplicação, dividindo os sistemas avaliados entre aceitáveis e não aceitáveis. O teste padronizado de confiança, por outro lado, consiste em avaliar os sistemas considerados aceitáveis de modo a estimar a probabilidade de existirem problemas de segurança ocultados ou difíceis de detectar (o objetivo do processo é lidar com as incertezas inerentes aos aspectos de segurança). O modelo proposto é demonstrado e avaliado no contexto de sistemas transacionais, os quais podem ser divididos em duas partes: a infraestrutura e as aplicações de negócio. Uma vez que cada uma destas partes possui objetivos de segurança distintos, o modelo é utilizado no desenvolvimento de metodologias adequadas para cada uma delas. Primeiro, a tese apresenta um teste padronizado de segurança para infraestruturas de sistemas transacionais, descrevendo e justificando todos os passos e decisões tomadas ao longo do seu desenvolvimento. Este teste foi aplicado a quatro infraestruturas reais, sendo os resultados obtidos cuidadosamente apresentados e analisados. Ainda no contexto das infraestruturas de sistemas transacionais, a tese discute o problema da seleção de componentes de software. Este é um problema complexo uma vez que a avaliação de segurança destas infraestruturas não é exequível antes da sua entrada em funcionamento. A ferramenta proposta, que tem por objetivo ajudar na seleção do software básico para suportar este tipo de infraestrutura, é aplicada na avaliação e análise de sete pacotes de software distintos, todos alternativas tipicamente utilizadas em infraestruturas reais. Finalmente, a tese aborda o problema do desenvolvimento de testes padronizados de confiança para aplicações de negócio, focando especificamente em aplicações Web. Primeiro, é proposta uma abordagem baseada no uso de ferramentas de análise de código, sendo apresentadas as diversas experiências realizadas para avaliar a validade da proposta, incluindo um cenário representativo de situações reais, em que o objetivo passa por selecionar o mais seguro de entre sete alternativas de software para suportar fóruns Web. Com base nas análises realizadas e nas limitações desta proposta, é de seguida definida uma abordagem genérica para a definição de testes padronizados de confiança para aplicações Web.
Most organizations nowadays depend on some kind of computer infrastructure to manage business critical activities. This dependence grows as computer systems become more reliable and useful, but so does the complexity and size of systems. Transactional systems, which are database-centered applications used by most organizations to support daily tasks, are no exception. A typical solution to cope with systems complexity is to delegate the software development task, and to use existing solutions independently developed and maintained (either proprietary or open source). The multiplicity of software and component alternatives available has boosted the interest in suitable benchmarks, able to assist in the selection of the best candidate solutions, concerning several attributes. However, the huge success of performance and dependability benchmarking markedly contrasts with the small advances on security benchmarking, which has only sparsely been studied in the past. his thesis discusses the security benchmarking problem and main characteristics, particularly comparing these with other successful benchmarking initiatives, like performance and dependability benchmarking. Based on this analysis, a general framework for security benchmarking is proposed. This framework, suitable for most types of software systems and application domains, includes two main phases: security qualification and trustworthiness benchmarking. Security qualification is a process designed to evaluate the most obvious and identifiable security aspects of the system, dividing the evaluated targets in acceptable or unacceptable, given the specific security requirements of the application domain. Trustworthiness benchmarking, on the other hand, consists of an evaluation process that is applied over the qualified targets to estimate the probability of the existence of hidden or hard to detect security issues in a system (the main goal is to cope with the uncertainties related to security aspects). The framework is thoroughly demonstrated and evaluated in the context of transactional systems, which can be divided in two parts: the infrastructure and the business applications. As these parts have significantly different security goals, the framework is used to develop methodologies and approaches that fit their specific characteristics. First, the thesis proposes a security benchmark for transactional systems infrastructures and describes, discusses and justifies all the steps of the process. The benchmark is applied to four distinct real infrastructures, and the results of the assessment are thoroughly analyzed. Still in the context of transactional systems infrastructures, the thesis also addresses the problem of the selecting software components. This is complex as evaluating the security of an infrastructure cannot be done before deployment. The proposed tool, aimed at helping in the selection of basic software packages to support the infrastructure, is used to evaluate seven different software packages, representative alternatives for the deployment of real infrastructures. Finally, the thesis discusses the problem of designing trustworthiness benchmarks for business applications, focusing specifically on the case of web applications. First, a benchmarking approach based on static code analysis tools is proposed. Several experiments are presented to evaluate the effectiveness of the proposed metrics, including a representative experiment where the challenge was the selection of the most secure application among a set of seven web forums. Based on the analysis of the limitations of such approach, a generic approach for the definition of trustworthiness benchmarks for web applications is defined.
APA, Harvard, Vancouver, ISO, and other styles
3

Li, Yanrong. "Techniques for improving clustering and association rules mining from very large transactional databases." Thesis, Curtin University, 2009. http://hdl.handle.net/20.500.11937/907.

Full text
Abstract:
Clustering and association rules mining are two core data mining tasks that have been actively studied by data mining community for nearly two decades. Though many clustering and association rules mining algorithms have been developed, no algorithm is better than others on all aspects, such as accuracy, efficiency, scalability, adaptability and memory usage. While more efficient and effective algorithms need to be developed for handling the large-scale and complex stored datasets, emerging applications where data takes the form of streams pose new challenges for the data mining community. The existing techniques and algorithms for static stored databases cannot be applied to the data streams directly. They need to be extended or modified, or new methods need to be developed to process the data streams.In this thesis, algorithms have been developed for improving efficiency and accuracy of clustering and association rules mining on very large, high dimensional, high cardinality, sparse transactional databases and data streams.A new similarity measure suitable for clustering transactional data is defined and an incremental clustering algorithm, INCLUS, is proposed using this similarity measure. The algorithm only scans the database once and produces clusters based on the user’s expectations of similarities between transactions in a cluster, which is controlled by the user input parameters, a similarity threshold and a support threshold. Intensive testing has been performed to evaluate the effectiveness, efficiency, scalability and order insensitiveness of the algorithm.To extend INCLUS for transactional data streams, an equal-width time window model and an elastic time window model are proposed that allow mining of clustering changes in evolving data streams. The minimal width of the window is determined by the minimum clustering granularity for a particular application. Two algorithms, CluStream_EQ and CluStream_EL, based on the equal-width window model and the elastic window model respectively, are developed by incorporating these models into INCLUS. Each algorithm consists of an online micro-clustering component and an offline macro-clustering component. The online component writes summary statistics of a data stream to the disk, and the offline components uses those summaries and other user input to discover changes in a data stream. The effectiveness and scalability of the algorithms are evaluated by experiments.This thesis also looks into sampling techniques that can improve efficiency of mining association rules in a very large transactional database. The sample size is derived based on the binomial distribution and central limit theorem. The sample size used is smaller than that based on Chernoff Bounds, but still provides the same approximation guarantees. The accuracy of the proposed sampling approach is theoretically analyzed and its effectiveness is experimentally evaluated on both dense and sparse datasets.Applications of stratified sampling for association rules mining is also explored in this thesis. The database is first partitioned into strata based on the length of transactions, and simple random sampling is then performed on each stratum. The total sample size is determined by a formula derived in this thesis and the sample size for each stratum is proportionate to the size of the stratum. The accuracy of transaction size based stratified sampling is experimentally compared with that of random sampling.The thesis concludes with a summary of significant contributions and some pointers for further work.
APA, Harvard, Vancouver, ISO, and other styles
4

Liu, Yufan. "A Survey Of Persistent Graph Databases." Kent State University / OhioLINK, 2014. http://rave.ohiolink.edu/etdc/view?acc_num=kent1395166105.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Niles, Duane Francis Jr. "Improving Performance of Highly-Programmable Concurrent Applications by Leveraging Parallel Nesting and Weaker Isolation Levels." Thesis, Virginia Tech, 2015. http://hdl.handle.net/10919/54557.

Full text
Abstract:
The recent development of multi-core computer architectures has largely affected the creation of everyday applications, requiring the adoption of concurrent programming to significantly utilize the divided processing power of computers. Applications must be split into sections able to execute in parallel, without any of these sections conflicting with one another, thereby necessitating some form of synchronization to be declared. The most commonly used methodology is lock-based synchronization; although, to improve performance the most, developers must typically form complex, low-level implementations for large applications, which can easily create potential errors or hindrances. An abstraction from database systems, known as transactions, is a rising concurrency control design aimed to circumvent the challenges with programmability, composability, and scalability in lock-based synchronization. Transactions execute their operations speculatively and are capable of being restarted (or rolled back) when there exist conflicts between concurrent actions. As such issues can occur later in the lifespans of transactions, entire rollbacks are not that effective for performance. One particular method, known as nesting, was created to counter that drawback. Nesting is the act of enclosing transactions within other transactions, essentially dividing the work into pieces called sub-transactions. These sub-transactions can roll back without affecting the entire main transaction, although general nesting models only allow one sub-transaction to perform work at a time. The first main contribution in this thesis is SPCN, an algorithm that parallelizes nested transactions while automatically processing any potential conflicts that may arise, eliminating the burden of additional processing from the application developers. Two versions of SPCN exist: Strict, which enforces the sub-transactions' work to be made visible in a serialized order; and Relaxed, which allows sub-transactions to distribute their information immediately as they finish (therefore invalidation may occur after-the-fact and must be handled). Despite the additional logic required by SPCN, it outperforms traditional closed nesting by 1.78x at the lowest and 3.78x at the highest in the experiments run. Another method to alter transactional execution and boost performance is to relax the rules of visibility for parallel operations (known as their isolation). Depending on the application, correctness is not broken even if some transactions see external work that may later be undone due to a rollback, or if an object is written while another transaction is using an older instance of its data. With lock-based synchronization, developers would have to explicitly design their application with varying amounts of locks, and different lock organizations or hierarchies, to change the strictness of the execution. With transactional systems, the processing performed by the system itself can be set to utilize different rulings, which can change the performance of an application without requiring it to be largely redesigned. This notion leads to the second contribution in this thesis: AsR, or As-Serializable transactions. Serializability is the general form of isolation or strictness for transactions in many applications. In terms of execution, its definition is equivalent to only one transaction running at a time in a given system. Many transactional systems use their own internal form of locking to create Serializable executions, but it is typically too strict for many applications. AsR transactions allow the internal processing to be relaxed while additional meta-data is used external to the system, without requiring any interaction from the developer or any changes to the given application. AsR transactions offer multiple orders of magnitude more in throughput in highly-contentious scenarios, due to their capability to outlast traditional levels of isolation.
Master of Science
APA, Harvard, Vancouver, ISO, and other styles
6

Bejaoui, Lofti. "Qualitative topological relationships for objects with possibly vague shapes: implications on the specification of topological integrity constraints in transactional spatial databases and in spatial data warehouses." Phd thesis, Université Blaise Pascal - Clermont-Ferrand II, 2009. http://tel.archives-ouvertes.fr/tel-00725614.

Full text
Abstract:
Dans les bases de données spatiales actuellement mises en oeuvre, les phénomènes naturels sont généralement représentés par des géométries ayant des frontières bien délimitées. Une telle description de la réalité ignore le vague qui caractérise la forme de certains objets spatiaux (zones d'inondation, lacs, peuplements forestiers, etc.). La qualité des données enregistrées est donc dégradée du fait de ce décalage entre la réalitée et sa description. Cette thèse s'attaque à ce problème en proposant une nouvelle approche pour représenter des objets spatiaux ayant des formes vagues et caractériser leurs relations topologiques. Le modèle proposé, appelé QMM model (acronyme de Qualitative Min-Max model), utilise les notions d'extensions minimale et maximale pour représenter la partie incertaine d'un objet. Un ensemble d'adverbes permet d'exprimer la forme vague d'un objet (ex : a region with a partially broad boundary), ainsi que l'incertitude des relations topologiques entre deux objets (ex : weakly Contains, fairly Contains, etc.). Cette approche est moins fine que d'autres approches concurrentes (modélisation par sous-ensembles flous ou modélisation probabiliste). Mais elle nécessite un processus d'acquisition complexe des données. De plus elle est relativement simple à mettre en oeuvre avec les systèmes existants de gestion de bases de données. Cette approche est ensuite utilisée pour contrôler la qualité des données dans les bases de données spatiales et les entrepôts de données spatiales en spécifiant les contraintes d'intégrité basé sur les concepts du modèle QMM. Une extension du langage de contraintes OCL (Object Constraint Language) a été étudiée pour spécifier des contraintes topologiques impliquant des objets ayant des formes vagues. Un logiciel existant (outil OCLtoSQL développé à l'Université de Dresden) a été étendu pour permettre la génération automatique du code SQL d'une contrainte lorsque la base de données est gérée par un système relationnel. Une expérimentation de cet outil a été réalisée avec une base de données utilisée pour la gestion des épandages agricoles. Pour cette application, l'approche et l'outil sont apparus très efficients. Cette thèse comprend aussi une étude de l'intégration de bases de données spatiales hétérogènes lorsque les objets sont représentés avec le modèle QMM. Des résultats nouveaux ont été produits et des exemples d'application ont été explicités.
APA, Harvard, Vancouver, ISO, and other styles
7

Bejaoui, Lotfi. "Qualitative Topological Relationships for Objects with Possibly Vague Shapes: Implications on the Specification of Topological Integrity Constraints in Transactional Spatial Databases and in Spatial Data Warehouses." Thesis, Université Laval, 2009. http://www.theses.ulaval.ca/2009/26348/26348.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Burger, Albert G. "Branching transactions : a transaction model for parallel database systems." Thesis, University of Edinburgh, 1996. http://hdl.handle.net/1842/15591.

Full text
Abstract:
In order to exploit parallel computers, database management systems must achieve a high level of concurrency when executing transactions. In a high contention environment, however, concurrency is severely limited due to transaction blocking, and the utilisation of parallel hardware resources, e.g. multiple CPUs, can be low. In this dissertation, a new transaction model, Branching Transactions, is proposed. Under branching transactions, more than one possible path of execution of a transaction is followed up in parallel, which allows us to avoid unnecessary transaction blockings and restarts. This approach uses additional hardware resources, mainly CPU - which would otherwise sit idle due to data contention - to improve transaction response time and throughput. A new transaction model has implications for many transaction processing algorithms, in particular concurrency control. A family of locking algorithms, based on multi-version two-phase locking, has been developed for branching transactions, including an algorithm which can dynamically switch between branching and non-branching modes. The issues of deadlock handling and recovery are also considered. The correctness of all new concurrency control algorithms is proved by extending traditional serializability theory so that it is able to cope with the notion of a branching transaction. Architectural descriptions of branching transaction systems for shared-memory parallel data-bases and hybrid shared-disk/shared-memory systems are discussed. In particular, the problem of cache coherence is addressed. The performance of branching transactions in a shared-memory parallel database system has been investigated, using discrete-event simulation. One field which may potentially benefit greatly from branching transactions is that of so-called "real-time" database systems, in which transactions have execution deadlines. A new real-time concurrency control algorithm based on branching transactions is introduced.
APA, Harvard, Vancouver, ISO, and other styles
9

Dias, Ricardo Jorge Freire. "Cooperative memory and database transactions." Master's thesis, Faculdade de Ciências e Tecnologia, 2008. http://hdl.handle.net/10362/4192.

Full text
Abstract:
Dissertação apresentada na Faculdade de Ciências e Tecnologia da Universidade Nova de Lisboa para a obtenção do Grau de Mestre em Engenharia Informática
Since the introduction of Software Transactional Memory (STM), this topic has received a strong interest by the scientific community, as it has the potential of greatly facilitating concurrent programming by hiding many of the concurrency issues under the transactional layer, being in this way a potential alternative to the lock based constructs, such as mutexes and semaphores. The current practice of STM is based on keeping track of changes made to the memory and, if needed, restoring previous states in case of transaction rollbacks. The operations in a program that can be reversible,by restoring the memory state, are called transactional operations. The way that this reversibility necessary to transactional operations is achieved is implementation dependent on the STM libraries being used. Operations that cannot be reversed,such as I/O to external data repositories (e.g., disks) or to the console, are called nontransactional operations. Non-transactional operations are usually disallowed inside a memory transaction, because if the transaction aborts their effects cannot be undone. In transactional databases, operations like inserting, removing or transforming data in the database can be undone if executed in the context of a transaction. Since database I/O operations can be reversed, it should be possible to execute those operations in the context of a memory transaction. To achieve such purpose, a new transactional model unifying memory and database transactions into a single one was defined, implemented, and evaluated. This new transactional model satisfies the properties from both the memory and database transactional models. Programmers can now execute memory and database operations in the same transaction and in case of a transaction rollback, the transaction effects in both the memory and the database are reverted.
APA, Harvard, Vancouver, ISO, and other styles
10

Aleksic, Mario. "Incremental computation methods in valid and transaction time databases." Thesis, Georgia Institute of Technology, 1996. http://hdl.handle.net/1853/8126.

Full text
APA, Harvard, Vancouver, ISO, and other styles
More sources

Books on the topic "Transactional databases"

1

N, Chorafas Dimitris. Transaction management: Managing complex transactions and sharing distributed databases. New York, N.Y: St. Martin's Press, 1998.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
2

Zawis, John A. Accessing hierarchical databases via SQL transactions in a multi-model database system. Monterey, Calif: Naval Postgraduate School, 1987.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
3

Walpole, Dennis A. Accessing network databases via SQL transactions in a multi-model database system. Monterey, Calif: Naval Postgraduate School, 1989.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
4

McManus, Thomas E. Telephone transaction-generated information: Rights and restrictions. Cambridge, Mass. (200 Aiken, Cambridge 02138): Program on Information Resources Policy, Harvard University, Center for Information Policy Research, 1990.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
5

Saake, Gunter, Kerstin Schwarz, and Can Türker, eds. Transactions and Database Dynamics. Berlin, Heidelberg: Springer Berlin Heidelberg, 2000. http://dx.doi.org/10.1007/3-540-46466-2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Freitag, Burkhard, Hendrik Decker, Michael Kifer, and Andrei Voronkov, eds. Transactions and Change in Logic Databases. Berlin, Heidelberg: Springer Berlin Heidelberg, 1998. http://dx.doi.org/10.1007/bfb0055493.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Kifer, M. Database systems: An application-oriented approach. 2nd ed. Boston: Pearson/Addison Wesley, 2005.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
8

1960-, Saake Gunter, Schwarz Kerstin, and Türker Can, eds. Transactions and database dynamics: 8th International Workshop on Foundations of Models and Languages for Data and Objects, Dagstuhl Castle, Germany, September 27-30, 1999 : selected papers. Berlin: Springer, 2000.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
9

Sowade, Olajide B. Transaction Management in Object Oriented Distributed Databases. London: University ofEast London, 1995.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
10

Kuhn, Darl, and Thomas Kyte. Oracle Database Transactions and Locking Revealed. Berkeley, CA: Apress, 2021. http://dx.doi.org/10.1007/978-1-4842-6425-6.

Full text
APA, Harvard, Vancouver, ISO, and other styles
More sources

Book chapters on the topic "Transactional databases"

1

Vossen, Gottfried. "Transactional workflows." In Deductive and Object-Oriented Databases, 20–25. Berlin, Heidelberg: Springer Berlin Heidelberg, 1997. http://dx.doi.org/10.1007/3-540-63792-3_4.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Haghir Chehreghani, Mostafa, and Morteza Haghir Chehreghani. "Transactional Tree Mining." In Machine Learning and Knowledge Discovery in Databases, 182–98. Cham: Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-46128-1_12.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Kempster, Tim, Gordon Brebner, and Peter Thanisch. "A Transactional Approach to Configuring Telecommunications Services." In Databases in Telecommunications, 40–53. Berlin, Heidelberg: Springer Berlin Heidelberg, 2000. http://dx.doi.org/10.1007/10721056_4.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Leung, Carson Kai-Sang, and Syed K. Tanbeer. "Mining Popular Patterns from Transactional Databases." In Data Warehousing and Knowledge Discovery, 291–302. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. http://dx.doi.org/10.1007/978-3-642-32584-7_24.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Kiran, R. Uday, and Masaru Kitsuregawa. "Discovering Chronic-Frequent Patterns in Transactional Databases." In Databases in Networked Information Systems, 12–26. Cham: Springer International Publishing, 2015. http://dx.doi.org/10.1007/978-3-319-16313-0_2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Tanbeer, Syed Khairuzzaman, Chowdhury Farhan Ahmed, Byeong-Soo Jeong, and Young-Koo Lee. "Discovering Periodic-Frequent Patterns in Transactional Databases." In Advances in Knowledge Discovery and Data Mining, 242–53. Berlin, Heidelberg: Springer Berlin Heidelberg, 2009. http://dx.doi.org/10.1007/978-3-642-01307-2_24.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Bharati, R. D., and V. Z. Attar. "Workload-Driven Transactional Partitioning for Distributed Databases." In Data Intelligence and Cognitive Informatics, 389–96. Singapore: Springer Singapore, 2021. http://dx.doi.org/10.1007/978-981-15-8530-2_31.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Bharati, R. D., and V. Z. Attar. "Workload-Driven Transactional Partitioning for Distributed Databases." In Data Intelligence and Cognitive Informatics, 389–96. Singapore: Springer Singapore, 2021. http://dx.doi.org/10.1007/978-981-15-8530-2_31.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Rashid, Md Mamunur, Md Rezaul Karim, Byeong-Soo Jeong, and Ho-Jin Choi. "Efficient Mining Regularly Frequent Patterns in Transactional Databases." In Database Systems for Advanced Applications, 258–71. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. http://dx.doi.org/10.1007/978-3-642-29038-1_20.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Nofong, Vincent Mwintieru. "Discovering Productive Periodic Frequent Patterns in Transactional Databases." In Data Science, 141–50. Cham: Springer International Publishing, 2015. http://dx.doi.org/10.1007/978-3-319-24474-7_20.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Transactional databases"

1

Xiang, Yang, Ruoming Jin, David Fuhry, and Feodor F. Dragan. "Succinct summarization of transactional databases." In the 14th ACM SIGKDD international conference. New York, New York, USA: ACM Press, 2008. http://dx.doi.org/10.1145/1401890.1401981.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Palmerini, P., S. Orlando, and R. Perego. "Statistical properties of transactional databases." In the 2004 ACM symposium. New York, New York, USA: ACM Press, 2004. http://dx.doi.org/10.1145/967900.968009.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

"TRANSACTIONAL SUPPORT IN NATIVE XML DATABASES." In 10th International Conference on Enterprise Information Systems. SciTePress - Science and and Technology Publications, 2008. http://dx.doi.org/10.5220/0001725303680373.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Sreedevi, M., and L. S. S. Reddy. "Mining regular closed patterns in transactional databases." In 2013 7th International Conference on Intelligent Systems and Control (ISCO). IEEE, 2013. http://dx.doi.org/10.1109/isco.2013.6481184.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Welke, Pascal. "Efficient Frequent Subgraph Mining in Transactional Databases." In 2020 IEEE 7th International Conference on Data Science and Advanced Analytics (DSAA). IEEE, 2020. http://dx.doi.org/10.1109/dsaa49011.2020.00044.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Tanbeer, Syed Khairuzzaman, Chowdhury Farhan Ahmed, and Byeong-Soo Jeong. "Mining Regular Patterns in Incremental Transactional Databases." In 2010 12th Asia Pacific Web Conference (APWEB). IEEE, 2010. http://dx.doi.org/10.1109/apweb.2010.69.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Dey, Akon, Alan Fekete, Raghunath Nambiar, and Uwe Rohm. "YCSB+T: Benchmarking web-scale transactional databases." In 2014 IEEE 30th International Conference on Data Engineering Workshops (ICDEW). IEEE, 2014. http://dx.doi.org/10.1109/icdew.2014.6818330.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Leis, Viktor, Alfons Kemper, and Thomas Neumann. "Exploiting hardware transactional memory in main-memory databases." In 2014 IEEE 30th International Conference on Data Engineering (ICDE). IEEE, 2014. http://dx.doi.org/10.1109/icde.2014.6816683.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Razdan, Varun, Aman Modi, Shruti Dumbare, and Rahul Jobanputra. "YCSB+T: Benchmark for transactional databases and performance." In 2017 International Conference of Electronics, Communication and Aerospace Technology (ICECA). IEEE, 2017. http://dx.doi.org/10.1109/iceca.2017.8212845.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Thomas Mason, Robert. "NoSQL Databases and Data Modeling for a Document-oriented NoSQL Database." In InSITE 2015: Informing Science + IT Education Conferences: USA. Informing Science Institute, 2015. http://dx.doi.org/10.28945/2245.

Full text
Abstract:
NoSQL databases are an important component of Big Data for storing and retrieving large volumes of data. Traditional Relational Database Management Systems (RDBMS) use the ACID theorem for data consistency, whereas NoSQL Databases use a non-transactional approach called BASE. RDBMS scale vertically and NoSQL Databases can scale both horizontally (sharding) and vertically. Four types of NoSQL databases are Document-oriented, Key-Value Pairs, Column-oriented and Graph. Data modeling for Document-oriented databases is similar to data modeling for traditional RDBMS during the conceptual and logical modeling phases. However, for a physical data model, entities can be combined (denormalized) by using embedding. What was once called a foreign key in a traditional RDBMS is now called a reference in a Document-oriented NoSQL database.
APA, Harvard, Vancouver, ISO, and other styles

Reports on the topic "Transactional databases"

1

Tolan, Gil D. Transaction Design Specification Medical Exam Databases System (MED) update Transaction. Fort Belvoir, VA: Defense Technical Information Center, December 1986. http://dx.doi.org/10.21236/ada271597.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Kang, Myong H., Oliver Costich, and Judith N. Froscher. A Practical Transaction Model and Untrusted Transaction Manager for a Multilevel-Secure Database System. Fort Belvoir, VA: Defense Technical Information Center, January 1992. http://dx.doi.org/10.21236/ada462360.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Costich, Oliver, and Sushil Jajodia. Maintaining Multilevel Transaction Atomicity in MLS Database Systems with Kernelized Architecture. Fort Belvoir, VA: Defense Technical Information Center, January 1993. http://dx.doi.org/10.21236/ada465420.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Costich, Oliver. Transaction Processing Using an Untrusted Scheduler in a Multilevel Database with Replicated Architecture. Fort Belvoir, VA: Defense Technical Information Center, January 1992. http://dx.doi.org/10.21236/ada462366.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Costich, Oliver, and John McDermott. A Multilevel Transaction Problem for Multilevel Secure Database Systems and its Solution for the Replicated Architecture. Fort Belvoir, VA: Defense Technical Information Center, January 1992. http://dx.doi.org/10.21236/ada462530.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Avellán, Leopoldo, Arturo Galindo, Giulia Lotti, and Juan Pablo Rodríguez Bonilla. Open configuration options Bridging the Gap: Mobilization of Multilateral Development Banks in Infrastructure. Inter-American Development Bank, February 2022. http://dx.doi.org/10.18235/0004006.

Full text
Abstract:
We explore how Multilateral Development Banks (MDBs) can help to fill a large infrastructure financing gap in developing countries by indirectly mobilizing resources from other entities. The analysis focuses on more than 6,500 transactions in 2005-2020 to developing and emerging markets from the Infrastructure Journal database. Using granular data, we analyze the dynamics of flows from different actors to infrastructure at the country-subsector level, and control for a wide range of fixed effects. MDB lending significantly increases the inflows from other sources. Cross-border and domestic resources are mobilized from both the public and the private sectors. Effects exhibit country heterogeneity. Mobilization occurs in countries of all income levels, though it is stronger in low and lower-middle income countries. In countries that use capital controls frequently mobilization effects are undermined. When the global financial crisis of 2008 hit, no difference in mobilization effects was found, unlike the COVID-19 pandemic when mobilization effects were weakened. The findings survive a long battery of robustness checks, and no evidence of anticipation effects is found.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography