Dissertations / Theses on the topic 'Bases de données réparties'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Bases de données réparties.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Vargas-Solar, Genoveva. "Service d'évènements flexible pour l'intégration d'applications bases de données réparties." Université Joseph Fourier (Grenoble ; 1971-2015), 2000. http://www.theses.fr/2000GRE10259.
Full textRobert, de Saint Victor Isabelle. "Système déductif dans le contexte de données réparties." Lyon 1, 1988. http://www.theses.fr/1988LYO10084.
Full textQuiané-Ruiz, Jorge-Alnulfo. "Allocation de requêtes dans des systèmes d'information distribués avec des participants autonomes." Nantes, 2008. https://tel.archives-ouvertes.fr/tel-00464475.
Full textIn large-scale distributed information systems, where participants (consumers and providers) are autonomous and have special interests for some queries, query allocation is a challenge. Much work in this context has focused on distributing queries among providers in a way that maximizes overall performance (typically throughput and response time). However, participants usually have certain expectations with respect to the mediator, which are not only performance-related. Such expectations mainly reflect their interests to allocate and perform queries, e. G. Their interests towards: providers (based on reputation for example), quality of service, topics of interests, and relationships with other participants. In this context, because of participants’ autonomy, dissatisfaction is a problem since it may lead participants to leave the mediator. Participant’s satisfaction means that the query allocation method meets its expectations. Thus, besides balancing query load, preserving the participants’ interests so that they are satisfied is also important. In this thesis, we address the query allocation problem in these environments and make the following main contributions. First, we provide a model to characterize the participants’ perception of the system regarding their interests and propose measures to evaluate the quality of query allocation methods. Second, we propose a framework for query allocation, called SbQA, that dynamically trades consumers’ interests for providers’ interests based on their satisfaction. Third, we propose an query allocation approach, called SbQA, that allows a query allocation method (specifically SbQA) to scale up in terms of the numbers of mediators, participants, and hence of performed queries. Fourth, we propose a query replication method, called SbQR, that allows to support participants’ failures when allocating queries while preserving participants’ satisfaction and good system performance. Last, but not least, we analytically and experimentally validate our proposals and demonstrate that they yield high efficiency while satisfying participants
Kerhervé, Brigitte. "Vues relationnelles : implantation dans les systèmes de gestion de bases de données centralisés et répartis." Paris 6, 1986. http://www.theses.fr/1986PA066090.
Full textSarr, Idrissa. "Routage des transactions dans les bases de données à large échelle." Paris 6, 2010. http://www.theses.fr/2010PA066330.
Full textMeynard, Michel. "Contrôle de la cohérence des bases de données réparties et dupliquées, sujettes aux partitionnements." Montpellier 2, 1990. http://www.theses.fr/1990MON20022.
Full textSahri, Soror. "Conception et implantation d'un système de bases de données distribuée & scalable : SD-SQL Server." Paris 9, 2006. https://portail.bu.dauphine.fr/fileviewer/index.php?doc=2006PA090013.
Full textOur thesis elaborates on the design of a scalable distributed database system (SD-DBS). A novel feature of an SD-DBS is the concept of a scalable distributed relational table, a scalable table in short. Such a table accommodates dynamic splits of its segments at SD-DBS storage nodes. A split occurs when an insert makes a segment to overflow, like in, e. G. , B-tree file. Current DBMSs provide the static partitioning only, requiring a cumbersome global reorganization from time to time. The transparency of the distribution of a scalable table is in this light an important step beyond the current technology. Our thesis explores the design issues of an SD-DBS, by constructing a prototype termed SD-SQL Server. As its name indicates, it uses the services of SQL-Server. SD-SQL Server repartitions a table when an insert overflows existing segments. With the comfort of a single node SQL Server user, the SD-SQL Server user has larger tables or a faster response time through the dynamic parallelism. We present the architecture of our system, its implementation and the performance analysis
Ravat, Franck. "Od3 : contribution méthodologique à la conception de bases de données orientées objet réparties." Toulouse 3, 1996. http://www.theses.fr/1996TOU30150.
Full textBergougnoux, Patrick. "MIME, un environnement de développement coopératif pour applications distribuées." Toulouse 3, 1992. http://www.theses.fr/1992TOU30014.
Full textHatimi, Mostafa. "Gestion des données dupliquées dans un environnement sujet aux partitionnements du réseau de communication." Montpellier 2, 1990. http://www.theses.fr/1990MON20133.
Full textNaacke, Hubert. "Modèle de coût pour médiateur de bases de données hétérogènes." Versailles-St Quentin en Yvelines, 1999. http://www.theses.fr/1999VERS0013.
Full textLes systemes distribues accedent a des sources d'informations diverses au moyen de requetes declaratives. Une solution pour resoudre les problemes lies a l'heterogeneite des sources repose sur l'architecture mediateur / adaptateurs. Dans cette architecture, le mediateur accepte en entree une requete de l'utilisateur, la traite en accedant aux sources via les adaptateurs concernes et renvoie la reponse a l'utilisateur. Le mediateur offre une vue globale et centralisee des sources. Les adaptateurs offrent un acces uniforme aux sources, au service du mediateur. Pour traiter une requete de maniere efficace, le mediateur doit optimiser le plan decrivant le traitement de la requete. Pour cela, plusieurs plans semantiquement equivalents sont envisages, le cout (i. E. Le temps de reponse) de chaque plan est estime afin de choisir celui de moindre cout qui sera execute. Le mediateur estime le cout des operations traitees par les sources en utilisant les informations de cout que les sources exportent. Or, a cause de l'autonomie des sources, les informations exportees peuvent s'averer insuffisantes pour estimer le cout des operations avec une precision convenable. Cette these propose une nouvelle methode permettant au developpeur d'adaptateur d'exporter un modele de cout d'une source a destination du mediateur. Le modele exporte contient des statistiques qui decrivent les donnees stockees dans la source ainsi que des fonctions mathematiques pour evaluer le cout des traitements effectues par la source. Lorsque le developpeur d'adaptateur manque d'information ou de moyen, il a la possibilite de fournir un modele de cout partiel qui est automatiquement complete avec le modele generique predefini au sein du mediateur. Nous validons experimentalement le modele de cout propose en accedant a des sources web. Cette validation montre l'efficacite du modele de cout generique ainsi que celle des modeles plus specialises selon les particularites des sources et les cas d'applications
Le, Sergent Thierry. "Méthodes d'exécution et machines virtuelles parallèles pour l'implantation distribuée du langage de programmation parallèle LCS." Toulouse 3, 1993. http://www.theses.fr/1993TOU30021.
Full textAcosta, Francisco. "Les arbres balances : spécification, performances et contrôle de concurrence." Montpellier 2, 1991. http://www.theses.fr/1991MON20201.
Full textSteff, Yann. "SMA et gestion coopérative de réseaux et systèmes : un cadre méthodologique pour une macro-organisation autonome." Toulouse 3, 2002. http://www.theses.fr/2002TOU30043.
Full textNicolle, Cécile. "Système d'Accès à des Bases de Données Hétérogènes réparties en vue d'une aide à la décision (SABaDH)." Lyon, INSA, 2001. http://theses.insa-lyon.fr/publication/2001ISAL0076/these.pdf.
Full textSince all time, for decision making, decider had to be faced with access problem of all needed data to take the better decision. Nowadays, most systems provide help for this decision making. But it's always difficult to know where the decider can find relevant data. Furthermore, decider can't know type of all data which he need to make his decision. That's why we propose an architecture of an access system which allows decider ask his request in language like natural language, without more detail about their location. Our system can find this data, and provides all information in relation with searched data, these information being relevant. Our system can alleviate some deficiency about search domain. Our system uses wrapper principle, and XML as internal language and request and answer language. Two prototype have been realised, one about search in legal texts base, the other about XML interrogation of Progress base with answer in XML
Mokadem, Riad. "Signatures algébriques dans la gestion de structures de données distribuées et scalables." Paris 9, 2006. https://portail.bu.dauphine.fr/fileviewer/index.php?doc=2006PA090014.
Full textRecent years saw emergence of new architectures, involving multiple computers. New concepts were proposed. Among most popular are those of a multicomputer or of a Network of Worksattion and more recently, of Peer to Peer and Grid Computing. This thesis consists on the design, implementation and performance measurements of a prototype SDDS manager, called SDDS-2005. It manages key based ordered files in distributed RAM of Windows machines forming a grid or P2P network. Our scheme can backup the RAM on each storage node onto the local disk. Our goal is to write only the data that has changed since the last backup. We interest also to update records and non key search (scans). Their common denominator was some application of the properties of new signature scheme based that we call algebraic signatures, which are useful in this context. Ones needs then to find only the areas that changed in the bucket since the last buckup. Our signature based scheme for updating records at the SDDS client should prove its advantages in client-server based database systems in general. It holds the promise of interesting possibilities for transactional concurrency control, beyond the mere avoidance of lost updates. We also update only data have been changed because of the using the algebraic signatures. Also, partly pre-computed algebraic signature of a string encodes each symbol by its cumulative signatures. They protect the SDDS data against incidental viewing by an unauthorized server’s administrator. The method appears attractive, it does not amply any storage overhead. It is also completly transparent for servers and occurs in client. Next, our cheme provide fast string search (match) directly on encoded data at the SDDS servers. They appear an alternative to known Karp-Rabin type schemes. Scans can explore the storage nodes in parallel. They match the records by entire non-key content or by its substring, prefix, longest common prefix or longest common string. The search complexity is almost O (1) for prefix search. One may use them also to detect and localize the silent corruption. These features should be of interest to P2P and grid computing. Then, we propose novel string search algorithm called n-Gramme search. It also appears then among the fastest known, e. G, probably often the faster one we know. It cost only a small fraction of existing records match, especially for larger strings search. The experiments prove high efficiency of our implementation. Our buckup scheme is substantially more efficient with the algebraic signatures. The signature calculus is itself substantially faster, the gain being about 30 %. Also, experiments prove that our cumulative pre-computing notably accelerates the string searchs which are faster than the partial one, at the expense of higher encoding/decoding overhead. They are new alternatives to known Karp-Rabin type schemes, and likely to be usually faster. The speed of string matches opens interesting perspectives for the popular join, group-by, rollup, and cube database operations. Our work has been subject of five publications in international conferences [LMS03, LMS05a, LMS05b, ML06, l&al06]. For convenience, we have included the latest publications. Also, the package termed SDDS-2005 is available for non-commercial use at http://ceria. Dauphine. Fr/. It builds up on earlier versions of the prototype, a cumulative effort of several folks and n-Gramme algorithm implementation. We have also presented our proposed prototype, SDDS-2005, at the Microsoft Research Academic Days 2006
Ghassany, Mohamad. "Contributions à l'apprentissage collaboratif non supervisé." Paris 13, 2013. http://www.theses.fr/2013PA132041.
Full textThe research outlined in this thesis concerns the development of collaborative clustering approaches based on topological methods, such as self-organizing maps (SOM), generative topographic mappings (GTM) and variational Bayesian GTM (VBGTM). So far, clustering methods performs on a single data set, but recent applications require data sets distributed among several sites. So, communication between the different data sets is necessary, while respecting the privacy of every site, i. E. Sharing data between sites is not allowed. The fundamental concept of collaborative clustering is that the clustering algorithms operate locally on individual data sets, but collaborate by exchanging information about their findings. The strength of collaboration, or confidence, is precised by a parameter called coefficient of collaboration. This thesis proposes to learn it automatically during the collaboration phase. Two data scenarios are treated in this thesis, referred as vertical and horizontal collaboration. The vertical collaboration occurs when data sets contain different objects and same patterns. The horizontal collaboration occurs when they have same objects and described by different Patterns
Vilarem, Jean-François. "Contrôle de concurrence mixte en environnement distribué : une méthode fusionnant verrouillage et certification." Montpellier 2, 1989. http://www.theses.fr/1989MON20023.
Full textAl, King Raddad. "Localisation de sources de données et optimisation de requêtes réparties en environnement pair-à-pair." Toulouse 3, 2010. http://thesesups.ups-tlse.fr/912/.
Full textDespite of their great success in the file sharing domain, P2P systems support only simple queries usually based on looking up a file by using its name. Recently, several research works have made to extend P2P systems to be able to share data having a fine granularity (i. E. Atomic attribute) and to process queries written with a highly expressive language (i. E. SQL). The characteristics of P2P systems (e. G. Large-scale, node autonomy and instability) make impractical to have a global catalog that stores often information about data, schemas and data source hosts. Because of the absence of a global catalog, two problems become more difficult: (i) locating data sources with taking into account the schema heterogeneity and (ii) query optimization. In our thesis, we propose an approach for processing SQL queries in a P2P environment. To solve the semantic heterogeneity between local schemas, our approach is based on domain ontology and on similarity formulas. As for the structural heterogeneity of local schemas, it is solved by the extension of a query routing method (i. E. Chord protocol) with Structure Indexes. Concerning the query optimization problem, we propose to take advantage of the data source localization phase to obtain all metadata required for generating a close to optimal execution plan. Finally, in order to show the feasibility and the validity of our propositions, we carry out performance evaluations and we discuss the obtained results
Everaere, Patricia. "Contribution à l'étude des opérateurs de fusion : manipulabilité et fusion disjonctive." Artois, 2006. http://www.theses.fr/2006ARTO0402.
Full textPropositional merging operators aim at defining the beliefs/goals of a group of agents from their individual beliefs/goals, represented by propositional formulae. Two widely used criteria for comparing existing merging operators are rationality and computational complexity. Our claim is that those two criteria are not enough, and that a further one has to be considered as well, namely strategy-proofness. A merging operator is said to be non strategy-proof if there is an agent involved in the merging process who can change the result of the merging, so as to make it closer to her expected one, by lying on her true beliefs/goals. A non strategy-proof merging operator does not give any guarantee that the results it provides are adequate to the beliefs/goals of the group, since it does not incite the agents to report their true beliefs/goals. A first contribution of this thesis consists of a study of the strategy-proofness of existing propositional merging operators. It shows that no existing merging operators fully satisfy the three criteria under consideration: rationality, complexity and strategy-proofness. Our second contribution consists of two new families of disjunctive merging operators, i. E. , operators ensuring that the result of the merging process entails the disjunction of the information given at start. The operators from both families are shown as valuable alternatives to formula-based merging operators, which are disjunctive, but exhibit a high computational complexity, are not strategy-proof, and are not fully rational
Jouis, Christophe. "Contributions à la conceptualisation et à la Modélisation des connaissances à partir d'une analyse linguistique de textes : réalisation d'un prototype : le système SEEK." Paris, EHESS, 1993. http://www.theses.fr/1993EHES0051.
Full textWe present a linguistic and computer model the aim of which is the understanding of linguistic items inserted in their context. This model is constituted by knowledge based systems of contextual exploration which consists in seeking linguistic clues in texts. It is shown that nothing else than a basic morpho-syntactical analysis and the use of the context of an examined linguistic item is required to build semantic representations. These contextual data express a knowledge of the language without using any other knowledge of the world. We present in details a program based on this model : seek. It is a help tool for knowledge extraction forme texts in natural language. This latter has been integrated in a workshop of cognitive engineering associated to a methodology of knowledge acquisition and modelling called metodac
Driouche, Mohamed. "Un système de gestion de base de données réparties dans un environnement temps réel." Paris 6, 1989. http://www.theses.fr/1989PA066730.
Full textLegtchenko, Sergey. "Adaptation dynamique des architectures réparties pour jeux massivement multijoueurs." Phd thesis, Université Pierre et Marie Curie - Paris VI, 2012. http://tel.archives-ouvertes.fr/tel-00931865.
Full textLoukil, Adlen. "Méthodologies, Modèles et Architectures de Référence pour la Gestion et l'Echange de Données Médicales Multimédia : Application aux Projets Européen OEDIPE et BRITER." Lyon, INSA, 1997. http://www.theses.fr/1997ISAL0016.
Full textInterchange and Integration of medical data is a fundamental task in modern medicine. However, a significant obstacle to the development of efficient interoperable information systems is the lack of software tools that provide transparent access to heterogeneous distributed databases. Currently most of the solutions are stand-alone ones fitting only one configuration. To solve this problems of integration and interoperability, we propose in this thesis an original approach which is based on the definition of communication protocols and the design of generic interface between the specific implementations of the protocols and the target databases associated to the Hospital Information Systems. The proposed solution is based on the development of a data dictionary modelling the communications protocols and the databases structures and generic module for the data storage and extraction. The design involves issues related to reverse engineering procedures and to automatic generation of SQL statements. To illustrate this approach, we present the demonstration prototype we have developed in the framework of the OEDIPE AIM project to experiment and to test open interchange of ECGs and associated clinical data. The second part is devoted to the modelling and integration of distributed electronic patient records using communications protocols. We first present a multidimensional approach for the structuring of patient records and propose a generic object oriented information model which integrates bio signals, images and accompanying clinical information. We then, describe a prototype system which has been developed in the framework of the BRITER AIM project for accessing and handling heterogeneous patient data stored in distributed electronic patient records in order to support Rehabilitation healthcare professional in making decisions. We thus demonstrate that the use of standard communications protocols allows and facilitate the development of portable and interoperable medical applications for the benefit of the health care field
Cazalens, Sylvie. "Formalisation en logique non standard de certaines méthodes de raisonnement pour fournir des réponses coopératives, dans des systèmes de bases de données et de connaissances." Toulouse 3, 1992. http://www.theses.fr/1992TOU30172.
Full textFauré, Fabienne. "Gestion de configuration et migration dans les systèmes coopératifs : une architecture répartie orientée services." Toulouse 3, 1994. http://www.theses.fr/1994TOU30253.
Full textBonnel, Nicolas Achille Jacques. "Adapnet : stratégies adaptatives pour la gestion de données distribuées sur un réseau pair-a pair." Lorient, 2008. http://www.theses.fr/2008LORIS134.
Full textIn the last few years, the amount of digital information produced has exponentially increased. This raises problems regarding the storage, the access and the availability of this data. Software and hardware architectures based on the peer-to-peer (p2p) paradigm seem to satisfy the needs of data storage but cannot handle efficiently both data accessibility and availability. We present ,in this thesis various contributions on p2p architectures for managing large volumes of information. We propose various strategies that operate on dedicated virtual topologies that can be maintained at low cost. More precisely, these topologies scale well because the cost for node arrival and node departure is on average constant, whatever the size of the network. We analyze the main paradigms of information sharing on a p2p network, considering successively the problem of access to typed information (semi-structured) and the general case that completely separates the nature of the queries and data location. We propose a routing strategy using structure and content of semi-structured information. We also propose strategies that efficiently explore the network when there is no assumption on the nature of data or queries. In order to manage a quality of service (which is expressed ln terms of speed and reliability), we, also investigate the problem of information availability, more precisely we replicate data stored ln the network. We propose a novel approach exploiting an estimation of local density of data
Longueville, Véronique. "Modélisation, calcul et évaluation de liens pour la navigation dans les grands ensembles d'images fixes." Toulouse 3, 1993. http://www.theses.fr/1993TOU30149.
Full textDuque, Hector. "Conception et mise en oeuvre d'un environnement logiciel de manipulation et d'accès à des données réparties : application aux grilles d'images médicales : le système DSEM / DM2." Lyon, INSA, 2005. http://theses.insa-lyon.fr/publication/2005ISAL0050/these.pdf.
Full textOur vision, in this thesis, is the one of a bio-medical grip as a partner of hospital's information systems, sharing computing resources as well as a platform for sharing information. Therefore, we aim at (i) providing transparent access to huge distributed medical data sets, (ii) querying these data by their content, and (iii), sharing computing resources within the grip. Assuming the existence of a grip infrastructure, we suggest a multi-layered architecture (Distributed Systems Engines – DSE). This architecture allows us to design High Performance Distributed Systems which are highly extensible, scalable and open. It ensures the connection between the grip, data storing systems, and medical platforms. The conceptual design of the architecture assumes a horizontal definition for each one of the layers, and is based on a multi-process structure. This structure enables the exchange of messages between processes by using the Message Passing Paradigm. These processes and messages allow one to define entities of a higher level of semantic significance, which we call Drivers and, which instead of single messages, deal with different kinds of transactions: queries, tasks and requests. Thus, we define different kinds of drivers for dealing with each kind of transaction, and in a higher level, we define services as an aggregation of drivers. The architectural framework of drivers and services eases the design of components of a Distributed System (DS), which we call engines, and also eases the extensibility and scalability of DS
Benzine, Mehdi. "Combinaison sécurisée des données publiques et sensibles dans les bases de données." Versailles-St Quentin en Yvelines, 2010. http://www.theses.fr/2010VERS0024.
Full textProtection of sensitive data is a major issue in the databases field. Many software and hardware solutions have been designed to protect data when stored and during query processing. Moreover, it is also necessary to provide a secure manner to combine sensitive data with public data. To achieve this goal, we designed a new storage and processing architecture. Our solution combines a main server that stores public data and a secure server dedicated to the storage and processing of sensitive data. The secure server is a hardware token which is basically a combination of (i) a secured microcontroller and (ii) a large external NAND Flash memory. The queries which combine public and sensitive data are split in two sub queries, the first one deals with the public data, the second one deals with the sensitive data. Each sub query is processed on the server storing the corresponding data. Finally, the data obtained by the computation of the sub query on public data is sent to the secure server to be mixed with the result of the computation on sensitive data. For security reasons, the final result is built on the secure server. This architecture resolves the security problems, because all the computations dealing with sensitive data are done by the secure server, but brings performance problems (few RAM, asymmetric cost of read/write operations. . . ). These problems will be solved by different strategies of query optimization
Desprats, Thierry. "Conception des systèmes coopératifs : maîtrise de la complexité par modélisation de l'interaction." Toulouse 3, 1993. http://www.theses.fr/1993TOU30167.
Full textBavueza, Munsana Dia Lemfu. "Ravir : un système de coopération des bases de données hétérogènes." Montpellier 2, 1987. http://www.theses.fr/1987MON20265.
Full textFaye, David Célestin. "Médiation de données sémantique dans SenPeer, un système pair-à-pair de gestion de données." Phd thesis, Université de Nantes, 2007. http://tel.archives-ouvertes.fr/tel-00481311.
Full textSauquet, Dominique. "Lied : un modèle de données sémantique et temporel : son intégration dans une architecture distribuée et son utilisation pour des applications médicales." Châtenay-Malabry, Ecole centrale de Paris, 1998. http://www.theses.fr/1998ECAP0586.
Full textBenslimane, Djamal. "Etudes de l'apport des techniques de parallélisme dans l'amélioration des performances des systèmes à base de règles de production." Clermont-Ferrand 2, 1990. http://www.theses.fr/1990CLF21287.
Full textLobry, Olivier. "Support Mémoire Adaptable Pour Serveurs de Données Répartis." Phd thesis, Université Joseph Fourier (Grenoble), 2000. http://tel.archives-ouvertes.fr/tel-00346893.
Full textIl n'est malheureusement pas possible d'offrir un serveur de données universel capable de répondre aux exigences de tous les SI. Ceux-ci diffèrent en effet significativement par le type des informations qu'ils traitent, la nature des traitements effectués, les propriétés de traitement qu'ils garantissent, les caractéristiques du matériel sous-jacent, etc. De ce fait, chaque système d'information intègre son ou ses propres serveurs de données implantant des politiques de gestion figées.
Les inconvénients d'une telle approche sont loin d'être négligeables. Tout d'abord, la ré-implantation de mécanismes élémentaires de gestion de ressources augmente le coût de conception. Ensuite, la rigidité comportementale réduit considérablement la réactivité à l'évolution tant en qualité qu'en quantité des informations, traitements et ressources matérielles. Enfin, l'opacité des tels systèmes rend difficile leur coexistence sur une même plate-forme.
Cette thèse montre qu'il n'existe pas de politique de gestion de la mémoire idéale. Plutôt que d'essayer d'offrir un serveur idéal, elle tente de définir une infrastructure permettant de concevoir des serveurs de données adaptés et évolutifs. Elle adresse plus particulièrement le problème de la gestion de la mémoire physique et se place dans le contexte des grappes de machines. Elle propose le support mémoire adaptable ADAMS basé sur un modèle de gestion hiérarchique et un modèle de communication par événements. Ce support facilite l'intégration de différents types de politiques tout en séparant bien leurs rôles respectifs sans faire d'hypothèse sur leur inter-dépendances.
Une intégration d'ADAMS à la mémoire virtuelle répartie permanente et adaptable du système Arias est ensuite exposée. ADAMS étend les caractéristiques de ce système afin de prendre en compte les besoins particulier de gestion des serveurs de données tout en réduisant le grain d'adaptabilité. Nous illustrons à travers un exemple comment le support résultant permet d'implanter un serveur de données dont les politiques de gestion peuvent être adaptées dynamiquement.
Bekele, Dawit. "Contribution à l'étude de la répartition d'applications écrites en langage ADA 83." Toulouse 3, 1994. http://www.theses.fr/1994TOU30069.
Full textPons, Jean-François. "Contrôle de la cohérence des accès aux objets dans les systèmes répartis : application des règles d'écriture recouverte." Montpellier 2, 1986. http://www.theses.fr/1986MON20071.
Full textGolenetskaya, Natalia. "Adressing scaling challenges in comparative genomics." Phd thesis, Université Sciences et Technologies - Bordeaux I, 2013. http://tel.archives-ouvertes.fr/tel-00865840.
Full textLarios, Rosillo Victor. "Stratégies en ligne pour une qualité de service satisfaisante dans un système de réalité virtuelle répartie." Compiègne, 2001. http://www.theses.fr/2001COMP1335.
Full textNicolas, Jean-Christophe. "Machines bases de données parallèles : contribution aux problèmes de la fragmentation et de la distribution." Lille 1, 1991. http://www.theses.fr/1991LIL10025.
Full textGrazziottin, Ribeiro Helena. "Un service de règles actives pour fédérations de bases de données." Université Joseph Fourier (Grenoble), 2000. http://www.theses.fr/2000GRE10084.
Full textJean, Sébastien. "Modèles et architectures d'interaction interne et externe pour cartes à microprocesseur ouvertes." Lille 1, 2001. https://pepite-depot.univ-lille.fr/RESTREINT/Th_Num/2001/50376-2001-329.pdf.
Full textNous présentons le résultat de recherches portant sur les deux derniers types d'arrangements et poursuivant deux objectifs. Nous souhaitions, d'une part, définir un cadre cohérent pour l'intégration de cartes ouvertes réactives au sein de systèmes d'information répartis (i. E. Où les applications encartées peuvent être à la fois clientes et serveurs) et, d'autre part, pouvoir disposer d'un support de coopération interne permettant un partage fin, contrôlé et évolutif, tant des informations que des compétences des applications encartées. Apporter une réponse au premier problème a un impact sur l'ensemble de l'infrastructure logicielle. La solution que nous avons proposée, baptisée AWARE, est le reflet de cette globalité. Elle s'articule autour d'un modèle d'exécution multi-tâches pour cartes ouvertes, où l'ordonnancement des tâches s'effectue depuis l'extérieur de la carte, et s'accompagne d'un modèle de programmation facilitant la conception d'applications réparties intégrant de telles cartes. Nous avons également proposé un modèle de coopération intra-carte basé sur quelques-uns des principes des bases de données relationnelles
El, Attar Ali. "Estimation robuste des modèles de mélange sur des données distribuées." Phd thesis, Nantes, 2012. https://archive.bu.univ-nantes.fr/pollux/show/show?id=b22726f5-f19e-4c6e-9dcb-451bb7a968e8.
Full textThis work proposes a contribution aiming at probabilistic model estimation, in the setting of distributed, decentralized, data-sharing computer systems. Such systems are developing over the internet, and also exist as sensor networks, for instance. Our general goal consists in estimating a probability distribution over a data set which is distributed into subsets located on the nodes of a distributed system. More precisely, we are at estimating the global distribution by aggregating local distributions, estimated on these local subsets. Our proposal exploits the following assumption: all distributions are modelled as a Gaussian mixture. Our contribution is a solution that is both decentralized and statistically robust to outlier local Gaussian mixture models. The proposed process only requires mixture parameters, rather than original data
El, Attar Ali. "Estimation robuste des modèles de mélange sur des données distribuées." Phd thesis, Université de Nantes, 2012. http://tel.archives-ouvertes.fr/tel-00746118.
Full textDaviet, Stéphane. "Étude du comportement humain grâce à la simulation multi-agents et aux méthodes de fouille de données temporelles." Nantes, 2009. http://www.theses.fr/2009NANT2018.
Full textSimulation tools has already proven its efficiency for studying numerous phenomena, but simulating human behavior is still a challenge for both computer and human sciences. In this context, we have worked on the simulation of individual demeanor and social interactions in human groups to study the emergence of collective beahvioral patterns, mixing techniques of artificial intelligence, multi-agent systems, psychological sciences and spatio-temporal data mining. This thesis presents a new model of emotional agent: the Emotion, Feeling, Temperament agent (EFT). Based on BDI architecture, our model integrates the OCC emotional model and the PerformanSe behavioral model. We also present a concrete implementation of our model: the simulation of brain-damaged people’s behavior on a production line. Then, we describe the modeling of the interactions of this system with AgentUML. We mine data produced by the simulation through knowledge discovery techniques applied to temporal data
Touzet, David. "Interrogation continue des systèmes d'information de proximité." Rennes 1, 2004. http://www.theses.fr/2004REN10007.
Full textEl, Merhebi Souad. "La gestion d'effet : une méthode de filtrage pour les environnements virtuels répartis." Toulouse 3, 2008. http://thesesups.ups-tlse.fr/243/1/El_Merhebi_Souad.pdf.
Full textDistributed virtual environments (DVEs) are intended to provide an immersive experience to their users within a shared virtual environment. For this purpose, DVEs try to supply participants with coherent views of the shared world. This requires a heavy message exchange between participants especially with the increasing popularity of massively multiplayer DVEs. This heavy message exchange consumes a lot of processing power and bandwidth, slowing down the system and limiting interactivity. Indeed, coherence, interactivity and scalability are basic requirements of DVEs. However, these requirements are conflicting because coherence requires the more important exchange of messages that we can have while interactivity and scalability demand to decrease this exchange to minimum. For this reason, the management of message exchange is essential for distributed virtual environments. To manage message exchange in an intelligent way, DVE systems use various filtering techniques. Among them, interest management techniques filter messages according to users' interests in the world. In this document, we present our interest management technique, the effect management. This technique expresses the interests and manifestations of participants in various media through conscience and effect zones. When the conscience zone of a participant collides the effect zone of another participant in a given medium, the first one becomes conscious of the second. ). .
Soutou, Christian. "Contribution à la conception d'une base de données fédérée : dérivation, évolution et intégration de schémas." Toulouse 3, 1994. http://www.theses.fr/1994TOU30187.
Full textPark, Young-Min. "Réseau virtuel : développement d'un système semi-réparti de gestion de bases de données non limitées sur le Web." Paris 8, 2004. http://www.theses.fr/2004PA082341.
Full textNotre projet porte sur la notion nouvelle de " réseau virtuel ". Cette notion peut être comparée à la notion de " réseau classique " ou de " réseau traditionnel ". Elle est utile surtout pour les gros réseaux et elle constitue l'une des meilleures solutions pour envisager le remplacement des réseaux classiques. D'un point de vue économique, le réseau classique oblige souvent à un important investissement. Plus la taille du réseau est grande, plus il nécessite de grosses dépenses liées au nombre des ordinateurs, des stations de travail, des serveurs et des personnes qui administrent et surveillent son fonctionnement. Le " réseau virtuel " fonctionne en utilisant le réseau déjà existant de l'Internet. A la place de nombreux serveurs et logiciels, le " réseau virtuel " se construit avec un serveur d'applications unique et plusieurs serveurs de bases de données qui ensemble effectuent les tâches que le réseau classique prend en charge. Le premier des avantages du " réseau virtuel " est qu'il est beaucoup moins coûteux que le réseau classique. Il utilise l'environnement d'Internet comme moyen de liaison entre les ordinateurs. Aucun autre matériel n'est exigé. Le second avantage est qu'il est plus facile de préserver la compatibilité et la cohérence des données. Le " réseau virtuel " utilise les mêmes applications pour toutes les machines connectées et les données. Le but de notre projet est de concevoir pour le " réseau virtuel ", la structure la plus efficace liant ses serveurs, de développer des applications souples, capables de traiter des bases de données de toutes tailles, et de mettre au point une stratégie de séparation des serveurs de bases de données permettant d'atteindre la capacité de stockage souhaitée