To see the other types of publications on this topic, follow the link: Component architecture.

Dissertations / Theses on the topic 'Component architecture'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Component architecture.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Banda, Ravi S. "Interoperable components across multiple component architectures." Morgantown, W. Va. : [West Virginia University Libraries], 1998. http://etd.wvu.edu/templates/showETD.cfm?recnum=232.

Full text
Abstract:
Thesis (M.S.)--West Virginia University, 1998.
Title from document title page. Document formatted into pages; contains vi, 53 p. : ill. Vita. Includes abstract. Includes bibliographical references (p. 50-51).
APA, Harvard, Vancouver, ISO, and other styles
2

Schmid, Stefan. "A component-based active router architecture." Thesis, Lancaster University, 2003. http://eprints.lancs.ac.uk/12227/.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Leeb, Angelika. "A flexible object architecture for component software." Thesis, Massachusetts Institute of Technology, 1996. http://hdl.handle.net/1721.1/40229.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Lewis, Joshua. "A COMMON COMPONENT-BASED SOFTWARE ARCHITECTURE FOR MILITARY AND COMMERCIAL PC-BASED VIRTUAL SIMULATION." Doctoral diss., University of Central Florida, 2006. http://digital.library.ucf.edu/cdm/ref/collection/ETD/id/2301.

Full text
Abstract:
Commercially available military-themed virtual simulations have been developed and sold for entertainment since the beginning of the personal computing era. There exists an intense interest by various branches of the military to leverage the technological advances of the personal computing and video game industries to provide low cost military training. By nature of the content of the commercial military-themed virtual simulations, a large overlap has grown between the interests, resources, standards, and technology of the computer entertainment industry and military training branches. This research attempts to identify these commonalities with the purpose of systematically designing and evaluating a common component-based software architecture that could be used to implement a framework for developing content for both commercial and military virtual simulation software applications.
Ph.D.
Other
Engineering and Computer Science
Modeling and Simulation
APA, Harvard, Vancouver, ISO, and other styles
5

Troup, Timothy J. "A component system architecture to enable user-directed component binding at run-time." Thesis, University of Glasgow, 2004. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.414045.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Jung, Georg. "Structured interrelations of component architectures." Diss., Manhattan, Kan. : Kansas State University, 2007. http://hdl.handle.net/2097/485.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Göbel, Steffen, Christoph Pohl, Ronald Aigner, Martin Pohlack, Simone Röttger, and Steffen Zschaler. "The COMQUAD Component Container Architecture and Contract Negotiation." Saechsische Landesbibliothek- Staats- und Universitaetsbibliothek Dresden, 2012. http://nbn-resolving.de/urn:nbn:de:bsz:14-qucosa-100181.

Full text
Abstract:
Component-based applications require runtime support to be able to guarantee non-functional properties. This report proposes an architecture for a real-time-capable, component-based runtime environment, which allows to separate non-functional and functional concerns in component-based software development. The architecture is presented with particular focus on three key issues: the conceptual architecture, an approach including implementation issues for splitting the runtime environment into a real-time-capable and a real-time-incapable part, and details of contract negotiation. The latter includes selecting component implementations for instantiantion based on their non-functional properties.
APA, Harvard, Vancouver, ISO, and other styles
8

Beckenkamp, Fábio Ghignatti. "A component architecture for artificial neural network systems." [S.l. : s.n.], 2002. http://deposit.ddb.de/cgi-bin/dokserv?idn=964923580.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Eleyan, Amna. "Extending web service architecture with a quality component." Thesis, University of Manchester, 2007. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.511251.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Ong, Elwin 1979. "Fault protection in a component-based spacecraft architecture." Thesis, Massachusetts Institute of Technology, 2003. http://hdl.handle.net/1721.1/82804.

Full text
APA, Harvard, Vancouver, ISO, and other styles
11

Ouimet, Martin 1975. "A diagnostics architecture for component-based system engineering." Thesis, Massachusetts Institute of Technology, 2004. http://hdl.handle.net/1721.1/17787.

Full text
Abstract:
Thesis (S.M.)--Massachusetts Institute of Technology, Dept. of Aeronautics and Astronautics, 2004.
Includes bibliographical references (leaves 58-60).
This work presents an approach to diagnosis to meet the challenging demands of modern engineering systems. The proposed approach is an architecture that is both hierarchical and hybrid. The hierarchical dimension of the proposed architecture serves to mitigate the complexity challenges of contemporary engineering systems. The hybrid facet of the architecture tackles the increasing heterogeneity of modern engineering systems. The architecture is presented and realized using a bus representation where various modeling and diagnosis approaches can coexist. The proposed architecture is realized in a simulation environment, the Specification Toolkit and Requirements Methodology (SpecTRM). This research also provides important background information concerning approaches to diagnosis. Approaches to diagnosis are presented, analyzed, and summarized according to their strengths and domains of applicability. Important characteristics that must be considered when developing a diagnostics infrastructure are also presented alongside design guidelines and design implications. Finally, the research presents important topics for further research.
by Martin Ouimet.
S.M.
APA, Harvard, Vancouver, ISO, and other styles
12

Göbel, Steffen, Christoph Pohl, Ronald Aigner, Martin Pohlack, Simone Röttger, and Steffen Zschaler. "The COMQUAD Component Container Architecture and Contract Negotiation." Technische Universität Dresden, 2004. https://tud.qucosa.de/id/qucosa%3A26291.

Full text
Abstract:
Component-based applications require runtime support to be able to guarantee non-functional properties. This report proposes an architecture for a real-time-capable, component-based runtime environment, which allows to separate non-functional and functional concerns in component-based software development. The architecture is presented with particular focus on three key issues: the conceptual architecture, an approach including implementation issues for splitting the runtime environment into a real-time-capable and a real-time-incapable part, and details of contract negotiation. The latter includes selecting component implementations for instantiantion based on their non-functional properties.
APA, Harvard, Vancouver, ISO, and other styles
13

Adhikari, Rabindra. "The pursuit of sustainability : architecture and component based design." Virtual Press, 2007. http://liblink.bsu.edu/uhtbin/catkey/1370875.

Full text
Abstract:
Architectural sustainability is ambiguous. The contemporary understanding of sustainability is often debated on the basis of environmental and humanitarian focus. This focus often brings ambiguity when the concept of sustainability is promoted as an end product and not as a perpetual process. For ecologically harmonious and sustainable human survival, sustainable methods and processes are believed to encourage reworking the course of human development.The Component-based design approach is a promising and appropriate method for achieving sustainability. The method involves systems thinking and illustrates characteristics like flexibility, resilience, adaptability and generative system. The focus of design should be in making relationships among these various components and processes, rather than emphasis on end product. This creative project illustrates these conceptual processes using computational logic to create a project for graphics and multimedia presentation.
Department of Architecture
APA, Harvard, Vancouver, ISO, and other styles
14

Ong, David C. C. "An intelligent multi-component distributed architecture for knowledge management." Thesis, Kingston University, 2009. http://eprints.kingston.ac.uk/20304/.

Full text
Abstract:
The aim of this thesis is to propose an integrated generic intelligent decision-making framework that can be employed in the design and construction of computing infrastructures where high flexibility and dynamism are essential. In fact, the main problem with many decision making systems is that they are designed for specific purposes which make them unsuitable to be deployed in a complex system, where the level of unknown / uncertainties is high. The proposed framework is generic enough to address this limitation as it could be deployed across different computing architectures or systems, or redeployed to serve a particular purpose. The research study starts with the proposal of two theoretical concepts as part of an intelligent information management approach for a new integrated intelligent decision-making framework. The first concentrates on the thinking and learning processes to achieve the best effort decision via logical reasoning strategies. It determines the best execution path under particular circumstances in a given computing environment. The second concept focuses on data capturing techniques using distributed sensing devices which act as sensors for a decision-making unit (i.e. an input / output (IO) interface for thinking and learning processes). A model to describe perceived sensory perception is proposed, as well as an observation technique to monitor the proposed model. These concepts are then translated into an intelligent decision-making framework, which is capable of interpreting and manipulating available information to offer the best effort solution based on available resources, rather than relying heavily on additional powerful physical resources to provide a precise solution. Therefore, the accuracy and precision of decision-making depends on the applied logic and learning processes. Indirectly, this framework attempts to solve integration problems related to the aspect of "Intelligent" into practical day-to-day problem solving applications. A working prototype based on the proposed framework was developed and presented for an evaluation, to verify the framework competence in operating with computing infrastructure, and whether it is capable of making sensible decisions upon request, and whether it is able to learn from its decisions via the received feedback. To achieve this, the behaviour of the prototype is accessed against the growth in the number of experiences and amount of knowledge collected during the execution process. Finally, it has concluded that proposed concepts and framework operates well in term of decision-making capabilities and reasoning strategies.
APA, Harvard, Vancouver, ISO, and other styles
15

Su, Jiancheng. "Component-based Intelligent Control Architecture for Reconfigurable Manufacturing Systems." Diss., Virginia Tech, 2007. http://hdl.handle.net/10919/29980.

Full text
Abstract:
The present dynamic manufacturing environment has been characterized by a greater variety of products, shorter life-cycles of products and rapid introduction of new technologies, etc. Recently, a new manufacturing paradigm, i.e. Reconfigurable Manufacturing Systems (RMS), has emerged to address such challenging issues. RMSs are able to adapt themselves to new business conditions timely and economically with a modular design of hardware/software system. Although a lot of research has been conducted in areas related to RMS, very few studies on system-level control for RMS have been reported in literature. However, the rigidity of current manufacturing systems is mainly from their monolithic design of control systems. Some new developments in Information Technology (IT) bring new opportunities to overcome the inflexibility that shadowed control systems for years. Component-based software development gains its popularity in 1990â s. However, some well-known drawbacks, such as complexity and poor real-time features counteract its advantages in developing reconfigurable control system. New emerging Extensible Markup Language (XML) and Web Services, which are based on non-proprietary format, can eliminate the interoperability problems that traditional software technologies are incompetent to accomplish. Another new development in IT that affects the manufacturing sector is the advent of agent technology. The characteristics of agent-based systems include autonomous, cooperative, extendible nature that can be advantageous in different shop floor activities. This dissertation presents an innovative control architecture, entitled Component-based Intelligent Control Architecture (CICA), designed for system-level control of RMS. Software components and open-standard integration technologies together are able to provide a reconfigurable software structure, whereas agent-based paradigm can add the reconfigurability into the control logic of CICA. Since an agent-based system cannot guarantee the best global performance, agents in the reference architecture are used to be exception handlers. Some widely neglected problems associated with agent-based system such as communication load and local interest conflicts are also studied. The experimental results reveal the advantage of new agent-based decision making system over the existing methodologies. The proposed control system provides the reconfigurability that lacks in current manufacturing control systems. The CICA control architecture is promising to bring the flexibility in manufacturing systems based on experimental tests performed.
Ph. D.
APA, Harvard, Vancouver, ISO, and other styles
16

Zou, Li. "A New Architecture for Developing Component-based Distributed Applications." University of Cincinnati / OhioLINK, 2000. http://rave.ohiolink.edu/etdc/view?acc_num=ucin974951548.

Full text
APA, Harvard, Vancouver, ISO, and other styles
17

Admodisastro, Novia. "An integrated architecture analysis framework for component-based software development." Thesis, Lancaster University, 2011. http://eprints.lancs.ac.uk/61630/.

Full text
Abstract:
The importance of architecture in reuse-driven development is widely recognized. Architecture provides a framework for establishing a match between available components and the system context. It is a key part of the system documentation; enforces the integrity of component composition and provides a basis for managing change. However, one of the most difficult problems in component-based system development (CBD) is ensuring that the software architecture provides an acceptable match with its intended application, business and evolutionary context. Unlike custom development where architectural design relies solely on a detailed requirements specification and where deficiencies in application context can be corrected by ‘tweaking’ the source code, in component-based system development the typical unit of development is often a black-box component whose source code is inaccessible to the developer. Getting the architecture right is therefore key to ensuring quality in a component-based system. Architecture analysis in CBD provides the developer with a means to expose interface mismatches, assess configurations with respect to specific structural and behavioural constraints and to verify the adequacy of compositions with respect to quality constraints. However, support for key component-based system design issues is still patchy in most architecture analysis approaches. My solution has been to develop, Component-based Software Architecture analysis FramEwork (CSAFE), a scenario-driven architecture analysis approach that combines and extends the strengths of current approaches using pluggable analysis. CSAFE is process- pluggable and recognises that negotiation (trade-off analysis) is central to black-box software development. However, while CSAFE is primarily intended to support black-box development, we recognise that there may be aspects of the system for which a black-box solution is not feasible. CSAFE supports custom development in such situations by treating abstract components as placeholders for custom development. CSAFE is supported by an extensible toolset.
APA, Harvard, Vancouver, ISO, and other styles
18

Gayard, Leonel Aguilar 1983. "CosmosLoader : uma ferramenta de apoio ao gerenciamento de configuração baseado no modelo Cosmos*." [s.n.], 2013. http://repositorio.unicamp.br/jspui/handle/REPOSIP/275621.

Full text
Abstract:
Orientador: Cecília Mary Fischer Rubira
Dissertação (mestrado) - Universidade Estadual de Campinas, Instituto de Computação
Made available in DSpace on 2018-08-23T21:56:13Z (GMT). No. of bitstreams: 1 Gayard_LeonelAguilar_M.pdf: 1532057 bytes, checksum: 8324b0d9bd3eebedc13daca6f31c7c6b (MD5) Previous issue date: 2013
Resumo: Nos últimos anos, o desenvolvimento baseado em componentes (DBC) e a arquitetura de software emergiram como disciplinas complementares para promover o reuso no desenvolvimento de software. O desenvolvimento baseado em componentes promove o desenvolvimento de componentes reutilizáveis e a formação de novos sistemas de software a partir da integração de componentes existentes. A arquitetura de um sistema de software descreve o sistema em termos de seus componentes arquiteturais, das propriedades destes e das conexões entre eles. Modelos de componentes possibilitam concretizar elementos de DBC como componentes e conectores a partir de conceitos tradicionais de desenvolvimento de software, como por exemplo, orientação a objetos e arquivos, de modo que um conjunto de classes e arquivos complementares podem formar um componente se seguirem as regras do modelo. A adequação a um modelo de componentes oferece benefícios para o sistema de software: por exemplo, o modelo de componentes EJB possibilita que um componente seja implantado em um contêiner e se beneficie do gerenciamento de segurança e transparência de localização oferecidos pelo contêiner; o modelo Cosmos estabelece regras para a criação de componentes baseados em conceitos de linguagens orientadas a objetos, como classes, interfaces e pacotes. Assim, a integração de componentes para a composição de um sistema de software de acordo com uma arquitetura e um modelo de componentes se torna uma atividade importante no desenvolvimento de um novo sistema de software. Uma configuração concreta é um conjunto de determinadas versões de componentes de software conectados de acordo com uma arquitetura. No entanto, observa-se que, se modelos facilitam a criação de componentes de software, ainda é necessário um esforço de desenvolvimento para integrar componentes e formar novos sistemas. O uso de ambientes de desenvolvimento orientados a arquitetura e DBC, e também de ferramentas de automatização, reduzem o esforço necessário para a integração de componentes. O ambiente Bellatrix é um ambiente de desenvolvimento integrado que permite a especificação de elementos arquiteturais e a modelagem de arquiteturas de software. No entanto, ainda existe um hiato entre a arquitetura de um sistema modelada no ambiente Bellatrix e a configuração do sistema a partir de componentes concretos. Esta dissertação apresenta a ferramenta CosmosLoader, que auxilia o gerenciamento de configurações de componentes baseados no modelo Cosmos. A solução proposta se baseia na extensão do ambiente Bellatrix e no modelo de componentes Cosmos* ("Cosmos estrela"), que estende o modelo Cosmos com o conceito de composição hierárquica de componentes. Por fim, são descritos estudos de caso realizados com essas ferramentas
Abstract: In the last years, Component-Based Development (CBD) and Software Architecture emerged as complementary disciplines that promote reuse in software development. Component-Based Development promotes the development of reusable software components and the creation of new software systems by integrating existing software components. The architecture of a software system describes such system in terms of its architectural components, their properties and the connections between them. Component models materialize concepts from CBD such as components and connectors from traditional concepts of software development, such as object orientation and files, so that a set of classes and complementary files form a component if they follow the component model's rules. The adequacy to a component model brings benefits to a software system: for instance, the EJB component model allows a component to be deployed to a container and benefit from the security management and location transparency provided by the container; the Cosmos component model allows the creation of components using only concepts from object-oriented languages, such as classes, interfaces and packages. Therefore, the integration of components to compose a new software system according to an architecture and a component model becomes an important activity in the development of a new software system. A concrete configuration is the set of specific versions of software components connected according to an architecture. However, it can be observed that while models ease the creation of software components, a development effort is still necessary to integrate components and form new systems. The use of development environments oriented towards architecture and CBD, and also of automations tools, reduce the effort needed to integrate components. The Bellatrix development environment is an integrated development environment that allows the specification of architectural elements and modeling of software architectures. However, there still is a gap between the architecture of a system modeled in Bellatrix and the configuration of a system from concrete software components. This dissertation presents the CosmosLoader tool, which assists in managing the configuration of components based on the Cosmos model. The proposed solution is based on an extension to the Bellatrix development environment and on the Cosmos* component model ("Cosmos star"), which extends the Cosmos component model with hierarchical composition of components. Finally, case studies using these tools are described
Mestrado
Ciência da Computação
Mestre em Ciência da Computação
APA, Harvard, Vancouver, ISO, and other styles
19

Aboud, Nour. "Service-Oriented Integration of Component and Organizational MultiAgent Models." Thesis, Pau, 2012. http://www.theses.fr/2012PAUU3019/document.

Full text
Abstract:
Les travaux présentés dans cette thèse concernent des problématiques d'architecture logicielle multi-domaines pour le développement d’applications distribuées. Ces applications sont caractérisées aujourd’hui comme des systèmes ouverts, complexes, hétérogènes et à large échelle. Les approches traditionnelles, telles que l’approche orienté objet, n’offrent plus un paradigme de conception suffisant pour appréhender la complexité de tels systèmes. Ces nouvelles tendances ont conduit à l’émergence d’approches de plus haut niveau telles que les approches orientées services, composants ou agents. Chacune de ces approches offrent des intérêts et des caractéristiques propres dans le développement d’applications distribuées. Les services offrent une abstraction et une interopérabilité à large échelle. Abstraction dans le sens où un service permet de spécifier un élément fonctionnel sans préciser comment cet élément est implémenté. Les composants sont une approche robuste basée sur la composition et la réutilisation d’éléments clairement définis par leurs interfaces. Les agents sont eux des éléments présentant un comportement dynamique dirigé par un but et des interactions de haut niveau avec les autres agents formant l’application, vue comme une organisation de services collaboratifs. D’un point de vue conceptuel, le service peut donc être perçu comme le modèle « métier » de l’application, alors que les composants et les agents constituent un modèle d’implémentation. L’étude de ces différents domaines et des modèles associés, a montré que les approches composants et agents sont complémentaires, les points forts d’une approche représentant les faiblesses de l’autre. Face à ce constat, il nous est paru intéressant d’intégrer ces deux approches, au sein d’une même démarche de conception. Cela permet, d’une part, qu’une approche puisse bénéficier des intérêts de l’autre et d’autre part, d’utiliser conjointement des agents et des composants dans la conception d'une même application. La démarche que nous avons adoptée consiste à considérer les services comme pivot d’interaction afin de rendre possible l’interopérabilité des agents et des composants. Pour supporter cette démarche, nous avons défini un processus de conception basé sur l’Ingénierie Des Modèles qui contient quatre modèles conceptuels (Domain Specific language) dont l’intérêt est de mettre l’accent sur les concepts de services et d’interaction. Nous avons ainsi défini un modèle de services, un modèle de composants et un modèle d’agents. Enfin, un modèle mixte appelé CASOM, Component Agent Service Oriented Model, permet de spécifier une application via une combinaison des trois domaines précédents. Ensuite, des règles de correspondances ont été définies entre les quatre modèles pour pouvoir par exemple transformer une spécification agents en une spécification composants ou mixte. L'implémentation de ces transformations a été réalisée en langage ATL (ATLAS Transformation Language)
The presented work considers problems related to multi-domain software architecture for the development of distributed applications. These applications are large-scaled, heterogeneous, open and complex software systems. Traditional approaches such as object-oriented are no longer sufficient to represent such complex systems. These trends lead to the emergence of higher-level approaches such as service-oriented, components or agents. Each one of these approaches offers interests and characteristics in the development of distributed applications. Services provide an abstraction and interoperability in a large scale. Abstraction is in the sense that a service can specify a functional element without specifying how this element is implemented. The components are a robust approach based on composition and reusability through their clearly defined interfaces. Agents are elements which are characterized by dynamic goal directed behaviours and high-level interactions with other agents forming the application, seen as an organization for collaborative services. From a conceptual point of view, the service can be seen as the “business” model of an application, while components and agents are the implementation models. The study of these different domains, with their related models, showed that the components and agents approaches are complementary; the strengths of one approach overcome the weaknesses of the other. Therefore, we are interested in the integration of these two approaches in a single design approach. This allows an approach to benefit from the interests of the other, on one hand and the use of agents and components jointly in the design of an application on the other hand. To reach our objective, we consider services as pivot of interaction between agents and components. The result of our analysis leads us to develop a design process based on Model-Driven Engineering which contains four conceptual models (Domain Specific Languages) with the main interest of focusing on the concepts of services and interaction. We then defined a service, component and agent models. Finally, a hybrid model called CASOM, Component Agent Service Oriented Model, was proposed that allows application specification via a combination of the three domains. Then, mapping rules have been defined between the four models in order to transform agents specification into components specification or mixed. The implementation of these transformations was done in ATL language (ATLAS Transformation Language)
APA, Harvard, Vancouver, ISO, and other styles
20

Zhu, Liming Computer Science &amp Engineering Faculty of Engineering UNSW. "Software architecture evaluation for framework-based systems." Awarded by:University of New South Wales. Computer Science and Engineering, 2007. http://handle.unsw.edu.au/1959.4/28250.

Full text
Abstract:
Complex modern software is often built using existing application frameworks and middleware frameworks. These frameworks provide useful common services, while simultaneously imposing architectural rules and constraints. Existing software architecture evaluation methods do not explicitly consider the implications of these frameworks for software architecture. This research extends scenario-based architecture evaluation methods by incorporating framework-related information into different evaluation activities. I propose four techniques which target four different activities within a scenario-based architecture evaluation method. 1) Scenario development: A new technique was designed aiming to extract general scenarios and tactics from framework-related architectural patterns. The technique is intended to complement the current scenario development process. The feasibility of the technique was validated through a case study. Significant improvements of scenario quality were observed in a controlled experiment conducted by another colleague. 2) Architecture representation: A new metrics-driven technique was created to reconstruct software architecture in a just-in-time fashion. This technique was validated in a case study. This approach has significantly improved the efficiency of architecture representation in a complex environment. 3) Attribute specific analysis (performance only): A model-driven approach to performance measurement was applied by decoupling framework-specific information from performance testing requirements. This technique was validated on two platforms (J2EE and Web Services) through a number of case studies. This technique leads to the benchmark producing more representative measures of the eventual application. It reduces the complexity behind the load testing suite and framework-specific performance data collecting utilities. 4) Trade-off and sensitivity analysis: A new technique was designed seeking to improve the Analytical Hierarchical Process (AHP) for trade-off and sensitivity analysis during a framework selection process. This approach was validated in a case study using data from a commercial project. The approach can identify 1) trade-offs implied by an architecture alternative, along with the magnitude of these trade-offs. 2) the most critical decisions in the overall decision process 3) the sensitivity of the final decision and its capability for handling quality attribute priority changes.
APA, Harvard, Vancouver, ISO, and other styles
21

Hall, Jeremy Graham. "An examination of the intangible business component of commercial real estate investments." Thesis, Massachusetts Institute of Technology, 1993. http://hdl.handle.net/1721.1/66757.

Full text
APA, Harvard, Vancouver, ISO, and other styles
22

Bühler, Frank Stefan. "Combining visual modelling with visual programming for CORBA component development." Thesis, De Montfort University, 2002. http://hdl.handle.net/2086/4068.

Full text
APA, Harvard, Vancouver, ISO, and other styles
23

Tran, Tam M. Allen James O. "Interoperability and security support for heterogeneous Cots/Gots/legacy component-based architecture." Monterey, Calif. : Springfield, Va. : Naval Postgraduate School ; Available from National Technical Information Service, 2000. http://handle.dtic.mil/100.2/ADA383767.

Full text
Abstract:
Thesis (M.S. in Software Engineering) Naval Postgraduate School, Sept. 2000.
Thesis advisor(s): LuQi; Shing, Mantak. "September 2000." Includes bibliographical references (p. 63-65). Also available in print.
APA, Harvard, Vancouver, ISO, and other styles
24

Juric, Radmila. "Component based software architecture for interoperable databases and extensions to the results." Thesis, Brunel University, 2005. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.428342.

Full text
APA, Harvard, Vancouver, ISO, and other styles
25

Tran, Tam M., and James O. Allen. "Interoperability and security support for heterogeneous Cots/Gots/legacy component-based architecture." Thesis, Monterey, California. Naval Postgraduate School, 2000. http://hdl.handle.net/10945/9419.

Full text
Abstract:
There is a need for Commercial-off-the-shelf (COTS), Government-off- the-shelf (GOTS) and legacy components to interoperate in a secure distributed computing environment in order to facilitate the development of evolving applications. This thesis researches existing open standards solutions to the distributed component integration problem and proposes an application framework that supports application wrappers and a uniform security policy external to the components. This application framework adopts an Object Request Broker (ORB) standard based on Microsoft Distributed Component Object Model (DCOM). Application wrapper architectures are used to make components conform to the ORB standard. The application framework is shown to operate in a common network architecture. A portion of the Naval Integrated Tactical Environmental System I (NITES I) is used as a case study to demonstrate the utility of this distributed component integration methodology (DCIM).
APA, Harvard, Vancouver, ISO, and other styles
26

Svensson, Tim. "Cloud based platform for real time Gait analysis : System component: Mobile architecture." Thesis, Högskolan i Halmstad, Akademin för informationsteknologi, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:hh:diva-34272.

Full text
Abstract:
Today gait analysis are performed in laboratories with expensive equipment and people must visit the labs to perform several supervised tests. The goal of this project is to develop a platform enables gait analysis with the accelerometer sensor in a mobile phone. This would allow more people to do gait analysis, as smart phones are widely available and cheap equipment compared to lab equipments. In order to solve this task a mobile application and a cloud server was created. The mobile application can gather data from the internal accelerometer sensor and a medical grade sensor simultaneously and send the data to the cloud server. When two sensors are used the symmetry between the left foot and the right foot can be measured, although the system works with only one sensor aswell. On the cloud server the accelerometer data is analysed and gait analysis is done on the data and visualized on a web page. The mobile application can collect data for 4 hours at a sampling rate of 120Hz and two sensors are used. When sending data collected from two sensors to the cloud at a sampling rate of 120 Hz the amount of data is approximately 21.96 Mb/h. The goal of the project was to create a proof-of-concept platform to do gait analysis and that goal was fullled and a fully functional platform was developed.
APA, Harvard, Vancouver, ISO, and other styles
27

Zhang, Huaxi Yulin. "Multi-dimensional architecture description language for forward and reverse evolution of component-based software." Thesis, Montpellier 2, 2010. http://www.theses.fr/2010MON20023/document.

Full text
Abstract:
Les approches basées sur les composants permettent de développer des logiciels en réutilisant des composants existant dans des bibliothèques. La structure d'un tel logiciel produit par assemblage de composants est définie à l'aide d'un langage de description d'architectures (ADL). Les processus de développement sont encore peu adaptés à ce paradigme. Ainsi, les ADL existants ne fournissent pas de véritable support au développement et à l'évolution des architectures logicielles à composants. Cette thèse propose Dedal, un ADL permettant de définir une architecture logicielle à différents niveaux d'abstraction qui correspondent aux étapes du cycle de vie du logiciel : sa spécification, son implémentation et son déploiement. La définition de l'architecture est complétée par un modèle du logiciel à l'exécution. La cohérence entre les différentes définitions d'une architecture doit être assurée : sa définition à un niveau d'abstraction doit être conforme à sa définition à un niveau supérieur. Ce principe permet de contrôler l'évolution d'une architecture, en validant les modifications réalisées à un certain niveau d'abstraction ou en motivant la création d'une nouvelle version pour propager les modifications entre niveaux d'abstraction et rétablir la cohérence. Ces mécanismes préviennent les problèmes d'érosion ou de dérive qui surviennent lors des évolutions entre les différents niveaux de définition des architectures. Un environnement couvrant le cycle de vie complet d'un logiciel à base de composants a été prototypé. Il comporte un atelier, permettant de décrire des architectures avec Dedal, puis un environnement d'exécution, extension des outils Fractal, capable de contrôler l'évolution des architectures déployées. L'évolution à l'exécution est réalisée de façon graduelle, de manière à faire fonctionner et à instrumenter les nouvelles versions pendant une phase de transition, avant de valider définitivement une modification
Component-based approaches promote software development by reuse of existing components from a repository. The structure of such software is described as an assembly of software components using an architecture description language (ADL). Software development processes often do not comply with this paradigm yet. Consequently, existing ADLs do not fully support component-based software architecture development and evolution. This thesis proposes Dedal, an ADL to describe software architectures at several abstraction levels that correspond to the steps of software lifecycle: its specification, its implementation and its deployment. The architecture definition is completed with a runtime model of the software. Consistency between the various definitions of a given architecture must be maintained: its definition at some abstraction level must conform to its definition at a higher abstraction level. This consistency principle enables to control the evolution of architectures either validating changes performed at an abstraction level or motivating the creation of a new version, to propagate changes from an abstraction level to the other and restore their consistency. These mechanisms prevent from architecture erosion or drift which might occur between two different description levels after evolution. An environment that covers the whole lifecycle of component-based software has been prototyped. It includes a CASE tool that supports the Dedal-based description of architectures and a runtime environment that extends Fractal tools to control the evolution of the deployed software. Runtime evolution is performed gradually in order to have new versions run and instrumented during a transition phase before committing changes
APA, Harvard, Vancouver, ISO, and other styles
28

Niemelä, Eila. "A component framework of a distributed control systems family /." Espoo [Finland] : Technical Research Centre of Finland, 1999. http://www.vtt.fi/inf/pdf/publications/1999/P402.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
29

Fiuczynski, Marc Eric. "Safe and efficient resource sharing in component-based systems /." Thesis, Connect to this title online; UW restricted, 2004. http://hdl.handle.net/1773/6992.

Full text
APA, Harvard, Vancouver, ISO, and other styles
30

McKean, David Keith. "Leveraging Model-Based Techniques for Component Level Architecture Analysis in Product-Based Systems." Thesis, The George Washington University, 2019. http://pqdtopen.proquest.com/#viewpdf?dispub=13812870.

Full text
Abstract:

System design at the component level seeks to construct a design trade space of alternate solutions comprising mapping(s) of system function(s) to physical hardware or software product components. The design space is analyzed to determine a near-optimal next-level allocated architecture solution that system function and quality requirements. Software product components are targeted to increasingly complex computer systems that provide heterogeneous combinations of processing resources. These processing technologies facilitate performance (speed) optimization via algorithm parallelization. However, speed optimization can conflict with electrical energy and thermal constraints. A multi-disciplinary architecture analysis method is presented that considers all attribute constraints required to synthesize a robust, optimum, extensible next-level solution. This paper presents an extensible, executable model-based architecture attribute framework that efficiently constructs a component-level design trade space. A proof-of-concept performance attribute model is introduced that targets single-CPU systems. The model produces static performance estimates that support optimization analysis and dynamic performance estimation values that support simulation analysis. This model-based approach replaces current architecture analysis of alternatives spreadsheet approaches. The capability to easily model computer resource alternatives that produces attribute estimates improves design space exploration productivity. Performance estimation improvements save time and money through reduced prototype requirements. Credible architecture attribute estimates facilitate more informed design tradeoff discussions with specialty engineers. This paper presents initial validation of a model-based architecture attribute analysis method and model framework using a single computation thread application on two laptop computers with different CPU configurations. Execution time estimates are calibrated for several data input sizes using the first laptop. Actual execution times on the second laptop are shown to be within 10 percent of execution time estimates for all data input sizes.

APA, Harvard, Vancouver, ISO, and other styles
31

Carlsson, Hampus, and Kärrman Marcus. "Cloud-based Mobile System for Free-Living Gait Analysis : System component : Server architecture." Thesis, Högskolan i Halmstad, Akademin för informationsteknologi, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:hh:diva-34293.

Full text
Abstract:
Progress in the fields of wearable sensor technologies together with specialized analysis algorithms has enabled systems for gait analysis outside labs. An example of a wearable sensor is the accelerometer embedded in a typical smartphone. The goal was to propose a system design capable of hosting existing gait analysis algorithms in a cloud environment, and tailor the design as to deliver fast results with the ambition of reaching near real-time.    The project identified a set of enabling technologies by examining existing systems for gait analysis; the technologies included cloud computing and WebSockets. The final system design is a hierarchical composition starting with a Linux VM running Node.js, which in turn connects to a database and hosts instances of the MatLab runtime. The results show the feasibility of mobile cloud based free-living gait analysis. The architectural design provides a solution to the critical problem of enabling existing algorithms to run in a cloud environment; and shows how  the graphical output of the native algorithm could be accurately reproduced in a web browser. The system can process a chunk of 1300 data points under 3 seconds for a client streaming at 128 Hz, while simultaneously streaming the real time signal.
APA, Harvard, Vancouver, ISO, and other styles
32

Stoicescu, Miruna. "Architecting Resilient Computing Systems : a Component-Based Approach." Thesis, Toulouse, INPT, 2013. http://www.theses.fr/2013INPT0120/document.

Full text
Abstract:
L'évolution des systèmes pendant leur vie opérationnelle est incontournable. Les systèmes sûrs de fonctionnement doivent évoluer pour s'adapter à des changements comme la confrontation à de nouveaux types de fautes ou la perte de ressources. L'ajout de cette dimension évolutive à la fiabilité conduit à la notion de résilience informatique. Parmi les différents aspects de la résilience, nous nous concentrons sur l'adaptativité. La sûreté de fonctionnement informatique est basée sur plusieurs moyens, dont la tolérance aux fautes à l'exécution, où l'on attache des mécanismes spécifiques (Fault Tolerance Mechanisms, FTMs) à l'application. A ce titre, l'adaptation des FTMs à l'exécution s'avère un défi pour développer des systèmes résilients. Dans la plupart des travaux de recherche existants, l'adaptation des FTMs à l'exécution est réalisée de manière préprogrammée ou se limite à faire varier quelques paramètres. Tous les FTMs envisageables doivent être connus dès le design du système et déployés et attachés à l'application dès le début. Pourtant, les changements ont des origines variées et, donc, vouloir équiper un système pour le pire scénario est impossible. Selon les observations pendant la vie opérationnelle, de nouveaux FTMs peuvent être développés hors-ligne, mais intégrés pendant l'exécution. On dénote cette capacité comme adaptation agile, par opposition à l'adaptation préprogrammée. Dans cette thèse, nous présentons une approche pour développer des systèmes sûrs de fonctionnement flexibles dont les FTMs peuvent s'adapter à l'exécution de manière agile par des modifications à grain fin pour minimiser l'impact sur l'architecture initiale. D'abord, nous proposons une classification d'un ensemble de FTMs existants basée sur des critères comme le modèle de faute, les caractéristiques de l'application et les ressources nécessaires. Ensuite, nous analysons ces FTMs et extrayons un schéma d'exécution générique identifiant leurs parties communes et leurs points de variabilité. Après, nous démontrons les bénéfices apportés par les outils et les concepts issus du domaine du génie logiciel, comme les intergiciels réflexifs à base de composants, pour développer une librairie de FTMs adaptatifs à grain fin. Nous évaluons l'agilité de l'approche et illustrons son utilité à travers deux exemples d'intégration : premièrement, dans un processus de développement dirigé par le design pour les systèmes ubiquitaires et, deuxièmement, dans un environnement pour le développement d'applications pour des réseaux de capteurs
Evolution during service life is mandatory, particularly for long-lived systems. Dependable systems, which continuously deliver trustworthy services, must evolve to accommodate changes e.g., new fault tolerance requirements or variations in available resources. The addition of this evolutionary dimension to dependability leads to the notion of resilient computing. Among the various aspects of resilience, we focus on adaptivity. Dependability relies on fault tolerant computing at runtime, applications being augmented with fault tolerance mechanisms (FTMs). As such, on-line adaptation of FTMs is a key challenge towards resilience. In related work, on-line adaption of FTMs is most often performed in a preprogrammed manner or consists in tuning some parameters. Besides, FTMs are replaced monolithically. All the envisaged FTMs must be known at design time and deployed from the beginning. However, dynamics occurs along multiple dimensions and developing a system for the worst-case scenario is impossible. According to runtime observations, new FTMs can be developed off-line but integrated on-line. We denote this ability as agile adaption, as opposed to the preprogrammed one. In this thesis, we present an approach for developing flexible fault-tolerant systems in which FTMs can be adapted at runtime in an agile manner through fine-grained modifications for minimizing impact on the initial architecture. We first propose a classification of a set of existing FTMs based on criteria such as fault model, application characteristics and necessary resources. Next, we analyze these FTMs and extract a generic execution scheme which pinpoints the common parts and the variable features between them. Then, we demonstrate the use of state-of-the-art tools and concepts from the field of software engineering, such as component-based software engineering and reflective component-based middleware, for developing a library of fine-grained adaptive FTMs. We evaluate the agility of the approach and illustrate its usability throughout two examples of integration of the library: first, in a design-driven development process for applications in pervasive computing and, second, in a toolkit for developing applications for WSNs
APA, Harvard, Vancouver, ISO, and other styles
33

Pintar, Freddie. "Investigation and Implementation of a Live Connection between Configura CET and Revit Architecture 2009." Thesis, Linköping University, PELAB - Programming Environment Laboratory, 2009. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-17357.

Full text
Abstract:

Building Information Modeling -BIM- is an innovative method to seamlessly bridgecommunication within the architecture, engineering and construction industries.With BIM software you can exchange information during the design, construction,and maintaining. BIM can be seen as a continuation of the CAD software, wherethe users exchanged information by word of mouth, now is made automatically.To get the effect required for BIM one or more CAD-systems have to work togetherto exchange information. Revit Architecture is an application by Autodeskwhere BIM is used from the design and construction to the documentation andmaintaining of a building. Configura is one of the major software developers of interiorsolutions and want to integrate their software with Revit Architecture. Theconcept of objects in both software system suit well to be used in BIM and witha live connection these could be shared between the applications. One of the conclusionsin this investigation was that the only way to have integration betweenthe applications was to use the API provided by Autodesk. And therefore theimplementation is limited to the function in it. Revit API is a powerful programmingenvironment that let 3rd party software extend the functionality in Revit.The results show how Remote Procedure Call as a communication tool can beused to exchange data between the applications, how different type of data can berepresented in both applications, and why we cannot have a live synchronization.

APA, Harvard, Vancouver, ISO, and other styles
34

Matas, Petr. "Connected component tree construction for embedded systems." Thesis, Paris Est, 2014. http://www.theses.fr/2014PEST1116/document.

Full text
Abstract:
L'objectif du travail présenté dans cette thèse est de proposer un avancement dans la construction des systèmes embarqués de traitement d'images numériques, flexibles et puissants. La proposition est d'explorer l'utilisation d'une représentation d'image particulière appelée « arbre des composantes connexes » (connected component tree – CCT) en tant que base pour la mise en œuvre de l'ensemble de la chaîne de traitement d'image. Cela est possible parce que la représentation par CCT est à la fois formelle et générale. De plus, les opérateurs déjà existants et basés sur CCT recouvrent tous les domaines de traitement d'image : du filtrage de base, passant par la segmentation jusqu'à la reconnaissance des objets. Une chaîne de traitement basée sur la représentation d'image par CCT est typiquement composée d'une cascade de transformations de CCT où chaque transformation représente un opérateur individuel. A la fin, une restitution d'image pour visualiser les résultats est nécessaire. Dans cette chaîne typique, c'est la construction du CCT qui représente la tâche nécessitant le plus de temps de calcul et de ressources matérielles. C'est pour cette raison que ce travail se concentre sur la problématique de la construction rapide de CCT. Dans ce manuscrit, nous introduisons le CCT et ses représentations possibles dans la mémoire de l'ordinateur. Nous présentons une partie de ses applications et analysons les algorithmes existants de sa construction. Par la suite, nous proposons un nouvel algorithme de construction parallèle de CCT qui produit le « parent point tree » représentation de CCT. L'algorithme est conçu pour les systèmes embarqués, ainsi notre effort vise la minimisation de la mémoire occupée. L'algorithme en lui-même se compose d'un grand nombre de tâches de la « construction » et de la « fusion ». Une tâche de construction construit le CCT d'une seule ligne d'image, donc d'un signal à une dimension. Les tâches de fusion construisent progressivement le CCT de l'ensemble. Pour optimiser la gestion des ressources de calcul, trois différentes stratégies d'ordonnancement des tâches sont développées et évaluées. Également, les performances des implantations de l'algorithme sont évaluées sur plusieurs ordinateurs parallèles. Un débit de 83 Mpx/s pour une accélération de 13,3 est réalisé sur une machine 16-core avec Opteron 885 processeurs. Les résultats obtenus nous ont encouragés pour procéder à une mise en œuvre d'une nouvelle implantation matérielle parallèle de l'algorithme. L'architecture proposée contient 16 blocs de base, chacun dédié à la transformation d'une partie de l'image et comprenant des unités de calcul et la mémoire. Un système spécial d'interconnexions est conçu pour permettre à certaines unités de calcul d'accéder à la mémoire partagée dans d'autres blocs de base. Ceci est nécessaire pour la fusion des CCT partiels. L'architecture a été implantée en VHDL et sa simulation fonctionnelle permet d'estimer une performance de 145 Mpx/s à fréquence d'horloge de 120 MHz
The aim of this work is to enable construction of embedded digital image processing systems, which are both flexible and powerful. The thesis proposal explores the possibility of using an image representation called connected component tree (CCT) as the basis for implementation of the entire image processing chain. This is possible, because the CCT is both simple and general, as CCT-based implementations of operators spanning from filtering to segmentation and recognition exist. A typical CCT-based image processing chain consists of CCT construction from an input image, a cascade of CCT transformations, which implement the individual operators, and image restitution, which generates the output image from the modified CCT. The most time-demanding step is the CCT construction and this work focuses on it. It introduces the CCT and its possible representations in computer memory, shows some of its applications and analyzes existing CCT construction algorithms. A new parallel CCT construction algorithm producing the parent point tree representation of the CCT is proposed. The algorithm is suitable for an embedded system implementation due to its low memory requirements. The algorithm consists of many building and merging tasks. A building task constructs the CCT of a single image line, which is treated as a one-dimensional signal. Merging tasks fuse the CCTs together. Three different task scheduling strategies are developed and evaluated. Performance of the algorithm is evaluated on multiple parallel computers. A throughput 83 Mpx/s at speedup 13.3 is achieved on a 16-core machine with Opteron 885 CPUs. Next, the new algorithm is further adapted for hardware implementation and implemented as a new parallel hardware architecture. The architecture contains 16 basic blocks, each dedicated to processing of an image partition and consisting of execution units and memory. A special interconnection switch is designed to allow some executions units to access memory in other basic blocks. The algorithm requires this for the final merging of the CCTs constructed by different basic blocks together. The architecture is implemented in VHDL and its functional simulation shows performance 145 Mpx/s at clock frequency 120 MHz
APA, Harvard, Vancouver, ISO, and other styles
35

Wortmann, Andreas [Verfasser]. "An Extensible Component & Connector Architecture Description Infrastructure for Multi-Platform Modeling / Andreas Wortmann." Aachen : Shaker, 2016. http://d-nb.info/1120864275/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
36

Graf, Sebastian [Verfasser]. "Design and Optimization of Multi-Variant Automotive E/E Architecture Component Platforms / Sebastian Graf." München : Verlag Dr. Hut, 2015. http://d-nb.info/1079768335/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
37

Myers, David S. "An Extensible Component-Based Architecture for Web-Based Simulation Using Standards-Based Web Browsers." Thesis, Virginia Tech, 2004. http://hdl.handle.net/10919/10123.

Full text
Abstract:
Web-based simulation (WBS) systems offer tradeoffs between user interactivity and hardware requirements striking to seek a balance between the differing concerns. Server-based systems offer little interactivity or concurrent visualization capabilities, while client-based systems have increased hardware requirements asking the user to provide high-end workstations. Concurrent visualization of simulation output proves execution intensive, or unusable in some situations. Creating an execution efficient and user friendly WBS system greatly improves user experience while gaining all of the benefits inherent in a web-based system such as high accessibility and ease of maintenance. In order to provide a usable concurrent visualization WBS this thesis developed the Web-Based Queuing System Simulation System (WebQS3). WebQS3 splits the responsibilities of simulation execution and simulation visualization into a client-server environment; the client is responsible for the visualization display and server is responsible for simulation execution. The system differs from many previous WBS systems in that the client-side application is developed using web-standard technologies such as HTTP, XML, SVG, and ECMAScript instead relying on Java Applets and associated technologies. Using web-standards as the foundation of the client agent opens the visualization and model construction functionality to any user that accesses the application using a web browser while also making the application more scalable in terms of user load. Implementing the client with web-standards also included the development of an asynchronous client-server communication protocol as opposed to traditional synchronous communication protocols used by Java WBS systems. The asynchronous protocol demonstrates similar or better execution performance than similar synchronous communication protocols in most quality characteristics. By creating a WBS system using web-standards implemented in most modern web browsers any user may visit the WebQS3 site and have simulation tools available for use. Providing simulation services on the web makes eases the creation of simulation models my making the tools to readily available while facilitating information sharing and collaboration over the web. The WebQS3 system serves as a model to drive research in WBS systems away from proprietary Java technologies to web standards for front-end visualization technologies.
Master of Science
APA, Harvard, Vancouver, ISO, and other styles
38

Tyan, Hung-ying. "Design, realization and evaluation of a component-based compositional software architecture for network simulation /." The Ohio State University, 2002. http://rave.ohiolink.edu/etdc/view?acc_num=osu1486402544591791.

Full text
APA, Harvard, Vancouver, ISO, and other styles
39

Karlsson, Claes. "Designing an Experiment to Compare Component Systems." Thesis, Växjö University, School of Mathematics and Systems Engineering, 2006. http://urn.kb.se/resolve?urn=urn:nbn:se:vxu:diva-819.

Full text
Abstract:

The aim of this work is to design an experiment, where client-server systems can be compared. They belong to the group of systems that are called component systems. Client-server systems are difficult to compare, because they are complex. The client-server systems are documented in different ways. Notations in the implementation of them are in different ways. There is a large difference in the communication between the client and server. The architectures between the client-server systems differ also, but they are not totally different. Therefore it is possible to construct an experiment for comparing them. Client-server systems that will be compared are Java RMI, Web Services, CORBA, and Enterprise JavaBeans. We are going to use Java as the programming language. Some of these systems, for example CORBA, can be implemented in other languages. The designed experiment is among other things going to answer how long time is needed to implement a specific application, how fast a specific client-server system is, and how long time is spent for learning about a specific system.

APA, Harvard, Vancouver, ISO, and other styles
40

Almeida, Junior José Valdvogel de. "Uma proposta de método para especificação de componentes de software." Pontifícia Universidade Católica de São Paulo, 2015. https://tede2.pucsp.br/handle/handle/18198.

Full text
Abstract:
Made available in DSpace on 2016-04-29T14:23:40Z (GMT). No. of bitstreams: 1 Jose Valdvogel de Almeida Junior.pdf: 1664725 bytes, checksum: b5162f7ac0d53d0a3c9e995e1e643e18 (MD5) Previous issue date: 2015-11-19
The component-based software development has proven effective for implementation in various fields of application, but still insufficient to create reusable and flexible parts. Processes and methods have been developed in order to address special care in this context. However, none of the proposed methods shows a simple specification of building blocks. To address this limitation we propose in this research work a split between the static and dynamic elements that make up the nature of an assembled component in a set of steps with application techniques using the best practices of software engineering. As a result, we reach the Component Specification method capable of generating well-designed models for building reusable and flexible software parts. A case study helps to illustrate the application of the method in an application context
O desenvolvimento de software baseado em componentes tem se provado efetivo para a implementação em diversos domínios de aplicação, mas ainda insuficiente para a criação de peças reusáveis e flexíveis. Processos e métodos foram desenvolvidos no intuito de endereçar cuidados especiais nesse contexto. No entanto, nenhum dos métodos propostos demonstra uma forma simples de especificação dos blocos construtores. Para tratar essa limitação propomos nesse trabalho de pesquisa uma divisão entre os elementos estáticos e dinâmicos que compõem a natureza de um componente reunidos em um conjunto de passos com aplicação de técnicas que utilizam as boas práticas da engenharia de software. Como resultado, chega-se ao Método de Especificação de Componentes capaz de gerar modelos bem elaborados para construção de peças de software reusáveis e flexíveis. Um estudo de caso ajuda a ilustrar a aplicação do método em um contexto de aplicação
APA, Harvard, Vancouver, ISO, and other styles
41

Rahman, Asim. "Metrics for the Structural Assessment of Product Line Architecture." Thesis, Blekinge Tekniska Högskola, Avdelningen för programvarusystem, 2004. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-3427.

Full text
Abstract:
The notion of maximizing software reuse among the family of products has gained considerable attention in the last decade. Lots of research has been done on designing and managing the commonalities and variabilities between the products. However, very few metrics have been developed to assist architects in designing product line architectures. The structure of the product line holds immense importance towards increasing the life span of the product line. Since many of the product line architecture design methodologies follow a component based approach, it seems logical to attempt to adapt the component based metrics to the product line domain. In this thesis, we attempt to derive metrics that quantify the structural quality of product line architecture.
+92-42-5727639
APA, Harvard, Vancouver, ISO, and other styles
42

Bari, Himanshu. "Design and implementation of a library to support the Common Component Architecture (CCA) over Legion." Diss., Online access via UMI:, 2004. http://wwwlib.umi.com/dissertations/fullcit/1424173.

Full text
APA, Harvard, Vancouver, ISO, and other styles
43

Murdoch, S. T. "An investigation into the use of a findings base as a component of an expert system." Thesis, Brunel University, 1986. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.374200.

Full text
APA, Harvard, Vancouver, ISO, and other styles
44

Bora, Prachi. "Runtime Algorithm Selection For Grid Environments: A Component Based Framework." Thesis, Virginia Tech, 2003. http://hdl.handle.net/10919/33823.

Full text
Abstract:
Grid environments are inherently heterogeneous. If the computational power provided by collaborations on the Grid is to be harnessed in the true sense, there is a need for applications that can automatically adapt to changes in the execution environment. The application writer should not be burdened with the job of choosing the right algorithm and implementation every time the resources on which the application runs are changed. A lot of research has been done in adapting applications to changing conditions. The existing systems do not address the issue of providing a unified interface to permit algorithm selection at runtime. The goal of this research is to design and develop a unified interface to applications in order to permit seamless access to different algorithms providing similar functionalities. Long running, computationally intensive scientific applications can produce huge amounts of performance data. Often, this data is discarded once the applicationâ s execution is complete. This data can be utilized in extracting information about algorithms and their performance. This information can be used to choose algorithms intelligently. The research described in this thesis aims at designing and developing a component based unified interface for runtime algorithm selection in grid environments. This unified interface is necessary so that the application code does not change if a new algorithm is used to solve the problem. The overhead associated with making the algorithm choice transparent to the application is evaluated. We use a data mining approach to algorithm selection and evaluate its potential effectiveness for scientific applications.
Master of Science
APA, Harvard, Vancouver, ISO, and other styles
45

Bora, Prachi Champalal. "Runtime Algorithm Selection For Grid Environments: A Component Based Framework." Thesis, Virginia Tech, 2003. http://hdl.handle.net/10919/33823.

Full text
Abstract:
Grid environments are inherently heterogeneous. If the computational power provided by collaborations on the Grid is to be harnessed in the true sense, there is a need for applications that can automatically adapt to changes in the execution environment. The application writer should not be burdened with the job of choosing the right algorithm and implementation every time the resources on which the application runs are changed. A lot of research has been done in adapting applications to changing conditions. The existing systems do not address the issue of providing a unified interface to permit algorithm selection at runtime. The goal of this research is to design and develop a unified interface to applications in order to permit seamless access to different algorithms providing similar functionalities. Long running, computationally intensive scientific applications can produce huge amounts of performance data. Often, this data is discarded once the applicationâ s execution is complete. This data can be utilized in extracting information about algorithms and their performance. This information can be used to choose algorithms intelligently. The research described in this thesis aims at designing and developing a component based unified interface for runtime algorithm selection in grid environments. This unified interface is necessary so that the application code does not change if a new algorithm is used to solve the problem. The overhead associated with making the algorithm choice transparent to the application is evaluated. We use a data mining approach to algorithm selection and evaluate its potential effectiveness for scientific applications.
Master of Science
APA, Harvard, Vancouver, ISO, and other styles
46

Mokni, Abderrahman. "A formal approach to automate the evolution management in component-based software development processes." Thesis, Montpellier, 2015. http://www.theses.fr/2015MONTS131/document.

Full text
Abstract:
Gérer l'évolution des logiciels est une tâche complexe mais nécessaire. Tout au long de son cycle de vie, un logiciel doit subir des changements, pour corriger des erreurs, améliorer ses performances et sa qualité, étendre ses fonctionnalités ou s’adapter à son environnement. A défaut d’évoluer, un logiciel se dégrade, devient obsolète ou inadapté et est remplacé. Cependant, sans évaluation de leurs impacts et contrôle de leur réalisation, les changements peuvent être sources d’incohérences et de dysfonctionnements, donc générateurs de dégradations du logiciel. Cette thèse propose une approche améliorant la gestion de l'évolution des logiciels dans les processus de développement orientés composants. Adoptant une démarche d’ingénierie dirigée par les modèles (IDM), cette approche s’appuie sur Dedal, un langage de description d’architecture (ADL) séparant explicitement trois niveaux d’abstraction dans la définition des architectures logicielles. Ces trois niveaux (spécification, configuration et assemblage) correspondent aux trois étapes principales du développement d’une architecture (conception, implémentation, déploiement) et gardent la trace des décisions architecturales prises au fil du développement. Ces informations sont un support efficace à la gestion de l’évolution : elles permettent de déterminer le niveau d’un changement, d’analyser son impact et de planifier sa réalisation afin d’éviter la survenue d’incohérences dans la définition de l’architecture (érosion, dérive, etc.). Une gestion rigoureuse de l’évolution nécessite la formalisation, d’une part, des relations intra-niveau liant les composants au sein des modèles correspondant aux différents niveaux de définition d’une architecture et, d’autre part, des relations inter-niveaux liant les modèles décrivant une même architecture aux différents niveaux d’abstraction. Ces relations permettent la définition des propriétés de consistance et de cohérence servant à vérifier la correction d’une architecture. Le processus d’évolution est ainsi décomposé en trois phases : initier le changement de la définition de l’architecture à un niveau d’abstraction donné ; vérifier et rétablir la consistance de cette définition en induisant des changements complémentaires ; vérifier et rétablir la cohérence globale de la définition de l’architecture en propageant éventuellement les changements aux autres niveaux d’abstraction.Ces relations et propriétés sont décrites en B, un langage de modélisation formel basé sur la théorie des ensembles et la logique du premier ordre. Elles s’appliquent à des architectures définies avec un adl formel écrit en B dont le méta-modèle, aligné avec celui de Dedal, permet d’outiller la transformation de modèles entre les deux langages. Cette intégration permet de proposer un environnement de développement conjuguant les avantages des approches IDM et formelle : la conception d’architectures avec l’outillage de Dedal (modeleur graphique); la vérification des architectures et la gestion de l’évolution avec l’outillage de B (animateur, model-checker, solver). Nous proposons en particulier d’utiliser un solver B pour calculer automatiquement des plans d’évolution conformes à notre proposition et avons ainsi défini l’ensemble des règles d’évolution décrivant les opérations de modification applicables à la définition d’une architecture. Le solver recherche alors automatiquement une séquence de modifications permettant la réalisation d’un changement cible tout en préservant les propriétés de consistance et de cohérence de l’architecture. Nous avons validé la faisabilité de cette gestion de l’évolution par une implémentation mêlant optimisation et génie logiciel (search-based software engineering), intégrant notre propre solver pourvu d’heuristiques spécifiques qui améliorent significativement les temps de calcul, pour expérimenter trois scénarios d’évolution permettant de tester la réalisation d’un changement à chacun des trois niveaux d’abstraction
Managing software evolution is a complex task. Indeed, throughout their whole lifecycle, software systems are subject to changes to extend their functionalities, correct bugs, improve performance and quality, or adapt to their environment. If not evolved, software systems degrade, become obsolete or inadequate and are replaced. While unavoidable, software changes may engender several inconsistencies and system dysfunction if not analyzed and handled carefully hence leading to software degradation and phase-out.This thesis proposes an approach to improve the evolution management activity in component-based software development processes. The solution adopts a Model-Driven Engineering (MDE) approach. It is based on Dedal, an Architecture Description Language (ADL) that explicitly separates software architecture descriptions into three abstraction levels: specification, configuration and assembly. These abstraction levels respectively correspond to the three major steps of component-based development (design, implementation and deployment) and trace architectural decisions all along development. Dedal hence efficiently supports evolution management: It enables to determine the level of change, analyze its impact and plan its execution in order to prevent architecture inconsistencies (erosion, drift, etc.). Rigorous evolution management requires the formalization, on the one hand, of intra-level relations linking components within models corresponding to different architecture abstraction levels and on the other hand, of the formalization of inter-level relations linking models describing the same architecture at different abstraction levels. These relations enable the definition of the consistency and coherence properties that prove necessary for architecture correctness analysis. The evolution process therefore consists of three steps: First, change is initiated on an architecture description at a given abstraction level; then, the consistency of the impacted description is checked out and restored by triggering additional changes; finally, the global coherence of the architecture definitions is verified and restored by propagating changes to other abstraction levels.Relations and properties are expressed in B, a set-theoretic and first-order logic language. They are applied on B formal ADL, the meta-model of which is mapped to Dedal's and helps automatic model transformations. This integration enables to implement a development environment that combines the benefits of both MDE and formal approaches: Software architecture design using Dedal tools (graphical modeler) and architecture analysis and evolution management using B tools (animator, model-checker, solver).In particular, we propose to use a B solver to automatically calculate evolution plans according to our approach. The solver explores a set of defined evolution rules that describe the change operations that can apply on architecture definitions. It automatically searches for a sequence of operations that both changes the architecture as requested and preserves architecture consistency and coherence properties. The feasibility of the evolution management approach is demonstrated through the experimentation of three evolution scenarios, each addressing a change at different abstraction level. The experimentation relies on an implementation of a search-based software engineering approach mixing software engineering and optimization and integrates our own solver with specific heuristics that significantly improve calculation time
APA, Harvard, Vancouver, ISO, and other styles
47

Subramanian, Gayatri. "Automating Component-Based System Assembly." Thesis, Georgia Institute of Technology, 2006. http://hdl.handle.net/1853/11508.

Full text
Abstract:
Owing to advancements in component re-use technology, component-based software development (CBSD) has come a long way in developing complex commercial software systems while reducing software development time and cost. However, assembling distributed resource-constrained and safety-critical systems using current assembly techniques is a challenge. Within complex systems when there are numerous ways to assemble the components unless the software architecture clearly defines how the components should be composed, determining the correct assembly that satisfies the system assembly constraints is difficult. Component technologies like CORBA and .NET do a very good job of integrating components, but they do not automate component assembly; it is the system developer's responsibility to ensure thatthe components are assembled correctly. In this thesis, we first define a component-based system assembly (CBSA) technique called "Constrained Component Assembly Technique" (CCAT), which is useful when the system has complex assembly constraints and the system architecture specifies component composition as assembly constraints. The technique poses the question: Does there exist a way of assembling the components that satisfies all the connection, performance, reliability, and safety constraints of the system, while optimizing the objective constraint? To implement CCAT, we present a powerful framework called "CoBaSA". The CoBaSA framework includes an expressive language for declaratively describing component functional and extra-functional properties, component interfaces, system-level and component-level connection, performance, reliability, safety, and optimization constraints. To perform CBSA, we first write a program (in the CoBaSA language) describing the CBSA specifications and constraints, and then an interpreter translates the CBSA program into a satisfiability and optimization problem. Solving the generated satisfiability and optimization problem is equivalent to answering the question posed by CCAT. If a satisfiable solution is found, we deduce that the system can be assembled without violating any constraints. Since CCAT and CoBaSA provide a mechanism for assembling systems that have complex assembly constraints, they can be utilized in several industries like the avionics industry. We demonstrate the merits of CoBaSA by assembling an actual avionic system that could be used on-board a Boeing aircraft. The empirical evaluation shows that our approach is promising and can scale to handle complex industrial problems.
APA, Harvard, Vancouver, ISO, and other styles
48

JUNIOR, CELSO GOMES BARRETO. "ADDING SYSTEM INTRASTRUCTURE FRAMEWORKS IN AN COMPONENT BASED ARCHITECTURE: A CASE STUDY WITHIN THE AULANET ENVIRONMENT." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2006. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=8623@1.

Full text
Abstract:
COORDENAÇÃO DE APERFEIÇOAMENTO DO PESSOAL DE ENSINO SUPERIOR
FUNDAÇÃO PADRE LEONEL FRANCA
Groupware é difícil de construir e de manter, pois envolve aspectos multidisciplinares. Além das dificuldades associadas ao desenvolvimento de aplicações colaborativas, usualmente o desenvolvedor de groupware deve se preocupar com outros aspectos de infra-estrutura. Nesta dissertação é proposta uma arquitetura multicamadas baseada em componentes para groupware, utilizando frameworks de infra-estrutura. Na camada de negócio são utilizados os frameworks Hibernate, responsável pela persistência dos dados da aplicação, e o framework Spring, que dentre outras coisas é responsável pelo controle de transações e pela exposição de serviços remotamente. Na camada de apresentação o framework JaveServer Faces provê meios para criar e reusar componentes de interface. Nesta dissertação também é apresentada uma forma de comparar frameworks de infra-estrutura, levando em consideração tanto aspectos técnicos, que definem se o framework atende aos requisitos da aplicação, quanto nãotécnicos, relacionados a aspectos como documentação disponível e aceitação no mercado. A arquitetura definida nesta dissertação é aplicada no AulaNet, groupware voltado para a aprendizagem desenvolvido no Laboratório de Engenharia de Software da PUC-Rio.
Groupware is difficult to develop and maintain because it involves multidisciplinary aspects in its construction. Besides the difficulties related to the development of collaborative applications, usually the developer must handle with other infrastructure aspects. In this dissertation, it is proposed a multilayer component based architecture with system infrastructure frameworks to deal with them. In the business layer, the Hibernate framework is responsible for the persistence of application data, and the Spring framework is responsible for, amongst others, transactions control and remote exposition of services. In the presentation layer the JaveServer Faces framework provides ways to create and to reuse user-interface components. This dissertation also presents a way to compare system infrastructure frameworks, considering both technical aspects, related to the application requirements fulfillment, and non- technical, related to aspects such as documentation availability and market acceptance. The architecture defined in this dissertation is applied to the AulaNet, which is a groupware for learning developed in the Software Engineering Laboratory of PUC-Rio.
APA, Harvard, Vancouver, ISO, and other styles
49

Shatnawi, Anas. "Supporting Reuse by Reverse Engineering Software Architecture and Component from Object-Oriented Product Variants and APIs." Thesis, Montpellier, 2015. http://www.theses.fr/2015MONTS140/document.

Full text
Abstract:
La réutilisation est reconnue comme une démarche intéressante pour améliorer la qualité des produits et la productivité des projets logiciels. L'ingénierie des logiciels à base de composants (CBSE en anglais) et l'ingénierie des lignes de produits logiciels (SPLE en anglais) sont considérées parmi les plus importants paradigmes de réutilisation systématique. L'ingénierie à base de composants permet de développer de nouveaux systèmes logiciels par composition de briques préconstruites appelées composants. L'ingénierie des lignes de produits logiciels permet de dériver (construire) de nouveaux produits par simple sélection de leurs caractéristiques (feature en anglais). Cette dérivation est rendue possible grâce à la représentation et à la gestion de la variabilité et de la similarité des produits d'une même famille. Cependant, une des difficultés vers une large adoption de l'ingénierie des logiciels à base de composants et des lignes de produits est le coût à investir pour construire, à partir de rien, les composants et les artefacts de lignes de produits. Dans ce contexte, les travaux de cette thèse proposent de réduire ce coût par une démarche basée sur la rétro-ingénierie.La première contribution de cette thèse consiste à proposer une approche permettant d'identifier, par l'analyse d'un ensemble de variantes d'un produit logiciel orienté objet, les éléments du code source pouvant servir à l'implémentation de composants. Au contraire des approches existantes d'identification de composants basées sur l'analyse d'un seul produit, l'originalité de notre approche consiste en l'analyse de plusieurs variantes de produits en même temps. Notre objectif est l'amélioration de la réutilisabilité des composants extraits. L'évaluation expérimentale menée dans le cadre de cette thèse a montré la pertinence de cette démarche.La deuxième contribution consiste en la proposition d'une approche pour l'extraction d'une architecture à base de composants d'un ensemble de variantes d'un produit logiciel orienté objet. Il s'agit d'identifier la variabilité des composants architecturaux et la configuration architecturale. L'identification de la configuration architecturale est principalement basée sur l'utilisation de l'analyse formelle de concepts pour trouver les dépendances entre les éléments architecturaux. L'expérimentation conduite pour l'évaluation de l'approche proposée confirme la pertinence des éléments identifiés.La troisième contribution de cette thèse permet de restructurer les APIs orientées objet en composants. Nous exploitons la spécificité des classes des APIs d'être conçues pour être utilisées par des applications clientes pour identifier ces composants. Le code source de ces APIs et celui de leurs applications clientes sont analysés afin d'identifier des groupes de classes qui peuvent servir à l'implémentation de composants à extraire. L'identification de ces groupes de classes est basée sur l'analyse des liens structurels entre ces classes et sur la probabilité que ces classes soient utilisées ensemble par les applications clientes. Nous montrons à travers les résultats expérimentaux que la restructuration des API orientées objet en composants facilite la réutilisation et la compréhension des éléments de ces APIs
It is widely recognized that software quality and productivity can be significantly improved by applying a systematic reuse approach. In this context, Component-Based Software Engineering (CBSE) and Software Product Line Engineering (SPLE) are considered as two important systematic reuse paradigms. CBSE aims at composing software systems based on pre-built software components and SPLE aims at building new products by managing commonalty and variability of a family of similar software. However, building components and SPL artifacts from scratch is a costly task. In this context, our dissertation proposes three contributions to reduce this cost.Firstly, we propose an approach that aims at mining reusable components from a set of similar object-oriented software product variants. The idea is to analyze the commonality and the variability of product variants, in order to identify pieces of code that may form reusable components. Our motivation behind the analysis of several existing product variants is that components mined from these variants are more reusable for the development of new software products than those mined from single ones. The experimental evaluation shows that the reusability of the components mined using our approach is better than those mined from single software.Secondly, we propose an approach that automatically recovers the component-based architecture of a set of object-oriented software product variants. Our contribution is twofold: the identification of the architectural component variability and the identification of the configuration variability. The configuration variability is based on the identification of dependencies between the architectural elements using formal concept analysis. The experimental evaluation shows that our approach is able to identify the architectural variability.Finally, we propose an approach that aims at restructuring object-oriented APIs as component-based ones. This approach exploits specificity of API entities by statically analyzing the source code of both APIs and their software clients to identify groups of API classes that are able to form components. Our assumption is based on the probability of classes to be reused together by API clients on the one hand, and on the structural dependencies between classes on the other hand. The experimental evaluation shows that structuring object-oriented APIs as component-based ones improves the reusability and the understandability of these APIs
APA, Harvard, Vancouver, ISO, and other styles
50

Ho, Quan. "Architecture and Applications of a Geovisual Analytics Framework." Doctoral thesis, Linköpings universitet, Medie- och Informationsteknik, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-91679.

Full text
Abstract:
The large and ever-increasing amounts of multi-dimensional, multivariate, multi-source, spatio-temporal data represent a major challenge for the future. The need to analyse and make decisions based on these data streams, often in time-critical situations, demands integrated, automatic and sophisticated interactive tools that aid the user to manage, process, visualize and interact with large data spaces. The rise of `Web 2.0', which is undisputedly linked with developments such as blogs, wikis and social networking, and the internet usage explosion in the last decade represent another challenge for adapting these tools to the Internet to reach a broader user community. In this context, the research presented in this thesis introduces an effective web-enabled geovisual analytics framework implemented, applied and verified in Adobe Flash ActionScript and HTML5/JavaScript. It has been developed based on the principles behind Visual Analytics and designed to significantly reduce the time and effort needed to develop customized web-enabled applications for geovisual analytics tasks and to bring the benefits of visual analytics to the public. The framework has been developed based on a component architecture and includes a wide range of visualization techniques enhanced with various interaction techniques and interactive features to support better data exploration and analysis. The importance of multiple coordinated and linked views is emphasized and a number of effective techniques for linking views are introduced. Research has so far focused more on tools that explore and present data while tools that support capturing and sharing gained insight have not received the same attention. Therefore, this is one of the focuses of the research presented in this thesis. A snapshot technique is introduced, which supports capturing discoveries made during the exploratory data analysis process and can be used for sharing gained knowledge. The thesis also presents a number of applications developed to verify the usability and the overall performance of the framework for the visualization, exploration and analysis of data in different domains. Four application scenarios are presented introducing (1) the synergies among information visualization methods, geovisualization methods and volume data visualization methods for the exploration and correlation of spatio-temporal ocean data, (2) effective techniques for the visualization, exploration and analysis of self-organizing network data, (3) effective flow visualization techniques applied to the analysis of time-varying spatial interaction data such as migration data, commuting data and trade flow data, and (4) effective techniques for the visualization, exploration and analysis of flood data.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography