Dissertations / Theses on the topic 'Epistemology|Business Administration, Management|Information Science'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Epistemology|Business Administration, Management|Information Science.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Pourshahid, Alireza. "A URN-based methodology for business process monitoring." Thesis, University of Ottawa (Canada), 2008. http://hdl.handle.net/10393/27605.

Full text
Abstract:
Business Process Management Systems (BPMS) are attracting much attention these days as core tools for process management. A BPMS consists of several modules including Business Activity Monitoring (BAM), which evaluates the performance of processes. Although measuring the performance of business processes is based on organizational goals and the impact of processes on such goals is an important aspect of the process evaluation, most of the existing BPMS do not offer appropriate capabilities. Several process improvements and quality methodologies have also been around for some time. Most of them, however, are based on statistical and management tools that are not integrated with current technologies including BPMS. In other words, the management and quality disciplines have not yet evolved to take full advantage of current technologies that can enhance process improvement efforts. One of the pitfalls here is the failure to use the information generated by different information systems dispersed across the organization to evaluate the impact of processes on their goals. In this thesis, we use the User Requirements Notation (URN) in a methodology for evaluating business processes against organizational goals. Although URN enables the modeling of processes and goals, its process monitoring capabilities need enhancements and we will address this issue by extending URN. In addition, we propose a methodology that exploits the new capabilities of URN for process analysis and improvement. Unlike other quality methodologies, the methodology and supporting tool proposed are capable of using different sources of information to measure the performance of modeled processes and evaluate their impact on the goal models. The new capabilities and their benefits are illustrated through examples from the healthcare domain.
APA, Harvard, Vancouver, ISO, and other styles
2

Noaman, Amin Yousef. "Reconciling formal and informal documentation in business modelling." Thesis, McGill University, 1995. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=22783.

Full text
Abstract:
Business modelling, the modelling of architectures and processes of organizations, should have a broad scope. It should not exclusively capture the basic information of the processes, but also address the various kinds of documentation related to the processes under consideration. In this combination, organizational models will be more expressive and useful.
The research reported here describes and demonstrates a new approach for reconciling formal and informal documentation in business modelling. It is based on the integration of an underlying formal modelling approach with hypertext concepts that provide mechanisms for capturing, manipulating and viewing informal model documentation.
We have developed the Hypertec tool which complements the Macrotec environment. Macrotec is a business modelling environment that is based on the formalism of extended colored Petri nets. Hypertec is a hypertext-based component supporting authoring, display and navigation of all the process documentation that cannot be captured by Macrotec. Our experience with Macrotec/Hypertec shows that their combined functionality substantially facilitates the understanding of business processes and clearly reduces problems such as miscommunication, misinterpretation, and misunderstandings about entire processes or some of their components.
APA, Harvard, Vancouver, ISO, and other styles
3

Glynn, Melissa Sue 1969. "Successful behaviors in information systems development teams." Diss., The University of Arizona, 1998. http://hdl.handle.net/10150/282674.

Full text
Abstract:
This dissertation research examines the impact of leadership, cohesion, and information sharing, and the application of group support systems on information design systems (ISD) project quality and project team satisfaction. Research has identified that after 40 years of developing information systems, there are still widespread difficulties in delivering systems on time and on budget. The research objective of this study is to examine the group level processes to understand how ISD team behavior can impact quality issues. A group support system was introduced to act as a sensemaking treatment to increase team performance. The following research questions were identified: (1) What is the impact of cohesion on project quality? (2) What is the impact of leadership on project quality? (3) What is the impact of information sharing on project quality? (4) What is the impact of cohesion on team satisfaction? (5) What is the impact of leadership on team satisfaction? (6) What is the impact of information sharing on team satisfaction? (7) Is there a relationship between group support systems use and project quality? (8) Can group support systems enable sensemaking activities? A longitudinal experiment was conducted with subjects who were enrolled in four sections of an upper-division Management Information Systems course in Systems Analysis and Design in consecutive semesters. Lectures and class-activities were identical in all four sections except that group support systems technology (GSS) was used by the second-semester classes, the treatment group. Student teams in all sections completed a semester-long ISD project.
APA, Harvard, Vancouver, ISO, and other styles
4

Chau, Michael C. "Searching and mining the Web for personalized and specialized information." Diss., The University of Arizona, 2003. http://hdl.handle.net/10150/280260.

Full text
Abstract:
With the rapid growth of the Web, users are often faced with the problem of information overload and find it difficult to search for relevant and useful information on the Web. Besides general-purpose search engines, there exist some alternative approaches that can help users perform searches on the Web more effectively and efficiently. Personalized search agents and specialized search engines are two such approaches. The goal of this dissertation is to study how machine learning and artificial intelligence techniques can be used to improve these approaches. A system development research process was adopted as the methodology in this dissertation. In the first part of the dissertation, five different personalized search agents, namely CI Spider, Meta Spider, Cancer Spider, Nano Spider, and Collaborative Spider, were developed. These spiders combine Web searching with various techniques such as noun phrasing, text clustering, and multi-agent technologies to help satisfy users' information needs in different domains and different contexts. Individual experiments were designed and conducted to evaluate the proposed approach and the experimental results showed that the prototype systems performed better than or comparable to traditional search methods. The second part of the dissertation aims to investigate how artificial intelligence techniques can be used to facilitate the development of specialized search engines. A Hopfield Net spider was proposed to locate from the Web URLs that are relevant to a given domain. A feature-based machine-learning text classifier also was proposed to perform filtering on Web pages. A prototype system was built for each approach. Both systems were evaluated and the results demonstrated that they both outperformed traditional approaches. This dissertation has two main contributions. Firstly, it demonstrated how machine learning and artificial intelligence techniques can be used to improve and enhance the development of personalized search agents and specialized search engines. Secondly, it provided a set of tools that can facilitate users in their Web searching and Web mining activities in various contexts.
APA, Harvard, Vancouver, ISO, and other styles
5

Celaya, Tracy A. "Cloud-Based Computing and human resource management performance| A Delphi study." Thesis, University of Phoenix, 2016. http://pqdtopen.proquest.com/#viewpdf?dispub=10004286.

Full text
Abstract:

The purpose of this qualitative study with a modified Delphi research design was to understand the reasons human resource (HR) leaders are slow to implement Cloud-based technologies and potentially identify how Cloud-Based Computing influences human resource management (HRM) and HR effectiveness, and potentially the overall performance of the organization. Business executives and HR leaders acknowledge the effect of technology on business processes and strategies, and the leader's influence on technology implementation and adoption. Cloud-Based Computing is fast becoming the standard for conducting HR processes and HR leaders must be prepared to implement the change effectively. Study findings revealed characteristics demonstrated by HR leaders successfully implementing cloud technology, best practices for successful implementation, factors championing and challenging Cloud-Based Computing adoption, and effects on HRM and organizational performance as a result of using Cloud-Based Computing. The outcomes of this study may provide the foundation of a model for implementing Cloud-Based Computing, a leadership model including characteristics of technology early adopters in HR, and identify factors impeding adoption and may assist HR leaders in creating effective change management strategies for adopting and implementing Cloud-Based Computing. Findings and recommendation from this study will enable HR professionals and leaders to make informed decisions on the adoption of Cloud-Based Computing and improve the effectiveness, efficiency, and strategic capability of HR.

APA, Harvard, Vancouver, ISO, and other styles
6

Thomas, Maurice A. "Evaluating Electronic Health Records Interoperability Symbiotic Relationship to Information Management Governance Security Risks." Thesis, Northcentral University, 2019. http://pqdtopen.proquest.com/#viewpdf?dispub=13808526.

Full text
Abstract:

A major initiative in the U.S. healthcare care industry is to establish a nationwide health information network securing the sharing of information between all involved U.S. healthcare stakeholders. However, implementing an interoperability solution is a massive, complex, and enduring effort with significant challenges such as inconsistent technology and data standards, as well as complex privacy and security issues. The purpose of this qualitative, case study is to examine the impacts of interoperability initiatives involving the U.S. government and to provide an understanding of the information governance and security risk as standards that are vendor-neutral and trustworthy. This qualitative case study was conducted using federal participants who are health information management (HIM) and health information technology (HIT) professionals working in the Washington DC metropolitan area. The participants' interview data revealed nine major themes; patient identification matching, payment claims and auditing, information sharing, data stewardship, regulatory compliance, technology enhancements, training and certification, standards optimization, and value-based care. The implication of the study's themes showed interoperability is beneficial to the healthcare industry, but there is a greater need for technology and data standardization, information governance, data stewardship, and a greater understanding of federal and state data privacy and security laws. Future recommendation for practices discussed; policy and regulatory adjustments to enhance auditing and compliance, establish a healthcare data ecosystem to improve data and information governance, and technology alternatives such as master data management and white space data. Recommendation for further research included expanding the sample population to compare other federal organizations or the United Kingdom's HIT interoperability project initiative.

APA, Harvard, Vancouver, ISO, and other styles
7

Waldner, Bruce W. "Determining Relationships Between Technology Acceptance and Employee Attitudes Toward Automated Workflows in the Oil Industry." Thesis, Northcentral University, 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=3577845.

Full text
Abstract:

Automated workflows are used to assist petroleum engineers in maximizing the productivity of oil reservoirs. However, for a workflow to be successful, end users must adopt the workflow. The purpose of this quantitative, ex post facto, correlational study was to determine the relationship between acceptance of automated workflow technology and attitudes toward automated workflow adoption. Participants included a cluster sample of 100 randomly selected petroleum engineers from five randomly chosen companies in the oil industry in the Middle East. Predictor variables were measured with a version of the Technology Acceptance Model (TAM) modified to address the issue of workflow adoption. The outcome variable was user attitude to automated workflow adoption, as measured with the Attitude subscale of the Theory of Reasoned Action (TRA) scale. A single multiple linear regression model was computed to answer all research questions. The results for the overall model were significant, adjusted R 2 = .43, F(6, 93) = 13.28, p < .001. After correcting for the number of years of engineering experience of the participants, attitudes toward using automated workflows were predicted by perceived ease of use, β = 0.30, p = .01; perceived usefulness, β = 0.24, p = .03; and computer self-efficacy, β = 0.20,p = .02. The level of automation, β = 0.12, p = .17, and perceived level of support, β = 0.08, p = .32, did not predict attitudes. The number of years of engineering experience was unrelated to attitudes to automated workflow adoption. These findings indicated three factors that have an effect on adoption of automated workflows: perceived ease of use, perceived usefulness, and the level of computer self-efficacy of the end-user. Areas for future research included confirmation of the reliability of the Level of Automation and the Perceived Level of Support subscales designed for this study. There is also a need for future research regarding how the level of automation and level of support affect the adoption of other technologies.

APA, Harvard, Vancouver, ISO, and other styles
8

Clay, Paul F. "Factors contributing to user choice between codification and personalization-based knowledge management systems a task-technology fit perspective /." [Bloomington, Ind.] : Indiana University, 2006. http://gateway.proquest.com/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqdiss&rft_dat=xri:pqdiss:3219901.

Full text
Abstract:
Thesis (Ph.D.)--Indiana University, Kelley School of Business, 2006.
Source: Dissertation Abstracts International, Volume: 67-06, Section: A, page: 2227. Adviser: Alan R. Dennis. "Title from dissertation home page (viewed June 21, 2007)."
APA, Harvard, Vancouver, ISO, and other styles
9

Zhao, Huimin. "Combining schema and instance information for integrating heterogeneous databases: An analytical approach and empirical evaluation." Diss., The University of Arizona, 2002. http://hdl.handle.net/10150/280014.

Full text
Abstract:
Critical to semantic integration of heterogeneous data sources, determining the semantic correspondences among the data sources is a very complex and resource-consuming task and demands automated support. In this dissertation, we propose a comprehensive approach to detecting both schema-level and instance-level semantic correspondences from heterogeneous data sources. Semantic correspondences on the two levels are identified alternately and incrementally in an iterative procedure. Statistical cluster analysis methods and the Self-Organizing Map (SOM) neural network method are used first to identify similar schema elements (i.e., relations and attributes). Based on the identified schema-level correspondences, classification techniques drawn from statistical pattern recognition, machine learning, and artificial neural networks are then used to identify matching tuples. Multiple classifiers are combined in various ways, such as bagging, boosting, concatenating, and stacking, to improve classification accuracy. Statistical analysis techniques, such as correlation and regression, are then applied to a preliminary integrated data set to evaluate the relationships among schema elements more accurately. Improved schema-level correspondences are fed back into the identification of instance-level correspondences, resulting in a loop in the overall procedure. Empirical evaluation using real-world and simulated data that has been performed is described to demonstrate the utility of the proposed multi-level, multi-technique approach to detecting semantic correspondences from heterogeneous data sources.
APA, Harvard, Vancouver, ISO, and other styles
10

Ganesan, Shankaranarayanan. "Dynamic schema evolution in a heterogeneous database environment: A graph theoretic approach." Diss., The University of Arizona, 1998. http://hdl.handle.net/10150/282767.

Full text
Abstract:
The objective of this dissertation is to create a theoretical framework and mechanisms for automating dynamic schema evolution in a heterogeneous database environment. The structure or schema of databases changes over time. Accommodating changes to the schema without loss of existing data and without significantly affecting the day to day operation of the database is the management of dynamic schema evolution. To address the problem of schema evolution in a heterogeneous database environment, we first propose a comprehensive taxonomy of schema changes and examine their implications. We then propose a formal methodology for managing schema evolution using graph theory with a well-defined set of operators and graph-based algorithms for tracking and propagating schema changes. We show that these operators and algorithms preserve the consistency and correctness of the schema following the changes. The complete framework is embedded in prototype software system called SEMAD (Schema Evolution Management ADvisor). We evaluate the system for its usefulness by conducting exploratory case studies using two different heterogeneous database domains, viz., a University database environment and a scientific database environment that is used by atmospheric scientists and hydrologists. The results of the exploratory case studies supported the hypothesis that SEMAD does help database administrators in their tasks. The results indicate that SEMAD helps the administrators identify and incorporate changes better than performing these tasks manually. An important overhead cost in SEMAD is the creation of the semantic data model, capturing the meta data associated with the model, and defining the mapping information that relates the model and the set of underlying databases. This task is a one-time effort that is performed at the beginning. The subsequent changes are incrementally captured by SEMAD. However, the benefits of using SEMAD in dynamically managing schema evolution appear to offset this overhead cost.
APA, Harvard, Vancouver, ISO, and other styles
11

Rodgers, Thomas Lee. "Software inspections: Collaboration and feedback." Diss., The University of Arizona, 1999. http://hdl.handle.net/10150/289015.

Full text
Abstract:
This dissertation studies impact of collaboration and feedback on software inspection productivity. Used as a software-engineering validation technique, software inspections can be a cost-effective method for identifying latent issues (defects) within design documents and program code. For over two years, Baan Company has used a generalized Electronic Meeting System (EMS also referred to as GroupWare) to support software inspections and reported EMS to be more productive than face-to-face paper-based inspections (Genuchten, 1998). Validation of this phenomenon and initial development of a potentially more effective specialized EMS (SEMS) tool is the basis of this dissertation. Explanations of the collaborative phenomenon are presented within a theoretical framework along with testable hypotheses. The framework is derived from Media Synchronicity Theory (Dennis and Valacich) and Focus Theory of Productivity (Briggs and Nunamaker). Two main research questions are explored. (1) Do collaboration tools improve software inspection productivity? (2) Can feedback dimensions that significantly improve productivity be identified and incorporated within software inspections? The first research question is supported. In a detailed reevaluation of the Baan study, EMS inspections are shown to be 32% more efficient than paper-based inspections. During the subsequent period, the results were more pronounced with EMS inspections being 66% more efficient even controlling for inspector proficiency. Significantly more conveying communication than convergent communication occurs during inspection meetings. EMS inspections enable more deliberation, less attention for communication, and more attention for information access compared to face-to-face paper-based inspections. The second research question is explored. Surveys and analysis probe some previously unexplored feedback dimensions (review rate, inspector proficiency and inspection process maturity). Experienced inspectors are surveyed regarding process maturity, inspector proficiency, and collaborative aspects of inspections. Preparation and review rates are necessary but not sufficient to explain productivity. Inspector proficiency is perceived to be important and multi-dimensional. Participation by highly proficient inspectors resulted in 49-76% more effective inspections. Significant inspection process variations exist within mature development organizations. Based on theory and experiences, the SEMS inspection tool is developed and a quasi-experiment proposed. Initial results using the SEMS inspection tool are reported and suggestions made for future enhancements.
APA, Harvard, Vancouver, ISO, and other styles
12

van, Ketwich Willem. "IT Governance of Cloud Computing| Performance Measures using an IT Outsourcing Perspective." Thesis, University of Melbourne (Australia), 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=1527429.

Full text
Abstract:

With the advent of cloud computing and the success of the cloud computing industry, organisations are beginning to adopt this service model and technology at an increasing rate. As the rate and level of use increases, organisations are faced with how best to govern these investments and obtain maximum benefit from the services offered by providers. This includes measuring the performance of these services, the corresponding organisational performance and the associated business value generated. In investigating these areas, this study compares cloud computing and IT outsourcing. It is found that while cloud measures relate, to a great extent, to the operational level of an organisation, IT outsourcing measures are concerned more with the strategic level. This highlights that cloud computing lacks strategic measures and that measures from IT outsourcing may be adopted to fill this gap.

APA, Harvard, Vancouver, ISO, and other styles
13

Wei, Chih-Ping 1965. "Schema management for large-scale multidatabase systems." Diss., The University of Arizona, 1996. http://hdl.handle.net/10150/290610.

Full text
Abstract:
Advances in networking and database technologies have made the concept of global information sharing possible. A rapidly growing number of applications require access to and manipulation of the data residing in multiple pre-existing database systems, which are usually autonomous and heterogeneous. A promising approach to the problems of interoperating multiple heterogeneous database systems is the construction of multidatabase systems. Among all of the research issues concerning multidatabase systems, schema management which involves with the management of various schemas at different levels in a dynamic environment has been largely overlooked in the previous research. Two most important research in schema management have been identified: schema translation and schema integration. The need for declarative and extensible approach to schema translation and the support for schema integration are accentuated in a large-scale environment. This dissertation presents a construct-equivalence-based methodology based on the implications of semantics characteristics of data models for schema translation and schema integration. The research was undertaken for the purposes of (1) overcoming the methodological inadequacies of existing schema translation approaches and the conventional schema integration process for large-scale MDBSs, (2) providing an integrated methodology for schema translation and schema normalization whose similarities of problem formulation has not been previously recognized, (3) inductively learning model schemas that provide a basis for declaratively specifying construct equivalences for schema translation and schema normalization. The methodology is based on a metamodel (Synthesized Object-Oriented Entity-Relationship (SOOER) model), an inductive metamodeling approach (Abstraction Induction Technique), a declarative construct equivalence representation (Construct Equivalence Assertion Language, CEAL), and its associated transformation and reasoning methods. The results of evaluation studies showed that Abstraction Induction Technique inductively learned satisfactory model schemas. CEAL's expressiveness and adequacy in meeting its design principles, well-defined construct equivalence transformation and reasoning methods, as well as the advantages realized by the construct-equivalence-based schema translation and schema normalization suggested that the construct-equivalence-based methodology be a promising approach for large-scale MDBSs.
APA, Harvard, Vancouver, ISO, and other styles
14

Brown, Mary Erin. "Data-Driven Decision Making as a Tool to Improve Software Development Productivity." Thesis, Walden University, 2013. http://pqdtopen.proquest.com/#viewpdf?dispub=3591716.

Full text
Abstract:

The worldwide software project failure rate, based on a survey of information technology software manager's view of user satisfaction, product quality, and staff productivity, is estimated to be between 24% and 36% and software project success has not kept pace with the advances in hardware. The problem addressed by this study was the limited information about software managers' experiences with data-driven decision making (DDD) in agile software organizations as a tool to improve software development productivity. The purpose of this phenomenological study was to explore how agile software managers view DDD as a tool to improve software development productivity and to understand how agile software development organizations may use DDD now and in the future to improve software development productivity. Research questions asked about software managers', project managers', and agile coaches' lived experiences with DDD via a set of interview questions. The conceptual framework for the research was based on the 3 critical dimensions of software organization productivity improvement: people, process, and tools, which were defined by the Software Engineering Institute's Capability Maturity Model Integrated published in 2010. Organizations focus on processes to align the people, procedures and methods, and tools and equipment to improve productivity. Positive social change could result from a better understanding of DDD in an agile software development environment; this increased understanding of DDD could enable organizations to create more products, offer more jobs, and better compete in a global economy.

APA, Harvard, Vancouver, ISO, and other styles
15

Hammoud, Wissam. "Attributes effecting software testing estimation; is organizational trust an issue?" Thesis, University of Phoenix, 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=3583478.

Full text
Abstract:

This quantitative correlational research explored the potential association between the levels of organizational trust and the software testing estimation. This was conducted by exploring the relationships between organizational trust, tester’s expertise, organizational technology used, and the number of hours, number of testers, and time-coding estimated by the software testers. The research conducted on a software testing department of a health insurance organization, employed the use of the Organizational Trust Inventory- Short Form (OTI-SF) developed by Philip Bromiley and Larry Cummings and revealed a strong relationship between organizational trust and software testing estimation. The research reviews historical theories of organizational trust and include a deep discussion about software testing practices and software testing estimation. By examining the significant impact of organizational trust on project estimating and time-coding in this research, software testing leaders can benefit from this research to improve project planning and managing process by improving the levels of trust within their organizations.

APA, Harvard, Vancouver, ISO, and other styles
16

Herman, Deirdre. "Asset Reuse of Images From a Repository." Thesis, Walden University, 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=3610257.

Full text
Abstract:

According to Markus's theory of reuse, when digital repositories are deployed to collect and distribute organizational assets, they supposedly help ensure accountability, extend information exchange, and improve productivity. Such repositories require a large investment due to the continuing costs of hardware, software, user licenses, training, and technical support. The problem addressed in this study was the lack of evidence in the literature on whether users in fact reused enough digital assets in repositories to justify the investment. The objective of the study was to investigate the organizational value of repositories to better inform architectural, construction, software and other industries whether repositories are worth the investment. This study was designed to examine asset reuse of medical images at a health information publisher. The research question focused on the amount of asset reuse over time, which was determined from existing repository transaction logs generated over an 8-year period by all users. A longitudinal census data analysis of archival research was performed on the entire dataset of 85,250 transaction logs. The results showed that 42 users downloaded those assets, including 11,059 images, indicating that the repository was used by sufficient users at this publisher of about 80 employees. From those images, 1,443 medical images were reused for new product development, showing a minimal asset reuse rate of 13%. Assistants (42%), writers (20%), and librarians (16%) were the primary users of this repository. Collectively, these results demonstrated the value of repositories in improving organizational productivity—through reuse of existing digital assets such as medical images to avoid unnecessary duplication costs—for social change and economic transformation.

APA, Harvard, Vancouver, ISO, and other styles
17

Kwan, Denise. "Senior librarians' perceptions on successful leadership skills| A case study." Thesis, University of Phoenix, 2013. http://pqdtopen.proquest.com/#viewpdf?dispub=3572917.

Full text
Abstract:

The purpose of this qualitative case study was to explore senior librarians’ perceptions of successful leadership skills in the 21st century. Library organizations in the postindustrial environment of the 21st century operate in a new context involving digitization, global libraries, cultural diversity, cultural intelligence, and cross-cultural management. This new context requires a postindustrial paradigm of leadership emphasizing influence and collaboration as defined by Rost (1991). The purposive sample used for this case study consisted of 10 senior library leaders. The data gathered consisted of demographic information and responses to six open-ended interview questions. Data were entered, coded, and analyzed for themes and patterns using NVivo 10 software. The analysis revealed significant emergent themes relating to successful library leadership skills. Results from this case study suggest successful library leadership skills in the 21st century involve two levels: foundational level and interpersonal level. At the foundational level are technical and knowledge skills, which are the building blocks for the next level of interpersonal skills. Interweaving these interpersonal skills are persuasion skills and collaborative skills, both at the core of the postindustrial paradigm of leadership. These two levels of skills, with an emphasis on persuasion skills, should form the basis of succession planning programs for next generation librarians. Implementing such programs could lead to increased leadership diversity, greater job satisfaction, improved job performance and effectiveness, helping the retention of librarians, and easing staff shortage. Further studies are recommended.

APA, Harvard, Vancouver, ISO, and other styles
18

Ng, Tobun Dorbin. "A concept space approach to semantic exchange." Diss., The University of Arizona, 2000. http://hdl.handle.net/10150/289095.

Full text
Abstract:
This dissertation work investigates the use of information technologies that clarify semantic meaning to help users elaborate their information needs by providing library-specific knowledge to the information seeking process. The research involved two interdependent semantic technologies: concept space consultation and library-specific, domain-specific, automatically generated concept spaces. The concept space consultation phase used spreading activation algorithms--branch-and-bound and Hopfield net algorithms--to explore knowledge sources in specific domains. This research demonstrated the comparable effectiveness of exploration of a library database using a man-made classification scheme and thesaurus as opposed to an automatically generated concept space. The results showed that the use of spreading activation algorithms identified more relevant concepts than the use of the manual browsing method. The concept space technique automatically identifies and extracts concept from a library collection while at the same time computing the strength of associations between concepts. This research demonstrated that the concept space technique was able to create human-recognizable concepts and their associations. In addition, the technique could be scaled to generate very large library-specific concept spaces for a very large underlying library collection. Moreover, the interdependent use of both semantic technologies creates a semantic medium for users and library-specific knowledge sources to exchange content with context--context in user information need and that in corporeal knowledge.
APA, Harvard, Vancouver, ISO, and other styles
19

Taylor, Stella D. "Facilitating Browsing with Information Visualization: Is Animation a Powerful Scent?" Diss., Temple University Libraries, 2009. http://cdm16002.contentdm.oclc.org/cdm/ref/collection/p245801coll10/id/35311.

Full text
Abstract:
Business Administration
Ph.D.
Search engines make vast amounts of information available to Internet users. Two types of tasks users engage in using search engines are closed-ended and open-ended. For closed-ended tasks, individuals have narrow objectives that require finding specific results. For open-ended tasks, individuals only have general objectives that require finding as much relevant information as possible about a topic, which can be difficult when large numbers of both relevant and irrelevant results are returned from a query. This can also leave users in a state of information overload. Some search engines have incorporated information visualization techniques (combining cognitive senses with visual cues that allow for better understanding the information) to facilitate browsing through results in order to reduce information overload. However, there is little research that identifies which visual cues are the most desirable for the presentation of search results. According to information foraging theory, cues that have strong scents will help users find information faster. In this study, we investigate the effects of augmenting visualizations with animation as a powerful scent to help users more easily identify relevant information in search engine results. This study employs cognitive fit theory to study the effect of different information formats on users' performance in completing the two different tasks. Overall, we find evidence that the effectiveness of cues such as animation is task-dependent. For example, we find that visualizations with animation are less effective than a standard textual display for subjects performing closed-ended web search tasks. The results of this study have strong implications for integrating appropriate cues into visualizations in order to help people find information.
Temple University--Theses
APA, Harvard, Vancouver, ISO, and other styles
20

Williams, Henry. "Black Business Owners Overcoming Barriers in Texas." ScholarWorks, 2018. https://scholarworks.waldenu.edu/dissertations/5958.

Full text
Abstract:
Black entrepreneurs in the state of Texas reported that they did not achieve the same level of success as nonminority enterprises. Many Black entrepreneurs have obtained education from top tier institutions and have the managerial experiences, skills, and working knowledge that facilitate business success, but they are not able to sustain business growth. The purpose of this phenomenological study was to explore the lived experiences of Black Texan entrepreneurs. This research was established using the conceptual framework of human capital. The overarching research question concerned the lived experiences of Black Texas entrepreneurs who were not as successful in business as their counterparts. Data were collected in the form of semistructured face-to-face interviews with 14 Black participants, using the African American Chamber of Commerce to facilitate networking among Black entrepreneurs. NVivo 11 software was used for data codification and thematic reduction. Emergent themes included (a) feeling good about being in business, (b) resourcefulness in business, (c) network advertising systems, (d) education, (e) having alternative plans if business closes, (f) feeling independent, and (g) activeness in business. The implications for positive social change include the potential for advising researchers on barriers confronted by Black entrepreneurs and potential steps to overcome those barriers. Additionally, scholars have recognized the need for greater understanding of how businesses are structured and managed by Black business-owners because this knowledge is vital to creating business prosperity for Black entrepreneurs who have dreams of becoming business owners.
APA, Harvard, Vancouver, ISO, and other styles
21

Mynyk, John. "Information technology programming standards and annual project maintenance costs." Thesis, University of Phoenix, 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=3578624.

Full text
Abstract:

Organizations that depend on the use of IT in their business models must maintain their systems and keep their systems current to survive (Filipek, 2008; Kulkarni, Kumar, Mookerjee, & Sethi, 2009; Unterkalmsteiner et al., 2012). As most IT departments allocate as much as 80% of their budget to maintain stability while leaving only the other 20% to allow improvements (Telea et al., 2010), high cost of stability may be a reason many IT organizations cannot afford efficient staffing and even jeopardize the existence of the organization (Filipek, 2008; Talib, Abdullah, Atan, & Murad, 2010). The purpose of this exploratory mixed methods study was to discover the IT programming standards used in IT departments that predict a decrease in project maintenance costs. This study employed an exploratory mixed methods data collection and analysis to develop and test a collection of universal programming standards. The qualitative portion of the study resulted in a list of IT programming standards from the Fortune 20 companies of 2011. Surveyed from IT departments in the Fortune 500 companies of 2011, the quantitative portion of this study correlate the degree of enforcement of each IT programming standard to a decrease in average project maintenance costs using a backward stepwise regression. Using a 95% confidence interval and a 5% margin of error (α = .05), the backward stepwise regression discarded 18 of the 22 IT programming standards. The remaining correlations give evidence that a) the more the department enforces waiting for feedback the higher the maintenance costs, b) the more the department enforces having the architectural team develop coding guidelines the lower the maintenance costs, and c) the more the IT department enforces the following of change management procedures, the higher the maintenance costs.

APA, Harvard, Vancouver, ISO, and other styles
22

Menon, Nirup 1967. "Productivity of information technology in the face of regulation and organizational choices: Evidence from the healthcare industry." Diss., The University of Arizona, 1997. http://hdl.handle.net/10150/282546.

Full text
Abstract:
The dissertation is on the impact of information technology (IT) on the economy. Theoretical and empirical analyses are reported in order to explain the productivity "paradox" surrounding IT. The techniques used in the study cover a wide range from simple deterministic production functions to panel data techniques and data envelopment analysis. The empirical study is set in the health care industry. A hospital, which is the the unit of analysis, is an organizational entity that provides an appropriate context for the study of the interactions between information technology, technological advancements, organizational factors and regulation. Overall, it was found that IT contributes in a positive manner to the production of healthcare services. However, measurement problems including "quality" of IT capital due to technological developments can lead to the underestimation of IT productivity. It is also seen that regulation plays a major role in the manner in which costs are allocated to input factors. In particular, the effect of the Prospective Payment System (PPS) legislated in 1983 is investigated. One of the findings of the cost function analysis is that, since PPS, hospitals have been moving toward cost containment. Non-parametric analyses is used to determine the allocative inefficiency in IT with respect to other inputs. The different techniques serve as means to triangulate the measure of IT impact on productivity and efficiency, and in addition, develop the theory underlying production techniques.
APA, Harvard, Vancouver, ISO, and other styles
23

Straw, Eric M. "Construction of a Conceptualization of Personal Knowledge within a Knowledge Management Perspective using Grounded Theory Methodology." Thesis, Nova Southeastern University, 2013. http://pqdtopen.proquest.com/#viewpdf?dispub=3590345.

Full text
Abstract:

The current research used grounded theory methodology (GTM) to construct a conceptualization of personal knowledge within a knowledge management (KM) perspective. The need for the current research was based on the use of just two categories of knowledge, explicit and tacit, within KM literature to explain diverse characteristics of personal knowledge. The construct of tacit knowledge has often been explicated and debated in KM literature. The debate over tacit knowledge arose from the complex epistemological roots of tacit knowing and the construct of tacit knowledge popularized by organizational knowledge creation theory. The ongoing debate over tacit knowledge in KM literature has shed little light on personal knowledge within a KM perspective. The current research set aside the debate over tacit knowledge and pursued the construct of personal knowledge from the perspective of the knower using GTM. Thirty-seven interviews were conducted with fourteen participants. Interviews were audio recorded and coding was accomplished with the qualitative data analysis software MAXQDA.

A total of eight categories were identified. These were organized into two groups. The core category being overwhelmed represented the absence of personal knowledge. The categories questioning self, seeking help, and microthinking fit under being overwhelmed. Together these categories were inverse indicators because they all decreased as knowledge acquisition progressed. The core category being confident represented the presence of personal knowledge. The categories remembering, multitasking, and speed fit under being overwhelmed. Together these categories were direct indicators because they all increased as knowledge acquisition progressed.

Three significant conclusions were drawn from the current research. These conclusions led to the conceptualization of personal knowledge from a KM perspective. The first significant conclusion was the conceptualization of a process of knowing as Integrated Complexity: From Overwhelmed to Confident (ICOC). The second significant conclusion was personal knowing as first-person epistemology is a universally lived experience that includes commitments to internal and external requirements as well as a bias toward integration. The third significant conclusion was personal knowledge can be viewed as a complex adaptive system. Finally, the current research concluded that personal knowledge within a KM perspective is a complex adaptive system maintained through acts of first-person epistemology.

APA, Harvard, Vancouver, ISO, and other styles
24

Lundahl, Philpot Eva. "Social media adoption and use among information technology professionals and implications for leadership." Thesis, University of Phoenix, 2013. http://pqdtopen.proquest.com/#viewpdf?dispub=3570894.

Full text
Abstract:

This sequential, mixed methods research addressed emerging social media use practices among IT professionals and explored lived experiences of senior IT leaders relative to successful organizational social media adoption and use. The study was informed by structuration theory and elements from the universal technology adoption and use (UTAUT) model, generation theory, and open leadership theory. In the first, quantitative descriptive research phase, an online survey was administered to describe IT professionals' uses of and attitudes toward social media in the workplace. Survey results based on 406 responses from IT professionals in the greater Seattle area indicated widespread use of different social media applications, and also showed that Millennial IT professionals use social media more extensively and are have more positive opinions about social media as compared to their older colleagues. Survey findings also indicated that an increasing number of employers are developing formal social media strategies and adopting policies and guidelines governing the use of social media in the workplace. The second, qualitative hermeneutic phenomenological research phase built on survey results and involved interviews with 13 senior IT leaders in the greater Seattle area. Findings indicated that despite the inherent user-driven nature of social media, senior leadership plays a key role in driving strategic social media adoption and in ensuring broad participation across generational cohorts and employee groups. Findings from the qualitative research phase further suggested that social media can help employees and stakeholders communicate and collaborate more effectively and efficiently, and that leaders can derive significant benefits from social media without compromising the integrity of their organizations.

APA, Harvard, Vancouver, ISO, and other styles
25

Reinicke, Bryan Alan. "Building a systems level theory of IS integration in mergers and acquisitions." [Bloomington, Ind.] : Indiana University, 2007. http://gateway.proquest.com/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqdiss&rft_dat=xri:pqdiss:3277979.

Full text
Abstract:
Thesis (Ph.D.)--Indiana University, Kelley School of Business, 2007.
Source: Dissertation Abstracts International, Volume: 68-09, Section: A, page: 3956. Adviser: Carol V. Brown. Title from dissertation home page (viewed May 5, 2008).
APA, Harvard, Vancouver, ISO, and other styles
26

Hu, Paul Jen-Hwa 1962. "Management of telemedicine technology in healthcare organizations: Technology acceptance, adoption, evaluation, and their implications." Diss., The University of Arizona, 1998. http://hdl.handle.net/10150/282579.

Full text
Abstract:
As an exciting information technology-based innovation, telemedicine has potential to enhance physicians' patient care and management, improve healthcare organizations' operations and performance, and cause a paradigmatic shift in health care toward a progressively emerging digital practice. Previous research has concentrated on technology developments and clinical applications and therefore offers limited discussion of technology management. Managing telemedicine technology in healthcare organizations is so complex and dynamic that it has been an important factor in the failure of many early telemedicine attempts. This dissertation research directly addressed organizational management of telemedicine technology. To deal with complexity and dynamism, the research took a multi-phase approach, using a research framework built upon a well-established theoretical foundation. Case study was used in the exploratory phase to provide detailed understanding of the underlying technology implementation process and to generate specific research questions or models for the subsequent descriptive/explanatory phase. Systematic linkage of these investigations was safeguarded by desired methodological triangulation. Findings from the case study and substantiating interviews identified technology acceptance, adoption and evaluation as problematic areas in organizational technology management. Findings of a survey study administered to most physicians practicing in public tertiary hospitals in Hong Kong suggested that perceived usefulness and ease of use, self-efficacy, and subjective norms were important to their accepting telemedicine technology. Similarly, a survey conducted with hospital executive officers, chiefs of service and center directors of all Hong Kong public healthcare establishments indicated that service needs, attitudes of medical staff, and the technology's benefits, risks and compatibility were essential to organizational technology adoption. In addition, results of an evaluative experimental study showed that the clinical decision- making of physicians can be improved through use of appropriate telemedicine technology. The combined findings suggested that these separate technology management issues were closely interrelated rather than isolated. Effects of a technology on physicians' patient care and management practice have important impacts on their technology acceptance, which, in turn, needs to be considered by their affiliating organization when making an adoption decision.
APA, Harvard, Vancouver, ISO, and other styles
27

Alao, 'Remi Kehinde Reuben. "Use of direct mail for improved electoral education that encourages civic behavior and election credibility." Thesis, University of Phoenix, 2013. http://pqdtopen.proquest.com/#viewpdf?dispub=3569144.

Full text
Abstract:

In this quantitative quasi-experimental study, correlation and regression analyses were used to test two research hypotheses. The experiment was to examine if a relationship existed between the mode of electoral information dissemination from the election organizing body (EMB) and voters’ behavior and attitudes associated with an electoral process. A Baptist church located in Otta, Nigeria gave permission to conduct the experiment within its premises, and 285 church members took part as participants in the study. Three main elements each of modern and traditional electoral information and communication channels including short message service (SMS), e-mail, surface post, newspaper, posters, and radio/television (TV) were tested to determine the validity of the research assumptions. Results of the study indicated voters’ preference for receiving direct electoral information from the organizer, preferably using mobile direct communication channels. There was an indication of knowledge and attitude changes because the electoral management body disseminated the electoral information directly to voters via direct mail. Knowledge and attitude changes could have implications for subsequent elections and other electoral management decisions. The implication of this study was that proper information system management could be a key remedy for unethical behavior during the electoral process. The suggestion following the result of the study was that if a sustainable, systematic planning and execution of electoral information and communication management is adopted, it could likely lead to improved voter knowledge and informed decision-making ability. It also has the prospect of reducing unethical stakeholder behavior during elections, and election of qualified candidates based on merit would be possible as may be demonstrated through informed participation by the electorates.

APA, Harvard, Vancouver, ISO, and other styles
28

Santanen, Eric Lawrence. "Directed brainstorming and the Cognitive Network Model of Creativity: An empirical investigation of cognitive factors related to the formation of creative solutions using an electronic brainstorming environment." Diss., The University of Arizona, 2002. http://hdl.handle.net/10150/279925.

Full text
Abstract:
This dissertation presents the Cognitive Network Model of Creativity. This causal model posits that creative solutions occur when new associations are formed between disparate elements from memory. The likelihood of forming new associations is a positive function of the disparity between these elements and an inverse function of the problem solver's cognitive load. Cognitive load is, in turn, a positive function of the disparity between elements and the quantity of stimuli per unit of time to which the problem solver is exposed. Cognitive load is also an inverse function of the extent to which elements may be combined. The disparity between elements is a positive function of stimuli diversity, while the extent to which elements may be combined is an inverse function of stimuli diversity. Thus, the Cognitive Network Model of creativity represents a highly plausible answer to the research question: "What is a basic cognitive mechanism responsible for producing creative solutions to a problem?" that is grounded in group support systems, cognitive psychology, problem solving, and creativity research. Sixty one four-person groups participated in one of two experimental problem solving tasks designed to evaluate the model. The solution space for each task was partitioned into five smaller domains based upon known criteria for good solutions. Four directed brainstorming prompts were derived from each domain. These twenty prompts were then arranged to create three treatment conditions with respect to stimuli diversity (low, medium, and high). In each treatment, one directed brainstorming prompt was delivered to the electronic brainstorming supported groups every two minutes; groups in the control condition received no facilitation. Initial findings consistent with the model suggest that people using directed electronic brainstorming produce higher concentrations of creative solutions than people using free brainstorming because directed brainstorming focuses the efforts of the problem solvers on specific goals while simultaneously providing ready access to discontiguous areas of memory that help problem solvers avoid bounded, familiar, and narrow thought patterns. Findings show directed brainstorming advantages for low stimuli diversity, resulting in sharply focused creative solutions, and high stimuli diversity, resulting in broadly focused creative solutions. Medium diversity facilitation bestowed no benefit.
APA, Harvard, Vancouver, ISO, and other styles
29

Shin, Bongsik 1960. "The implication of information technology in telework: Adoption model and influencing factors of communication media choice among teleworkers." Diss., The University of Arizona, 1997. http://hdl.handle.net/10150/288732.

Full text
Abstract:
As the post-industrial environment requires more flexibility in organizational operations, telework is gradually gaining an acceptance as a distributed organizational design. Academic research, however, has not been effective in providing rich theoretical and empirical support for the planning and implementation of telework at organizations. This dissertation is motivated to contribute telework research through the investigation of implications that general-purpose communication media have on distributive telework. First, an empirical study is performed to understand the role of individual-, contextual-, and social-level factors on the media choice behaviors of remotely scattered teleworkers. Then, examination is made on how the media choice and other individual and social factors affect the perception on information-carrying capability and productivity of a communication medium, email. For this investigation, a hypothetical model that depicts the relationship among the constructs is proposed and relevant hypotheses are developed. Prior to the empirical study, existing literatures of telework are reviewed and characterized for the conceptual analysis of the problems and issues in telework. It was recognized that the narrow focus of research on teleworkers and the lack of a theoretical foundation are impeding broad understanding of telework. Data analysis indicates that teleworkers' media choice is the result of dynamics of individual-, contextual-, and social-level variables. Management support as a social influence showed the strongest impact on teleworkers' media use. The study confirms that, though regarded as a lean medium, email could become an effective and rich communication tool through an active social structuration process. Teleworkers belonging to an email-oriented communication network not only recognized email as a rich medium, but also had higher perception of work productivity from its use. The study confirms that, when email is recognized as an information-rich, as well as a function-rich, medium by teleworkers, a telework program could be benefited from the reduced loss of internal processes and enhanced work productivity.
APA, Harvard, Vancouver, ISO, and other styles
30

Romano, Nicholas Charles 1963. "A Web-based system for collaboration." Diss., The University of Arizona, 1998. http://hdl.handle.net/10150/288860.

Full text
Abstract:
Today's complex environmental and organizational pressures lead to business teams distributed along the dimensions of space, time, and computing resources. Distributed teams often need to collaborate to solve complex problems together. Many information systems support simple information sharing, however group research has shown that productive problem solving extends far beyond this. Group Support Systems (GSS) researchers and product reviewers suggest that simple discussion tools fall short of supporting additional phases of group problem solving. Research into distributed collaboration is needed to understand this complex domain. This dissertation describes an investigation into distributed collaboration to design, develop, implement, evaluate, and iteratively refine a prototype World-Wide-Web (Web) based distributed GSS. The research addresses requirements derivation, architecture design, prototype implementation, evaluation, iterative refinement, and the nature of roles played by participants. The literature review examines the areas of meeting analysis, hypertext, the Web, GSS, distributed GSS, collaborative interface design, group facilitation and systems engineering. The systems development research method is applied according to the following stages: Conceptual Framework Development, Requirements Identification, Systems Architecture Development, Systems Design, Systems Implementation and Systems Evaluation. An evolutionary prototyping approach incorporates evaluator suggestions and evolving technology into the system. Systems requirements are derived from the literature review, interviews with dozens of GSS researchers, practitioners and developers and hundreds of users from around the world. The initial specification employs the data model and interface design of an existing face-to-face GSS. A Distributed GSS architecture, consisting of a Hypertext Markup Language (HTML)/JavaScript client interface and a centralized HyperText Transport Protocol (HTTP)/Common Graphical Interface (CGI) server, is proposed, designed, and implemented. The logical data model is extended for distributed collaboration through identification of relevant data entities and relationships. The interface is extended through iterative prototyping based on observations, user feedback and technical enhancements. Prototype systems functionality is extended based on lab and field observations and direct feedback from users. Research contributions include a new distributed architecture, knowledge about distributed GSS interfaces, functionality, facilitation, leadership and participation, a prototype for additional research, and knowledge about the processes and group dynamics for distributed teams.
APA, Harvard, Vancouver, ISO, and other styles
31

McQuaid, Michael Joseph. "The impact of time-based text visualization on situation awareness." Diss., The University of Arizona, 2003. http://hdl.handle.net/10150/290011.

Full text
Abstract:
Overabundance of information has been well documented as a drag on human performance. Theoretical constructs such as visual dominance have led to visualization systems as a palliative, integrating cumbersome amounts of incoming data into more tractable forms. Most such systems have accepted only numerical inputs. I designed, developed, and evaluated a time-based visualization system for textual information, using situation awareness theory to evaluate the system. The system analyzes texts, such as emails and instant messages, as they arrive on the desktop and presents a labeled display, using suffix tree clustering and multidimensional scaling, of similarity between texts, updated as the mix of incoming texts changes. The system aids decision makers subject to information overload. I evaluated the system's performance under a simulated crisis in which groups of 30 students played the role of advisors to the University president during a recreation of an actual campus shooting. Half of the students used the visualization system to receive input, while half used a competing text-based system. Subjects were evaluated using an instrument drawn from situation awareness theory and found to achieve higher levels of situation awareness using the visualization system. Additional measures collected provide some guidelines for design of text-oriented, time-based visualization systems.
APA, Harvard, Vancouver, ISO, and other styles
32

Aakhus, Mark Alan 1964. "The communication logics of computer-supported facilitative interventions: A study of the community of practice and social technologies surrounding the use of group decision support systems in process facilitation." Diss., The University of Arizona, 1997. http://hdl.handle.net/10150/288721.

Full text
Abstract:
Computer supported facilitation is a form of third party intervention that uses advanced information technology to deliver non-authoritative intervention on organizational decision making. The goal of this type of intervention is to create communication events where decision making and decision outcomes are collaboratively produced by those who have a stake in the decision. The facilitator's role is to assure decision making progress without taking sides or rendering a final decision. The obligations of facilitation form competing injunctions for practicing facilitation. Facilitative intervention must be performed so that it influences decision activity without influencing decision outcomes, facilitators must participate in decision making without becoming a party to the decision, and facilitators must enforce decision procedures without coercing participant acceptance of the procedures. The evolution of the field is marked by innovations in practice and role definitions that seek more effective means to reconcile the competing demands of the role and the changing circumstances of the intervention context. Computer supported facilitation is a technologically advanced form of intervention that combine skills of facilitators with the capacities of collaborative computing technology to more efficiently and effectively deliver decisions for organizations facing a choice. The facilitator designs and carries out interventions by using group decision support systems that enable anonymous participation, simultaneous communication of ideas, geographic and chronological distribution of participation, and the electronic storage of contributions. This investigation finds that while technical advances help facilitators overcome the numerous barriers to decision making communication, the advances in technique and technology are prescriptions for decision making communication built on inadequate descriptive assumptions about the nature of communication. The community of facilitation practice and its technologies operate on the dubious assumption that communication process and content are in fact distinct. The community of practice, however, is caught up in preserving this distinction as its solution to the paradoxes of doing non-authoritative intervention. The dissertation demonstrates this state of affairs by showing the set of premises for facilitative action embodied in the process management view of the practice, the methods of transparency work which uphold intervention neutrality, and the way the community treats an innovation on practice.
APA, Harvard, Vancouver, ISO, and other styles
33

Sibona, Christopher. "Predicting social networking sites continuance intention| Should I stay or should I go?" Thesis, University of Colorado at Denver, 2015. http://pqdtopen.proquest.com/#viewpdf?dispub=3667242.

Full text
Abstract:

This research develops and tests models to predict continuance intention on social networking sites. The models adds new factors which are relevant to social networking sites continuance intention. The social networking site continuance model adds five factors: personal innovativeness, habit, attitude toward alternatives, interpersonal influence, and consumer switching costs to enhance the predictive power of information systems continuance. Interpersonal influence, alternative perceptions and procedural and relational costs are theorized to have a direct effect on continuance intention. Personal innovativeness and habit are theorized to have a direct and moderating effects on continuance intention. The results have a large positive effect of the explanatory power in explaining more of the variance of continuance intention on a social networking site. The information systems (IS) continuance model explains approximately 66.8% of the variance and the social networking site continuance model with the five added factors explains 76.7% of the variance and is considered to have a large effect in the explained variance. All of the factors have statistical significance; the factors with the largest path coefficients are, in order, satisfaction & perceived usefulness (β = 0.3686), consumer switching costs (β = 0.2496), alternative perceptions (β = -0.2069), habit (β = 0.1642), personal innovativeness (β = -0.0589) and interpersonal influence (β = -0.0451). Habit and personal innovativeness, as moderators, were not statistically significant and did not substantially aid in the interpretation of the factors. The research helps explains the relevant factors for why users of social networking sites will continue to use or abandon a site.

APA, Harvard, Vancouver, ISO, and other styles
34

Malopinsky, Larissa V. "Facilitating organizational change the use of activity theory as a framework for social construction of strategic knowledge /." [Bloomington, Ind.] : Indiana University, 2007. http://gateway.proquest.com/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqdiss&rft_dat=xri:pqdiss:3297945.

Full text
Abstract:
Thesis (Ph.D.)--Indiana University, Dept. of Instructional Systems Technology, 2007.
Source: Dissertation Abstracts International, Volume: 69-03, Section: A, page: 1062. Adviser: Thomas M. Schwen. Title from dissertation home page (viewed Sept. 30, 2008).
APA, Harvard, Vancouver, ISO, and other styles
35

Robert, Lionel P. "Social capital and knowledge integration in virtual teams." [Bloomington, Ind.] : Indiana University, 2006. http://gateway.proquest.com/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqdiss&rft_dat=xri:pqdiss:3243774.

Full text
Abstract:
Thesis (Ph.D.)--Indiana University, Kelley School of Business, 2006.
Title from PDF t.p. (viewed Nov. 17, 2008). Source: Dissertation Abstracts International, Volume: 67-12, Section: A, page: 4612. Adviser: Alan R. Dennis.
APA, Harvard, Vancouver, ISO, and other styles
36

Tian, Xiaoguang. "Hybrid Models in Automobile Insurance: Technology Adoption and Customer Relations." Thesis, University of North Texas, 2019. https://digital.library.unt.edu/ark:/67531/metadc1538717/.

Full text
Abstract:
Customer relationship management (CRM), a primary activity in the business value chain to relate to the customer, involves solicitation, analysis, and the use of the knowledge about the customer to provide goods and services through effective and efficient methods. It is a wise strategy and source of competitive advantage for customer behavior understanding and business performance management. The use of information technology (IT) in CRM allows companies to simplify their processes, to integrate product or service related decision making with the business strategies, and to optimize their operations by embracing analytical techniques. The insurance industry is facing unprecedented challenges and decisions in this data-driven business paradigm. It is a strategic necessity for customer-centric insurers to utilize emerging IT capability to support interactions between customers and business operations. The research in the dissertation seeks to provide insights into the application of early technology innovation and data-driven strategies by investigating the following two groups of CRM technology issues: technology adoption and data-driven technology application. Through three essays, the dissertation explores the use of information technology and data analytical tools to provide insight into how automobile insurance companies make decisions regarding their relationships with their customers. The results from these studies provide a framework for managers to devise effective approaches to enhancing the performance of their business.
APA, Harvard, Vancouver, ISO, and other styles
37

Liginlal, Divakaran. "Building fuzzy front-end decision support systems for new product information in global telecommunication markets: A measure theoretical approach." Diss., The University of Arizona, 1999. http://hdl.handle.net/10150/284835.

Full text
Abstract:
In today's highly competitive business environment, innovation and new product introduction are recognized as the sustaining forces of corporate success. The early phases of new product development, collectively known as the 'front-end', are crucial to the success of new products. Building a fuzzy front-end decision support system, balancing the needs for analytical soundness and model robustness while incorporating decision-maker's subjectivity and adaptability to different business situations, is a challenging task. A process model and a structural model focusing on the different forms of uncertainties involved in new product introduction in a global telecommunication market are presented in this dissertation. Fuzzy measure theory and fuzzy set theory are used to build a quantitative model of the executive decision-process at the front-end. Solutions to the problem of exponential complexity in defining fuzzy measures are also proposed. The notion of constrained fiizzy integrals demonstrates how the fuzzy measure-theoretical model integrates resource allocation in the presence of project interactions. Forging links between business strategies and expert evaluations of critical success factors is attempted through fuzzy rule-based techniques in the framework of the proposed model. Interviews with new product managers of several American business firms have confirmed the need for building an intelligent front-end decision support system for new product development. The outline of a fuzzy systems development methodology and the design of a proof-of-concept prototype serve as significant contributions of this research work toward this end. In the context of executive decision making, a usability inspection of the prototype is carried out and results are discussed. A computational analysis, based upon methods of tactical systems simulation, measures the rank order consistency of the fuzzy measure theoretical approach in comparison with two competing fuzzy multiple attribute decision models under structural variations of the underlying models. The results demonstrate that (1) the modeling of the fuzzy numbers representing the linguistic variables, (2) the selection of the granularity of the linguistic scales, and (3) the selection of the model dimensions significantly affect the quality of the decisions suggested by the decision aid. A comprehensive plan for future validation of the decision aid is also presented.
APA, Harvard, Vancouver, ISO, and other styles
38

Shelby, Kenneth R. Jr. "Systems Engineering Knowledge Asset (SEKA) Management for Higher Performing Engineering Teams| People, Process and Technology toward Effective Knowledge-Workers." Thesis, The George Washington University, 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=3611730.

Full text
Abstract:

Systems engineering teams' value-creation for enterprises is slower than possible due to inefficiencies in communication, learning, common knowledge collaboration and leadership conduct. This dissertation outlines the surrounding people, process and technology dimensions for higher performing engineering teams. It describes a true experiment investigation of opportunities to improve communication, learning and common knowledge collaboration.

The art and practice of Systems Engineering contributes business value by orchestrating large numbers of knowledge-workers as engineering teams in the achievement of complex goals. During the creation of new systems, engineering team performance modulates business efficiencies to realize those complex goals. Higher performing engineering teams share a vision providing purpose, rely on personal knowledge convolved with collaborators knowledge to unleash potential, leverage common knowledge in their team mental models, and execute synergistically. Why do non-high performing teams exist? Culture change is hard. Humans prefer the familiar. Without Leadership and systematic enablement, teams usually do not naturally find the high performing team traits.

This research investigates a unique Information Technology based Systems Engineering Knowledge Asset (SEKA) management mechanism. The selected mechanism integrates multiple techniques for improved collaboration efficacy. The research methodology was a modified true experiment with dual-posttest only, using an A and B group for comparative controls. Research findings reflect, with 99% confidence, that SEKA represented in 3-way Multiple Informational Representations Required of Referent (MIRRoR) knowledge constructs improves systems engineering teams' consumption of a common knowledge base.

Engineering teams can consume a set of information, which generates knowledge common with their collaborators, in a shorter period. More knowledge that is common facilitates increased ability to collaborate. Increased collaboration accelerates team learning, leading to shorter systems delivery schedules, lower cost to produce and earlier actionable intelligence. Shorter delivery times increase customer satisfaction; lower costs improve profit margin potential, and earlier actionable intelligence supports "left of boom" intervention.

APA, Harvard, Vancouver, ISO, and other styles
39

Ogunlade, Jacob Olusola. "Assessing the collaborative knowledge management of the market dominant organization." ScholarWorks, 2009. https://scholarworks.waldenu.edu/dissertations/683.

Full text
Abstract:
Dominant firms enjoy economic strengths which enable them to compete effectively in relevant markets through the use of collaborative knowledge management (CKM). While the literature is replete with general guiding principles for companies to adopt successful business strategies, there is very limited empirical research on effectively using CKM to improve company performance and market domination. The purpose of this study was to evaluate strategies for information sharing by companies to achieve better operations management and control, a wider range of customers, and stronger competitive edge in the global economy. Epistemological foundation for the study was provided by the literature on knowledge management and organizational dynamics. Data were collected by an electronically self-administered questionnaire on a convenience sample of 80 employees of three small businesses in Memphis, Tennessee. A quantitative method using Poisson regression was applied to test the hypotheses about relationships between six independent variables of value proposition, culture building, responsibilities, information technology, approaches and assessment and the dependent variable, collaborative knowledge management. Results indicate that value proposition, information technology, and building an organizational culture of responsibilities and best practices play significant roles in effective CKM. Social change implications of the study suggest that high-intensity collaborative knowledge management would produce creative leaders and workers, improved leader-worker collaboration, and more effective use of information technologies in organizational intelligence and decision making.
APA, Harvard, Vancouver, ISO, and other styles
40

Herman, Elizabeth M. "Exploring knowledge loss in the contact center: Key loss areas, contributing factors, and performance metric implications." ScholarWorks, 2009. https://scholarworks.waldenu.edu/dissertations/704.

Full text
Abstract:
Knowledge loss, or organizational forgetting, is often overlooked in knowledge management frameworks, yet it costs organizations money, personnel, efficiency, and customer service. The purpose of this mixed model case study was to understand, using a model of forgetting as the conceptual framework, where and why loss occurred and to examine performance implications. An inbound telephone contact center was studied because of prolific changes in that industry regarding knowledge complexity and performance. The researcher interviewed 20 participants, observed 63 calls, and reviewed 3 months of performance data including average handle time, first call resolution percentages, and time spent on after call work. Key research questions addressed areas of loss, contributing factors, and operational impacts. Using theme-based coding and chi-square goodness of fit analyses for the qualitative data coupled with descriptive analyses and frequency distributions for the quantitative data, results showed that loss occurred because of attitudinal resistance to change, unlearning, and lack of organizational standards. Average handle time and first call resolution metrics were negatively impacted. Contributing factors included culture, leadership support, and limited follow-through from lessons learned. Recommendations include establishing a formal disposal process to remove outdated knowledge from knowledge management tools and establishing incentives to encourage employees to contribute knowledge, which can lead to higher staff engagement of those tools and improved customer service. The social change significance is that addressing knowledge loss can promote fiscal sustainability and revenue generation, thereby preventing layoffs or organizational closures.
APA, Harvard, Vancouver, ISO, and other styles
41

Mullaseril, Paul Abraham 1959. "Capacitated rural postman problem with time windows and split delivery." Diss., The University of Arizona, 1997. http://hdl.handle.net/10150/282300.

Full text
Abstract:
The importance of effective and efficient distribution is evident from its associated costs. Transportation and shipping alone comprise roughly 15 percent of a product's sales in the U.S. Physical distribution is very energy and labor intensive, which have both become relatively more expensive in the last 10-15 years. Not surprisingly, there is a growing demand for automated planning systems that produce economical routes. Other than the cost savings, introduction of these systems enables companies to maintain a higher level of service for their customers, it makes them less dependent on human planners, it supplies better management information facilities and it makes distribution planning work faster and simpler.
APA, Harvard, Vancouver, ISO, and other styles
42

Dedge, Parks Dana M. "Defining Data Science and Data Scientist." Scholar Commons, 2017. http://scholarcommons.usf.edu/etd/7014.

Full text
Abstract:
The world’s data sets are growing exponentially every day due to the large number of devices generating data residue across the multitude of global data centers. What to do with the massive data stores, how to manage them and defining who are performing these tasks has not been adequately defined and agreed upon by academics and practitioners. Data science is a cross disciplinary, amalgam of skills, techniques and tools which allow business organizations to identify trends and build assumptions which lead to key decisions. It is in an evolutionary state as new technologies with capabilities are still being developed and deployed. The data science tasks and the data scientist skills needed in order to be successful with the analytics across the data stores are defined in this document. The research conducted across twenty-two academic articles, one book, eleven interviews and seventy-eight surveys are combined to articulate the convergence on the terms data science. In addition, the research identified that there are five key skill categories (themes) which have fifty-five competencies that are used globally by data scientists to successfully perform the art and science activities of data science. Unspecified portions of statistics, technology programming, development of models and calculations are combined to determine outcomes which lead global organizations to make strategic decisions every day. This research is intended to provide a constructive summary about the topics data science and data scientist in order to spark the dialogue for us to formally finalize the definitions and ultimately change the world by establishing set guidelines on how data science is performed and measured.
APA, Harvard, Vancouver, ISO, and other styles
43

Walsh, Kenneth Ronald 1966. "Generating insight for reengineering." Diss., The University of Arizona, 1996. http://hdl.handle.net/10150/290606.

Full text
Abstract:
As organizations face stiff competition and changing environments their structure and systems can become less effective to a point at which they require radical change. It is becoming clear that change is often inevitable. However, radical change is difficult. The popular term "reengineering" is used to describe a process by which organizations undergo the challenging radical process change that is critical to remaining competitive. Although the processes of reengineering appears to have more potential than other change methods that merely tweak a poor system or apply outdated principles, it can often fail. A critical step in the reengineering process is the design of the "to-be" model, however no research has been done on how this should be achieved. It appears that improving this step can have a dramatic impact on the success of the reengineering effort as a whole. This study developed a method for creating "to-be" models and tested it with an organization undergoing reengineering. The method used the latest computer support including both group support systems and animated simulation. The design of the "to-be" models was framed as a group problem solving process and therefore the focus theory of group productivity was used to guide the design of the process, including selection of computer support tools. Because so little is known about the process of creating "to-be" models, this study used an exploratory action science approach. What is known about organizational change is that it takes place in a politically charged arena where individuals have strong vested interests in its outcomes. This environment is difficult to create in the laboratory and therefore the research was conducted using a real organization undergoing reengineering. Results suggest that the combined use of collaborative technology and process animation gives organization members a better understanding of current processes and problems associated with them and helps to generate significant ideas for process improvement. Because groups viewing a process animation have sometimes been observed to focus on incremental improvement at the expense of radical change, special attention must be given to facilitation methods and idea generation techniques that are designed to elicit radical change ideas.
APA, Harvard, Vancouver, ISO, and other styles
44

Gilleo, Wayne M. "Examining continued use of software as a service in organizations." Thesis, Capella University, 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=3646094.

Full text
Abstract:

Information systems can be used to increase competitiveness by increasing the effectiveness and speed of decision-making. Software as a Service (SaaS) is a subset of cloud computing that provides information systems functionality through a web browser. Organizations that adopt SaaS can receive value over time if they continue to use the SaaS solution after implementation. This study analyzed the extent to which SaaS adoption and continued use factors affect the continued use of SaaS in organizations. The research can help organizations maximize the value of SaaS by identifying success factors for continued use. The study determined the extent to which the independent factors of Rapport, Responsiveness, Reliability, Features, Security, Flexibility, and Marketing Effort affected the dependent variable of the decision makers' intent to continue use of SaaS within their organization. Participants in the study included IT managers and other technologists familiar with the use of SaaS in their workplace. Invitations were sent via email to IT managers and technologists working at a large aerospace organization. There was also an open invitation posted on a social media site used by customers of a large SaaS provider. The invitations resulted in 101 valid responses. The Extended SaaS-Qual model developed in this study is an extension of an earlier model developed to analyze SaaS continuance. The original SaaS-Qual model was extended to understand the effect of marketing on SaaS continuance intention. Statistical analysis including loadings, internal consistency, convergent validity, and discriminate validity showed strong support for the Extended SaaS-Qual model. The Extended SaaS-Qual model will help with the establishment of standardized measures for the benefit of both SaaS providers and SaaS users.

APA, Harvard, Vancouver, ISO, and other styles
45

Zhang, Zhewei. "THREE ESSAYS ON DIGITAL EVOLUTION." Diss., Temple University Libraries, 2016. http://cdm16002.contentdm.oclc.org/cdm/ref/collection/p245801coll10/id/391804.

Full text
Abstract:
Business Administration/Management Information Systems
Ph.D.
Digital products are rapidly shaping our world into a ubiquitous computing world. Because of its unique characteristics, digital artifacts are generative and highly evolving through the recombination of existing elements as well as by the invention of new elements. In this thesis, I first propose an evolutionary view to examine how digital artifacts can evolve over time based on the discussion of limitations of existing approaches. Then, inspired by contemporary evolutionary thinking and phylogenetics in biology, I develop a set of analytic approaches to classify different evolutionary patterns based on the evolutionary rate and evolutionary direction of the digital artifact. In three studies, I investigate digital evolution in specific contexts: web Mashup development, open source software development and digital design practice. I conclude my thesis by highlighting the contributions of my work, and summarizing current limitations and opportunities for future research.
Temple University--Theses
APA, Harvard, Vancouver, ISO, and other styles
46

Bates, Anthony Shawn. "The Trilogy of Science: Filling the Knowledge Management Gap with Knowledge Science and Theory." ScholarWorks, 2017. https://scholarworks.waldenu.edu/dissertations/3977.

Full text
Abstract:
The international knowledge management field has different ways of investigating, developing, believing, and studying knowledge management. Knowledge management (KM) is distinguished deductively by know-how, and its intangible nature establishes different approaches to KM concepts, practices, and developments. Exploratory research and theoretical principles have formed functional intelligences from 1896 to 2013, leading to a knowledge management knowledge science (KMKS) concept that derived a grounded theory of knowledge activity (KAT). This study addressed the impact of knowledge production problems on KM practice. The purpose of this qualitative meta-analysis study was to fit KM practice within the framework of knowledge science (KS) study. Themed questions and research variables focused on field mechanisms, operative functions, principle theory, and relationships of KMKS. The action research used by American practitioners has not established a formal structure for KS. The meta-data-analysis examined 385 transdisciplinary peer-reviewed articles using social science, service science, and systems science databases, with a selection of interdisciplinary studies that had a practice-research-theory framework. Key attributes utilizing Boolean limiters, words, phrases and publication dates, along with triangulation, language analysis and coding through analytic software identified commonalities of the data under study. Findings reflect that KM has not become a theoretically saturated field. KS as the forensic science of KM creates a paradigm shift, causes social change that averts rapid shifts in management direction and uncertainty, and connects KM philosophy and science of knowledge. These findings have social change implications by informing the work of managers and academics to generate a methodical applied science.
APA, Harvard, Vancouver, ISO, and other styles
47

Schluderberg, Larry E. "Addressing the cybersecurity Malicious Insider threat." Thesis, Utica College, 2015. http://pqdtopen.proquest.com/#viewpdf?dispub=1571095.

Full text
Abstract:

Malicious Insider threats consist of employees, contractors, or business partners who either have current authorized access, or have had authorized access to an organization's critical information and have intentionally misused that access in a manner that compromised the organization. Although incidents initiated by malicious insiders are fewer in number than those initiated by external threats, insider incidents are more costly on average because the threat is already trusted by the organization and often has privileged access to the organization's most sensitive information. In spite of the damage they cause there are indications that the seriousness of insider incidents are underappreciated as threats by management. The purpose of this research was to investigate who constitutes MI threats, why and how they initiate attacks, the extent to which MI activity can be modeled or predicted, and to suggest some risk mitigation strategies. The results reveal that addressing the Malicious Insider threat is much more than just a technical issue. Dealing effectively with the threat involves managing the dynamic interaction between employees, their work environment and work associates, the systems with which they interact, and organizational policies and procedures. Techniques for detecting and mitigating the threat are available and can be effectively applied. Some of the procedural and technical methods include definition of, follow through, and consistent application of corporate, and dealing with adverse events indigenous to the business environment. Other methods include conduct of a comprehensive Malicious Insider risk assessment, selective monitoring of employees in response to behavioral precursors, minimizing unknown access paths, control of the organization's production software baseline, and effective use of peer reporting.

Keywords: Cybersecurity, Professor Paul Pantani, CERT, insider, threat, IDS, SIEMS. FIM, RBAC, ABAC, behavioral, peer, precursors, access, authentication, predictive, analytics, system, dynamics, demographics.

APA, Harvard, Vancouver, ISO, and other styles
48

Herman, Deirdre. "Asset Reuse of Images From a Repository." ScholarWorks, 2011. https://scholarworks.waldenu.edu/dissertations/1110.

Full text
Abstract:
According to Markus's theory of reuse, when digital repositories are deployed to collect and distribute organizational assets, they supposedly help ensure accountability, extend information exchange, and improve productivity. Such repositories require a large investment due to the continuing costs of hardware, software, user licenses, training, and technical support. The problem addressed in this study was the lack of evidence in the literature on whether users in fact reused enough digital assets in repositories to justify the investment. The objective of the study was to investigate the organizational value of repositories to better inform architectural, construction, software and other industries whether repositories are worth the investment. This study was designed to examine asset reuse of medical images at a health information publisher. The research question focused on the amount of asset reuse over time, which was determined from existing repository transaction logs generated over an 8-year period by all users. A longitudinal census data analysis of archival research was performed on the entire dataset of 85,250 transaction logs. The results showed that 42 users downloaded those assets, including 11,059 images, indicating that the repository was used by sufficient users at this publisher of about 80 employees. From those images, 1,443 medical images were reused for new product development, showing a minimal asset reuse rate of 13%. Assistants (42%), writers (20%), and librarians (16%) were the primary users of this repository. Collectively, these results demonstrated the value of repositories in improving organizational productivity---through reuse of existing digital assets such as medical images to avoid unnecessary duplication costs---for social change and economic transformation.
APA, Harvard, Vancouver, ISO, and other styles
49

Brown, Mary Erin. "Data-Driven Decision Making as a Tool to Improve Software Development Productivity." ScholarWorks, 2011. https://scholarworks.waldenu.edu/dissertations/1075.

Full text
Abstract:
The worldwide software project failure rate, based on a survey of information technology software manager's view of user satisfaction, product quality, and staff productivity, is estimated to be between 24% and 36% and software project success has not kept pace with the advances in hardware. The problem addressed by this study was the limited information about software managers' experiences with data-driven decision making (DDD) in agile software organizations as a tool to improve software development productivity. The purpose of this phenomenological study was to explore how agile software managers view DDD as a tool to improve software development productivity and to understand how agile software development organizations may use DDD now and in the future to improve software development productivity. Research questions asked about software managers', project managers', and agile coaches' lived experiences with DDD via a set of interview questions. The conceptual framework for the research was based on the 3 critical dimensions of software organization productivity improvement: people, process, and tools, which were defined by the Software Engineering Institute's Capability Maturity Model Integrated published in 2010. Organizations focus on processes to align the people, procedures and methods, and tools and equipment to improve productivity. Positive social change could result from a better understanding of DDD in an agile software development environment; this increased understanding of DDD could enable organizations to create more products, offer more jobs, and better compete in a global economy.
APA, Harvard, Vancouver, ISO, and other styles
50

Ecelbarger, Richard Paul 1944. "Leadership, process improvement, and information technology: The implementation of network-based applications in a financial services organization." Diss., The University of Arizona, 1998. http://hdl.handle.net/10150/282724.

Full text
Abstract:
The basis of this study is that traditional models of leadership do not adequately address the complexity of challenges presented to leaders engaged in IT-facilitated process improvements. The result is that most IT (information technology) implementations either fail outright or do not meet expectations. This study was conducted therefore, to close the knowledge gap that exists between the leadership decision to seek IT-facilitated improvement and the leadership approach needed to successfully accomplish it. Case study and qualitative research methodologies were employed to reveal assumptions and beliefs of leadership that influenced the development of network-based applications in a financial service organization--how improvement projects were initiated, implemented, and accepted by the staff for whom they were intended. The analysis of data led to the description of seven leadership themes--Institutional Climate, Involvement, Communication, Learning, Development Strategy, Resources, and Impact. Of particular interest are the assumptions and beliefs of the Impact theme, which provides evidence that IT-facilitated process improvements were successfully implemented in the financial service organization studied. The assumptions and beliefs of leadership are then used to determine how the described framework of leadership is aligned with the Postindustrial Model of Leadership suggested by Lee Bolman and Terrence Deal (1991)--a model that they contend extends beyond the limitations of traditional models. The framework revealed by the seven leadership themes was compared to evaluation criteria developed for Bolman and Deal's leadership model. Using the organizational frame definitions developed by Bolman and Deal in their book Reframing Organizations: Artistry, Choice, and Leadership (1991), the following ranking was observed: Structural, Political, Human Resource, and Symbolic. This analysis provides evidence that the framework of leadership in the financial services organization is aligned with the Postindustrial Model of Leadership suggested by Bolman and Deal. In summary, the study reveals that successful IT-implementations were achieved in the financial service organization with Bolman and Deal's Postindustrial Model of Leadership in place. Additional research is recommended, however, to refine the assumptions and beliefs of leadership such that the conclusions reached in this study can be confirmed and generalized to expanded populations and circumstances.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography