To see the other types of publications on this topic, follow the link: Big quality.

Dissertations / Theses on the topic 'Big quality'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Big quality.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Blahová, Leontýna. "Big Data Governance." Master's thesis, Vysoká škola ekonomická v Praze, 2016. http://www.nusl.cz/ntk/nusl-203994.

Full text
Abstract:
This master thesis is about Big Data Governance and about software, which is used for this purposes. Because Big Data are huge opportunity and also risk, I wanted to map products which can be easily use for Data Quality and Big Data Governance in one platform. This thesis is not only on theoretical knowledge level, but also evaluates five key products (from my point of view). I defined requirements for every kind of domain and then I set up the weights and points. The main objective is to evaluate software capabilities and compere them.
APA, Harvard, Vancouver, ISO, and other styles
2

Serra-Diaz, Josep M., Brian J. Enquist, Brian Maitner, Cory Merow, and Jens-C. Svenning. "Big data of tree species distributions: how big and how good?" SPRINGER HEIDELBERG, 2018. http://hdl.handle.net/10150/626611.

Full text
Abstract:
Background: Trees play crucial roles in the biosphere and societies worldwide, with a total of 60,065 tree species currently identified. Increasingly, a large amount of data on tree species occurrences is being generated worldwide: from inventories to pressed plants. While many of these data are currently available in big databases, several challenges hamper their use, notably geolocation problems and taxonomic uncertainty. Further, we lack a complete picture of the data coverage and quality assessment for open/public databases of tree occurrences. Methods: We combined data from five major aggregators of occurrence data (e.g. Global Biodiversity Information Facility, Botanical Information and Ecological Network v.3, DRYFLOR, RAINBIO and Atlas of Living Australia) by creating a workflow to integrate, assess and control data quality of tree species occurrences for species distribution modeling. We further assessed the coverage - the extent of geographical data - of five economically important tree families (Arecaceae, Dipterocarpaceae, Fagaceae, Myrtaceae, Pinaceae). Results: Globally, we identified 49,206 tree species (84.69% of total tree species pool) with occurrence records. The total number of occurrence records was 36.69 M, among which 6.40 M could be considered high quality records for species distribution modeling. The results show that Europe, North America and Australia have a considerable spatial coverage of tree occurrence data. Conversely, key biodiverse regions such as South-East Asia and central Africa and parts of the Amazon are still characterized by geographical open-public data gaps. Such gaps are also found even for economically important families of trees, although their overall ranges are covered. Only 15,140 species (26.05%) had at least 20 records of high quality. Conclusions: Our geographical coverage analysis shows that a wealth of easily accessible data exist on tree species occurrences worldwide, but regional gaps and coordinate errors are abundant. Thus, assessment of tree distributions will need accurate occurrence quality control protocols and key collaborations and data aggregation, especially from national forest inventory programs, to improve the current publicly available data.
APA, Harvard, Vancouver, ISO, and other styles
3

Palmqvist, Simon. "Validating the Quality of a Big Data Java Corpus." Thesis, Linnéuniversitetet, Institutionen för datavetenskap och medieteknik (DM), 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-75410.

Full text
Abstract:
Recent research within the field of Software Engineering have used GitHub, the largest hub for open source projects with almost 20 million users and 57 million repositories, to mine large amounts of source code to get more trustworthy results when developing machine and deep learning models. Mining GitHub comes with many challenges since the dataset is large and the data does not only contain quality software projects. In this project, we try to mine projects from GitHub based on earlier research by others and try to validate the quality by comparing the projects with a small subset of quality projects with the help of software complexity metrics.
APA, Harvard, Vancouver, ISO, and other styles
4

Yu, Dong Michael. "The effect of big four office size on audit quality." Diss., Columbia, Mo. : University of Missouri-Columbia, 2007. http://hdl.handle.net/10355/4827.

Full text
Abstract:
Thesis (Ph. D.)--University of Missouri-Columbia, 2007.
The entire dissertation/thesis text is included in the research.pdf file; the official abstract appears in the short.pdf file (which also appears in the research.pdf); a non-technical general description, or public abstract, appears in the public.pdf file. Title from title screen of research.pdf file (viewed on October 15, 2007) Vita. Includes bibliographical references.
APA, Harvard, Vancouver, ISO, and other styles
5

Tian, Chao. "Towards effective analysis of big graphs : from scalability to quality." Thesis, University of Edinburgh, 2017. http://hdl.handle.net/1842/29578.

Full text
Abstract:
This thesis investigates the central issues underlying graph analysis, namely, scalability and quality. We first study the incremental problems for graph queries, which aim to compute the changes to the old query answer, in response to the updates to the input graph. The incremental problem is called bounded if its cost is decided by the sizes of the query and the changes only. No matter how desirable, however, our first results are negative: for common graph queries such as graph traversal, connectivity, keyword search and pattern matching, their incremental problems are unbounded. In light of the negative results, we propose two new characterizations for the effectiveness of incremental computation, and show that the incremental computations above can still be effectively conducted, by either reducing the computations on big graphs to small data, or incrementalizing batch algorithms by minimizing unnecessary recomputation. We next study the problems with regards to improving the quality of the graphs. To uniquely identify entities represented by vertices in a graph, we propose a class of keys that are recursively defined in terms of graph patterns, and are interpreted with subgraph isomorphism. As an application, we study the entity matching problem, which is to find all pairs of entities in a graph that are identified by a given set of keys. Although the problem is proved to be intractable, and cannot be parallelized in logarithmic rounds, we provide two parallel scalable algorithms for it. In addition, to catch numeric inconsistencies in real-life graphs, we extend graph functional dependencies with linear arithmetic expressions and comparison predicates, referred to as NGDs. Indeed, NGDs strike a balance between expressivity and complexity, since if we allow non-linear arithmetic expressions, even of degree at most 2, the satisfiability and implication problems become undecidable. A localizable incremental algorithm is developed to detect errors using NGDs, where the cost is determined by small neighbors of nodes in the updates instead of the entire graph. Finally, a rule-based method to clean graphs is proposed. We extend graph entity dependencies (GEDs) as data quality rules. Given a graph, a set of GEDs and a block of ground truth, we fix violations of GEDs in the graph by combining data repairing and object identification. The method finds certain fixes to errors detected by GEDs, i.e., as long as the GEDs and the ground truth are correct, the fixes are assured correct as their logical consequences. Several fundamental results underlying the method are established, and an algorithm is developed to implement the method. We also parallelize the method and guarantee to reduce its running time with the increase of processors.
APA, Harvard, Vancouver, ISO, and other styles
6

Rizk, Raya. "Big Data Validation." Thesis, Uppsala universitet, Informationssystem, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-353850.

Full text
Abstract:
With the explosion in usage of big data, stakes are high for companies to develop workflows that translate the data into business value. Those data transformations are continuously updated and refined in order to meet the evolving business needs, and it is imperative to ensure that a new version of a workflow still produces the correct output. This study focuses on the validation of big data in a real-world scenario, and implements a validation tool that compares two databases that hold the results produced by different versions of a workflow in order to detect and prevent potential unwanted alterations, with row-based and column-based statistics being used to validate the two versions. The tool was shown to provide accurate results in test scenarios, providing leverage to companies that need to validate the outputs of the workflows. In addition, by automating this process, the risk of human error is eliminated, and it has the added benefit of improved speed compared to the more labour-intensive manual alternative. All this allows for a more agile way of performing updates on the data transformation workflows by improving on the turnaround time of the validation process.
APA, Harvard, Vancouver, ISO, and other styles
7

TANNEEDI, NAREN NAGA PAVAN PRITHVI. "Customer Churn Prediction Using Big Data Analytics." Thesis, Blekinge Tekniska Högskola, Institutionen för kommunikationssystem, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-13518.

Full text
Abstract:
Customer churn is always a grievous issue for the Telecom industry as customers do not hesitate to leave if they don’t find what they are looking for. They certainly want competitive pricing, value for money and above all, high quality service. Customer churning is directly related to customer satisfaction. It’s a known fact that the cost of customer acquisition is far greater than cost of customer retention, that makes retention a crucial business prototype. There is no standard model which addresses the churning issues of global telecom service providers accurately. BigData analytics with Machine Learning were found to be an efficient way for identifying churn. This thesis aims to predict customer churn using Big Data analytics, namely a J48 decision tree on a Java based benchmark tool, WEKA. Three different datasets from various sources were considered; first includes Telecom operator’s six month aggregate active and churned users’ data usage volumes, second includes globally surveyed data and third dataset comprises of individual weekly data usage analysis of 22 android customers along with their average quality, annoyance and churn scores by accompanying theses. Statistical analyses and J48 Decision trees were drawn for three different datasets. From the statistics of normalized volumes, autocorrelations were small owing to reliable confidence intervals, but confidence intervals were overlapping and close by, therefore no much significance could be noticed, henceforth no strong trends could be observed. From decision tree analytics, decision trees with 52%, 70% and 95% accuracies were achieved for three different data sources respectively.      Data preprocessing, data normalization and feature selection have shown to be prominently influential. Monthly data volumes have not shown much decision power. Average Quality, Churn Risk and to some extent, Annoyance scores may point out a probable churner. Weekly data volumes with customer’s recent history and necessary attributes like age, gender, tenure, bill, contract, data plan, etc., are pivotal for churn prediction.
APA, Harvard, Vancouver, ISO, and other styles
8

Sonsa-ardjit, Pitchaya, and Ramon Vejaratpimol. "Clients’ Perspectives Toward Audit Service Quality of the Big 4 inThailand." Thesis, Karlstad University, Faculty of Economic Sciences, Communication and IT, 2010. http://urn.kb.se/resolve?urn=urn:nbn:se:kau:diva-6198.

Full text
Abstract:

Purpose

The purpose of this thesis is, firstly, to investigate clients’ perspective toward the Big 4’s financial audit service quality. Secondly, the gaps between clients’ perceptions and expectations of audit service quality provided by the Big 4 audit firms will be studied. Finally, factors influencing clients’ expectations of audit service quality will be categorised.

Method

A combination of qualitative and quantitative approach is used in the form of a web-based self-completion questionnaire. A qualitative approach is used in one section of the questionnaire which is an open-ended question asking about the

clients’ perception toward audit service quality. A quantitative approach is used in the rest of the 2 sections of the questionnaire; firstly, asking the respondents to score the level of perception and expectation of audit service quality; secondly, asking for types of clients’ industries. The respondents are 25 clients who have direct experience with the Big 4 audit firms located in Thailand.

Finding 

Clients strongly expect assurance, reliability, and responsiveness while strongly perceive assurance and reliability of the Big4’s audit service quality. However, it is obvious that clients’ perception of all 5 dimensions is less than those of expectation; assurance, reliability and responsiveness are significantly different at .05 level. Moreover, eight factors from given expectation score are re-categorised in order from the most important issue to the least important as follows; Factor 1: Trust & Confidence, Factor 2: Responsiveness & Accuracy, Factor 3: Knowledge and skills in clients’ industry, Caring and Independence, Factor 4: Understanding of Clients, Factor 5: Timing/Scheduling & Right Service, Factor 6: Physical Facilities, Factor 7: Professional appearance & Professional Procedures, and Factor 8: Information & Communication Channels and Materials.

Conclusion 

In conclusion, the factors that are not satisfied by the clients; assurance, reliability, responsiveness, should be taken account of by the Big 4. Not only the Big 4 operating in Thailand have to be aware of their service quality, the other audit firms both international brands and local brands should also be aware of their service quality in order to satisfy their clients and to avoid damages of the firms and markets from audit failure. Both the audit firms and the clients together can help in audit quality improvement.

Recommendation 

To improve audit service quality, it is not only the Big4 audit firms’ responsibility but also the good cooperation from the clients could be the crucial support, and the ongoing policies are needed because it takes some time to see the consequences. When the quality level of audit service becomes a win-win situation, both audit firms and clients receive mutual benefits. Moreover, the Big 4 are the big actors in the audit industry in Thailand with promptly financial and human resource, they should support non-Big 4 to improve audit service quality. Because it means the overall image of audit service in Thailand would be improve somehow.

APA, Harvard, Vancouver, ISO, and other styles
9

Santos, Lúcio Fernandes Dutra. "Similaridade em big data." Universidade de São Paulo, 2017. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-07022018-104929/.

Full text
Abstract:
Os volumes de dados armazenados em grandes bases de dados aumentam em ritmo sempre crescente, pressionando o desempenho e a flexibilidade dos Sistemas de Gerenciamento de Bases de Dados (SGBDs). Os problemas de se tratar dados em grandes quantidades, escopo, complexidade e distribuição vêm sendo tratados também sob o tema de big data. O aumento da complexidade cria a necessidade de novas formas de busca - representar apenas números e pequenas cadeias de caracteres já não é mais suficiente. Buscas por similaridade vêm se mostrando a maneira por excelência de comparar dados complexos, mas até recentemente elas não estavam disponíveis nos SGBDs. Agora, com o início de sua disponibilidade, está se tornando claro que apenas os operadores de busca por similaridade fundamentais não são suficientes para lidar com grandes volumes de dados. Um dos motivos disso é que similaridade\' é, usualmente, definida considerando seu significado quando apenas poucos estão envolvidos. Atualmente, o principal foco da literatura em big data é aumentar a eficiência na recuperação dos dados usando paralelismo, existindo poucos estudos sobre a eficácia das respostas obtidas. Esta tese visa propor e desenvolver variações dos operadores de busca por similaridade para torná-los mais adequados para processar big data, apresentando visões mais abrangentes da base de dados, aumentando a eficácia das respostas, porém sem causar impactos consideráveis na eficiência dos algoritmos de busca e viabilizando sua execução escalável sobre grandes volumes de dados. Para alcançar esse objetivo, este trabalho apresenta quatro frentes de contribuições: A primeira consistiu em um modelo de diversificação de resultados que pode ser aplicado usando qualquer critério de comparação e operador de busca por similaridade. A segunda focou em definir técnicas de amostragem e de agrupamento de dados com o modelo de diversificação proposto, acelerando o processo de análise dos conjuntos de resultados. A terceira contribuição desenvolveu métodos de avaliação da qualidade dos conjuntos de resultados diversificados. Por fim, a última frente de contribuição apresentou uma abordagem para integrar os conceitos de mineração visual de dados e buscas por similaridade com diversidade em sistemas de recuperação por conteúdo, aumentando o entendimento de como a propriedade de diversidade pode ser aplicada.
The data being collected and generated nowadays increase not only in volume, but also in complexity, requiring new query operators. Health care centers collecting image exams and remote sensing from satellites and from earth-based stations are examples of application domains where more powerful and flexible operators are required. Storing, retrieving and analyzing data that are huge in volume, structure, complexity and distribution are now being referred to as big data. Representing and querying big data using only the traditional scalar data types are not enough anymore. Similarity queries are the most pursued resources to retrieve complex data, but until recently, they were not available in the Database Management Systems. Now that they are starting to become available, its first uses to develop real systems make it clear that the basic similarity query operators are not enough to meet the requirements of the target applications. The main reason is that similarity is a concept formulated considering only small amounts of data elements. Nowadays, researchers are targeting handling big data mainly using parallel architectures, and only a few studies exist targeting the efficacy of the query answers. This Ph.D. work aims at developing variations for the basic similarity operators to propose better suited similarity operators to handle big data, presenting a holistic vision about the database, increasing the effectiveness of the provided answers, but without causing impact on the efficiency on the searching algorithms. To achieve this goal, four mainly contributions are presented: The first one was a result diversification model that can be applied in any comparison criteria and similarity search operator. The second one focused on defining sampling and grouping techniques with the proposed diversification model aiming at speeding up the analysis task of the result sets. The third contribution concentrated on evaluation methods for measuring the quality of diversified result sets. Finally, the last one defines an approach to integrate the concepts of visual data mining and similarity with diversity searches in content-based retrieval systems, allowing a better understanding of how the diversity property is applied in the query process.
APA, Harvard, Vancouver, ISO, and other styles
10

Grillo, Aderibigbe. "Developing a data quality scorecard that measures data quality in a data warehouse." Thesis, Brunel University, 2018. http://bura.brunel.ac.uk/handle/2438/17137.

Full text
Abstract:
The main purpose of this thesis is to develop a data quality scorecard (DQS) that aligns the data quality needs of the Data warehouse stakeholder group with selected data quality dimensions. To comprehend the research domain, a general and systematic literature review (SLR) was carried out, after which the research scope was established. Using Design Science Research (DSR) as the methodology to structure the research, three iterations were carried out to achieve the research aim highlighted in this thesis. In the first iteration, as DSR was used as a paradigm, the artefact was build from the results of the general and systematic literature review conduct. A data quality scorecard (DQS) was conceptualised. The result of the SLR and the recommendations for designing an effective scorecard provided the input for the development of the DQS. Using a System Usability Scale (SUS), to validate the usability of the DQS, the results of the first iteration suggest that the DW stakeholders found the DQS useful. The second iteration was conducted to further evaluate the DQS through a run through in the FMCG domain and then conducting a semi-structured interview. The thematic analysis of the semi-structured interviews demonstrated that the stakeholder's participants' found the DQS to be transparent; an additional reporting tool; Integrates; easy to use; consistent; and increases confidence in the data. However, the timeliness data dimension was found to be redundant, necessitating a modification to the DQS. The third iteration was conducted with similar steps as the second iteration but with the modified DQS in the oil and gas domain. The results from the third iteration suggest that DQS is a useful tool that is easy to use on a daily basis. The research contributes to theory by demonstrating a novel approach to DQS design This was achieved by ensuring the design of the DQS aligns with the data quality concern areas of the DW stakeholders and the data quality dimensions. Further, this research lay a good foundation for the future by establishing a DQS model that can be used as a base for further development.
APA, Harvard, Vancouver, ISO, and other styles
11

Wong, Ho-chuen. "Analysis and prediction of beach water quality in Hong Kong with special reference to Big Wave Bay Beach /." Click to view the E-thesis via HKUTO, 2009. http://sunzi.lib.hku.hk/hkuto/record/B43278619.

Full text
APA, Harvard, Vancouver, ISO, and other styles
12

Isacsson, Johannes, and William Ryan. "Total Kvalitetsstyrning inom Non-Big X Revisionsbyråer : Praktiskt arbete med ISQC 1." Thesis, Linköpings universitet, Företagsekonomi, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-139363.

Full text
Abstract:
Background and problem: The Enron scandal is one of many audit scandals during the 21st century, which in combination with the global harmonization has led to increased demands on the audit sector and its quality. Due to this the ISQC 1 was designed, which purpose is to control and guide an internal quality control that will achieve and maintain a high level of quality. Research on audit quality has risen with the designing of ISQC 1, where focus mostly have been on how audit firms have managed the implementation of and the work with the standard. However, the research focuses on big international audit firms, which results in audit firms with a lower number of employees being left in the dark. The established field of Total Quality Management can be used to study how audit firms with a lower number of employees work with ISQC 1. Purpose: The purpose of this study is to create an understanding of how non-Big X audit firms use ISQC 1 to assure the quality of their services. Method: This study uses a qualitative method and an inductive approach, with elements of a deductive approach. Empirical primary data has been gathered by longer semi-structured interviews with eight auditors who all works on different non-Big X audit firms. A content analysis has also been executed, which compares ISQC 1 with the cornerstones of TQM. Conclusion: This study indicates that non-Big X audit firms mostly uses meetings, teachings and audit manuals to create a quality culture and work as a basis for the audit process. Furthermore, the study implies that ISQC 1 is experienced as resource-demanding, copious and more adjusted to big international audit firms. To face these issues, this study indicates that a cooperation with audit networks can help. When it comes to audit firms assuring the quality of their services, this study points out that preventing errors is critical for the non-Big X audit firms. Ultimately this study supports that there is an expectation gap, and that customers demand different things than what ISQC 1 advocate.
Bakgrund och problem: Enronskandalen är en i raden av olika revisionsskandaler under 2000-talet, som i kombination med den globala harmoniseringen har lett till ökade krav på revisionsbranschen och dess kvalitet. Till följd av detta utformades standarden ISQC 1 vars syfte är att styra och vägleda en intern kvalitetskontroll för att på så sätt uppnå en jämn och hög nivå på kvaliteten. I och med utformningen av ISQC 1 har också forskning uppkommit där fokus har varit på hur revisionsbyråer klarat av att implementera och arbeta med standarden. Denna forskning fokuserar dock främst på stora internationella revisionsbyråer, vilket gör att revisionsbyråer med färre anställda kommer i skymundan. För att undersöka hur revisionsbyråer med färre anställda arbetar med ISQC 1 kan då det etablerade fältet total kvalitetsstyrning användas.   Syfte: Syftet med denna studie är att skapa förståelse för hur non-Big X revisionsbyråer arbetar med ISQC 1 för att kvalitetssäkra sina tjänster i praktiken.  Metod: Denna studie använder sig av en kvalitativ metod och en induktiv ansats med inslag av deduktion. Empirisk primärdata har samlats in genom längre semi-strukturerade intervjuer med åtta revisorer som arbetar på olika non-Big X byråer. Vidare har även en innehållsanalys genomförts som jämför ISQC 1 med hörnstensmodellen. Slutsats: Studien indikerar på att non-Big X revisionsbyråer främst använder sig av möten, utbildningar och revisionshandböcker för att både skapa en kvalitetskultur och ligga till grund för revisionsprocessen. Studien ger vidare antydningar om att ISQC 1 upplevs som resurskrävande, överflödig och mer anpassad efter stora internationella byråer. För att bemöta dessa problem ges indikationer på att ett samarbete med revisionsnätverk kan underlätta. I arbetet med att kvalitetssäkra tjänster indikerar vidare denna studie på att förebyggande av fel är kritiskt för non-Big X byråer. Slutligen stärker studien att ett förväntningsgap finns, då kunderna efterfrågar annat än vad ISQC 1 förespråkar.
APA, Harvard, Vancouver, ISO, and other styles
13

Decker, Timothy Joseph. "An assessment of water quality on Little and Big Duck Creeks near Elwood, Indiana." Virtual Press, 1987. http://liblink.bsu.edu/uhtbin/catkey/539626.

Full text
Abstract:
A water quality study was conducted on Big Duck Creek and Little Duck Creek near Elwood, Indiana during the summer, autumn and winter of 1978 - 1979 and compared to measurements made in 1938.Samples were analyzed for dissolved oxygen, biochemical oxygen demand, suspended solids, ammonia, total coliform bacteria, hydrogen ion concentration and temperature. Significant differences in concentrations were observed for each environmental parameter measured in 1979 when evaluated on a basis of sampling location. With the exception of dissolved oxygen concentrations, the effect of stream volume on the concentration of environmental parameters was in most instances small.When the up and downstream stations were compared, a definite decrease in water quality was noted. The dissolved oxygen concentrations decreased as the water flowed through the city. Due to increased organic loading, the biochemical oxygen demand increased in the downstream area.Elwood increased the suspended solids level of Big Duck Creek. Agricultural and urban runoff together with untreated sewage discharges significantly increased the suspended solids level within the inner city region. Dilution by treated sewage as well as cleaner water from Little Duck Creek help reduce the suspended solids level in the downstream location. Relatively high ammonia concentrations were observed in the upstream portions of the creeks. This was apparently related to farm practices. The higher readings of ammonia were noted after animal manure was spread on the fields along the stream. The downstream levels were also high. This was probably due to ammonia in the effluents of the sewage treatment plant. In contrast, only small changes in the hydrogen ion concentration was observed throughout the creek.The number of total coliform bacteria increased in the center of town because of a sewer bypass into the creek. Below Elwood the concentration of bacteria decreased due to dilution with disinfected effluents from the sewage treatment plant.Except for a noticeable increase in biochemical oxygen demand and ammonia, the Elwood sewage treatment plant effluent together with flow from Little Duck Creek appears to improve Big Duck Creek's condition as it leaves the city to join White River.Significant differences were observed between measurements made in 1938 and 1978 - 79. Results of the study showed a significant improvement in water quality of Big Duck Creek since 1938. This was probably due to the construction of Elwood's wastewater collection system and sewage treatment plant in the 1940s.However, there was still degradation of water quality within the city due to untreated wastewater discharges. Consequently, emphasis should be focused on the inner city problem since Elwood residents would be in the proximity to this area.
Department of Natural Resources
APA, Harvard, Vancouver, ISO, and other styles
14

Landelius, Cecilia. "Data governance in big data : How to improve data quality in a decentralized organization." Thesis, KTH, Industriell ekonomi och organisation (Inst.), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-301258.

Full text
Abstract:
The use of internet has increased the amount of data available and gathered. Companies are investing in big data analytics to gain insights from this data. However, the value of the analysis and decisions made based on it, is dependent on the quality ofthe underlying data. For this reason, data quality has become a prevalent issue for organizations. Additionally, failures in data quality management are often due to organizational aspects. Due to the growing popularity of decentralized organizational structures, there is a need to understand how a decentralized organization can improve data quality. This thesis conducts a qualitative single case study of an organization currently shifting towards becoming data driven and struggling with maintaining data quality within the logistics industry. The purpose of the thesis is to answer the questions: • RQ1: What is data quality in the context of logistics data? • RQ2: What are the obstacles for improving data quality in a decentralized organization? • RQ3: How can these obstacles be overcome? Several data quality dimensions were identified and categorized as critical issues,issues and non-issues. From the gathered data the dimensions completeness, accuracy and consistency were found to be critical issues of data quality. The three most prevalent obstacles for improving data quality were data ownership, data standardization and understanding the importance of data quality. To overcome these obstacles the most important measures are creating data ownership structures, implementing data quality practices and changing the mindset of the employees to a data driven mindset. The generalizability of a single case study is low. However, there are insights and trends which can be derived from the results of this thesis and used for further studies and companies undergoing similar transformations.
Den ökade användningen av internet har ökat mängden data som finns tillgänglig och mängden data som samlas in. Företag påbörjar därför initiativ för att analysera dessa stora mängder data för att få ökad förståelse. Dock är värdet av analysen samt besluten som baseras på analysen beroende av kvaliteten av den underliggande data. Av denna anledning har datakvalitet blivit en viktig fråga för företag. Misslyckanden i datakvalitetshantering är ofta på grund av organisatoriska aspekter. Eftersom decentraliserade organisationsformer blir alltmer populära, finns det ett behov av att förstå hur en decentraliserad organisation kan arbeta med frågor som datakvalitet och dess förbättring. Denna uppsats är en kvalitativ studie av ett företag inom logistikbranschen som i nuläget genomgår ett skifte till att bli datadrivna och som har problem med att underhålla sin datakvalitet. Syftet med denna uppsats är att besvara frågorna: • RQ1: Vad är datakvalitet i sammanhanget logistikdata? • RQ2: Vilka är hindren för att förbättra datakvalitet i en decentraliserad organisation? • RQ3: Hur kan dessa hinder överkommas? Flera datakvalitetsdimensioner identifierades och kategoriserades som kritiska problem, problem och icke-problem. Från den insamlade informationen fanns att dimensionerna, kompletthet, exakthet och konsekvens var kritiska datakvalitetsproblem för företaget. De tre mest förekommande hindren för att förbättra datakvalité var dataägandeskap, standardisering av data samt att förstå vikten av datakvalitet. För att överkomma dessa hinder är de viktigaste åtgärderna att skapa strukturer för dataägandeskap, att implementera praxis för hantering av datakvalitet samt att ändra attityden hos de anställda gentemot datakvalitet till en datadriven attityd. Generaliseringsbarheten av en enfallsstudie är låg. Dock medför denna studie flera viktiga insikter och trender vilka kan användas för framtida studier och för företag som genomgår liknande transformationer.
APA, Harvard, Vancouver, ISO, and other styles
15

Brabbins, Lucinda J. "Accepting the 'Big C' : exploring the acceptance-quality of life relationship in a cancer population." Thesis, University of Lincoln, 2016. http://eprints.lincoln.ac.uk/29724/.

Full text
Abstract:
There is a high prevalence of distress amongst cancer patients, with up to 40% reporting clinically-significant levels of distress, such as depression, anxiety, and death anxiety. Acceptance and Commitment Therapy (ACT) has a growing evidence base in health populations, such as in diabetes and chronic pain, counteracting the avoidant behaviours which are negatively implicated in outcomes. ACT claims that experiential acceptance is key to shifting the avoidant responses people exhibit towards their inner experiences and psychological pain. Thus far, there is limited research exploring ACT processes within a cancer population. Traditional existentially-informed theory has claimed that accepting death anxiety would be psychologically paralysing and that mortality must be defended against, yet evidence for acceptance suggests the opposite. This study therefore aimed to explore the relationship between acceptance and outcomes of quality of life and distress in a cancer population, in order to explore the implications for an accepting response style, and also whether avoidant or approach response styles were implicated in better or worse psychological outcomes. Using a longitudinal design, six standardised questionnaires, well-utilised in cancer populations, were issued to 72 adults with experience of cancer, measuring clinical variables, cancer appraisals, response styles, and both quality of life and distress outcomes. The addition of a follow-up questionnaire, completed by 31 participants after three months, allowed for predictive and cross-lag analyses to be carried out. Results showed acceptance to be an independent explanatory and predictive response style variable for both quality of life and distress outcomes, in the direction of psychological health. ‘Avoidant’ response styles were negatively implicated in outcomes, and many ‘approach’ response styles other than acceptance did not reach correlational significance with outcome variables. Acceptance and avoidant response styles remained stable over time, whilst outcomes continued to change, and acceptance was thought to be driving future outcomes in a one-way direction. 5 The findings support the notion that experiential acceptance is implicated in beneficial psychological outcomes for cancer populations, rather than being detrimental to functioning, as suggested by existential theorists. Acceptance demonstrated significantly more influence over outcomes than either disease characteristics or threatening illness appraisals, and may influence future functioning. Furthermore, avoidant responses to cancer-related experiences were negatively implicated in psychological distress and also quality of life, which continued to deteriorate over time despite avoidant response styles remaining stable. Acceptance may be a helpful and influential response style, given that it predicts and potentially influences future functioning, and may provide a reasonable treatment target for psychological intervention with cancer patients, over and above the direct targeting of cancer appraisals. Objective. 40% of cancer patients may experience clinically-significant levels of distress, yet evidence for appropriate psycho-oncological interventions remains lacking. This study aimed to explore acceptance, as defined by acceptance and commitment therapy (ACT), in cancer patients. Primary aims investigated whether acceptance was related to and predictive of better quality of life and distress outcomes, and whether acceptance interacted with cancer appraisals to influence outcomes. Design. Longitudinal, quantitative design with a follow-up after three months. Participants completed a battery of questionnaires at times one and two. Measures. 72 participants completed standardised questionnaire batteries comprising: Brief Illness Perception Questionnaire; Brief COPE; Acceptance and Action Questionnaire II; Hospital Anxiety and Depression Scale; Functional Assessment of Cancer Therapy – General (FACT-G); and Death Anxiety Scale (DAS). 31 participants repeated the battery after three months. Results. Acceptance was an independent explanatory and predictive variable for quality of life and distress scores, in the direction of psychological health, and predicted functioning over time. Acceptance had greater explanatory power for outcomes than either cancer appraisals or avoidant response styles, including denial and self-distraction. Avoidant response styles showed significant associations with outcomes, but in the directions of greater distress and poorer quality of life. Conclusions. The findings support the role of an accepting response style in improved psychological outcomes. Avoidant responses were consistently related to poorer outcomes in a population with distress and death anxiety. Acceptance predicts functioning, and is supported as a helpful response style for cancer patients.
APA, Harvard, Vancouver, ISO, and other styles
16

Washha, Mahdi. "Information quality in online social media and big data collection : an example of Twitter spam detection." Thesis, Toulouse 3, 2018. http://www.theses.fr/2018TOU30080/document.

Full text
Abstract:
La popularité des médias sociaux en ligne (Online Social Media - OSM) est fortement liée à la qualité du contenu généré par l'utilisateur (User Generated Content - UGC) et la protection de la vie privée des utilisateurs. En se basant sur la définition de la qualité de l'information, comme son aptitude à être exploitée, la facilité d'utilisation des OSM soulève de nombreux problèmes en termes de la qualité de l'information ce qui impacte les performances des applications exploitant ces OSM. Ces problèmes sont causés par des individus mal intentionnés (nommés spammeurs) qui utilisent les OSM pour disséminer des fausses informations et/ou des informations indésirables telles que les contenus commerciaux illégaux. La propagation et la diffusion de telle information, dit spam, entraînent d'énormes problèmes affectant la qualité de services proposés par les OSM. La majorité des OSM (comme Facebook, Twitter, etc.) sont quotidiennement attaquées par un énorme nombre d'utilisateurs mal intentionnés. Cependant, les techniques de filtrage adoptées par les OSM se sont avérées inefficaces dans le traitement de ce type d'information bruitée, nécessitant plusieurs semaines ou voir plusieurs mois pour filtrer l'information spam. En effet, plusieurs défis doivent être surmontées pour réaliser une méthode de filtrage de l'information bruitée . Les défis majeurs sous-jacents à cette problématique peuvent être résumés par : (i) données de masse ; (ii) vie privée et sécurité ; (iii) hétérogénéité des structures dans les réseaux sociaux ; (iv) diversité des formats du UGC ; (v) subjectivité et objectivité. Notre travail s'inscrit dans le cadre de l'amélioration de la qualité des contenus en termes de messages partagés (contenu spam) et de profils des utilisateurs (spammeurs) sur les OSM en abordant en détail les défis susmentionnés. Comme le spam social est le problème le plus récurant qui apparaît sur les OSM, nous proposons deux approches génériques pour détecter et filtrer le contenu spam : i) La première approche consiste à détecter le contenu spam (par exemple, les tweets spam) dans un flux en temps réel. ii) La seconde approche est dédiée au traitement d'un grand volume des données relatives aux profils utilisateurs des spammeurs (par exemple, les comptes Twitter)
The popularity of OSM is mainly conditioned by the integrity and the quality of UGC as well as the protection of users' privacy. Based on the definition of information quality as fitness for use, the high usability and accessibility of OSM have exposed many information quality (IQ) problems which consequently decrease the performance of OSM dependent applications. Such problems are caused by ill-intentioned individuals who misuse OSM services to spread different kinds of noisy information, including fake information, illegal commercial content, drug sales, mal- ware downloads, and phishing links. The propagation and spreading of noisy information cause enormous drawbacks related to resources consumptions, decreasing quality of service of OSM-based applications, and spending human efforts. The majority of popular social networks (e.g., Facebook, Twitter, etc) over the Web 2.0 is daily attacked by an enormous number of ill-intentioned users. However, those popular social networks are ineffective in handling the noisy information, requiring several weeks or months to detect them. Moreover, different challenges stand in front of building a complete OSM-based noisy information filtering methods that can overcome the shortcomings of OSM information filters. These challenges are summarized in: (i) big data; (ii) privacy and security; (iii) structure heterogeneity; (iv) UGC format diversity; (v) subjectivity and objectivity; (vi) and service limitations In this thesis, we focus on increasing the quality of social UGC that are published and publicly accessible in forms of posts and profiles over OSNs through addressing in-depth the stated serious challenges. As the social spam is the most common IQ problem appearing over the OSM, we introduce a design of two generic approaches for detecting and filtering out the spam content. The first approach is for detecting the spam posts (e.g., spam tweets) in a real-time stream, while the other approach is dedicated for handling a big data collection of social profiles (e.g., Twitter accounts)
APA, Harvard, Vancouver, ISO, and other styles
17

Zhu, Wei. "Non-Lattice Based Ontology Quality Assurance." Case Western Reserve University School of Graduate Studies / OhioLINK, 2019. http://rave.ohiolink.edu/etdc/view?acc_num=case1558509364811856.

Full text
APA, Harvard, Vancouver, ISO, and other styles
18

Wong, Ho-chuen, and 黃浩川. "Analysis and prediction of beach water quality in Hong Kong: with special reference to Big Wave BayBeach." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2009. http://hub.hku.hk/bib/B43278619.

Full text
APA, Harvard, Vancouver, ISO, and other styles
19

Dahlström, Viktor, and Robin Danielsson. "Levererar Big-4 en högre revisionskvalitet jämfört med Non-Big 4? : En kvantitativ studie som jämför större och mindre revisionsbolags revisionskvalitet relaterat till revisionsarvodet." Thesis, Högskolan i Gävle, Företagsekonomi, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:hig:diva-23802.

Full text
Abstract:
Syfte: Större revisionsbolag har länge inom redovisningslitteraturen associerats med bättre revisionskvalitet jämfört med mindre revisionsbolag. På senare tid har frågan lyfts om större revisionsbolags höga revisionsarvoden beror på högre revisionskvalitet eller marknadsmakt. Denna studie bidrar med nya empiriska underlag inom jämförelsestudier mellan större och mindre revisionsbolag, där revisionsarvodet används som proxy för revisionskvalitet. Vidare tar denna studie, till skillnad från tidigare studier, även hänsyn till revisionskvalitet inom olika riskmiljöer.   Metod: Studien har använt sig av en kvantitativ metod med ett positivistiskt förhållningssätt genom arbetet. En deduktiv forskningsansats har tillämpats där tidigare forskning har legat som grund till studiens framförda hypoteser. Insamling av finansiella sekundärdata för totalt 2518 företag har utförts via databasen Thomson Reuters Datastream.   Resultat & slutsats: Studiens resultat påvisar signifikanta skillnader i revisionskvalitet mellan större och mindre revisionsbolag i studiens olika riskmiljöer. För studiens europeiska länder är förhållandet mellan större och mindre revisionsbolag likvärdiga medan de amerikanska revisionsbolagen skiljer sig signifikant revisionskvalitetsmässigt.   Förslag till fortsatt forskning: Studien har genomförts utan hänsyn tagen till kvalitativa faktorer som kan komma att påverka revisionskvalitet, vilket öppnar ett utrymme för komparativa studier med en kvalitativ inriktning. Det finns även möjlighet att utöka antalet börsmarknader för respektive land eller utvidga antalet länder i olika riskmiljöer.   Uppsatsens bidrag: Studien lämnar två bidrag till redovisningslitteraturen i form av nya empiriska bevis inom revisionskvalitet mellan större och mindre revisionsbolag samt unik forskning kring revisionskvalitetsstudier mellan riskmiljöer. Vidare lämnar studiens resultat incitament åt praktiker att granska revisionsmarknaden för eget vinstintresse samt svarar på normgivares funderingar kring marknadsbalansen mellan större och mindre revisionsbolag.
Aim: Big audit firms have long been associated with higher audit quality, compared to smaller audit firms. Recent studies suggest that the higher audit fees from bigger audit firm is affected by market misuse rather than better audit quality. This study provides new empirical evidence between the comparison of big vs small audit firm, where audit fees are used as proxy for audit quality. Furthermore, this study investigates different litigation environment that could affect audit quality.   Method: This study uses an quantitative based method with an positivist, deductive approach, were earlier studies have had an impact on our hypotheses. Financial information from 2518 companies has been collected from Thomson Reuters Datastream.   Result & Conclusions: This study's result provides significant differences of audit quality between big and small audit firms in different risk environments. For this study, the audit quality relationship between big and small audit firms are equivalent for the European countries while audit quality between big and small audit firms in the US differ significantly.   Contribution of the thesis: This study leaves two contributions to the extent audit literature, in terms of empirical evidence of audit quality between big and small audit firms and unique research results of audit quality in different litigation environments. Furthermore, the results of this study creates incentives for practitioners to review the audit market for self interests and answer legal setters concerns about unbalanced audit markets.   Suggestions for future research: The study has been carried out without consideration of qualitative factors that may affect audit quality. It’s opening a space for comparative studies with an qualitative approach. It is also possible to expand the number of stock exchanges for a country or expand the number of countries in different risk environments.
APA, Harvard, Vancouver, ISO, and other styles
20

Loughman, Kathleen Riha. "The effects of dams in the Big Sandy watershed using a novel bacteria-based bioindicator of water quality." Huntington, WV : [Marshall University Libraries], 2005. http://www.marshall.edu/etd/descript.asp?ref=531.

Full text
Abstract:
Theses (M.S.)--Marshall University, 2005.
Title from document title page. Includes abstract. Document formatted into pages: contains viii, 105 p. including illustrations and maps. Bibliography: p. 37-40.
APA, Harvard, Vancouver, ISO, and other styles
21

Hasselblom, Miranda, and Erik Ernstsson. "Hur skiljer sig kvalitetskontroller vid revision mellan Big 4 och små revisionsbyråer?" Thesis, Högskolan i Halmstad, Akademin för ekonomi, teknik och naturvetenskap, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:hh:diva-42512.

Full text
Abstract:
Dagens revisionsbransch är stöpt ur företagsskandaler och har under drygt hundra år utvecklats. Många revisionsbyråer har kommit och gått och det samma gäller reglering kring yrket. ISQC 1 är en standard som började gälla i Sverige under 2009 och som ställer krav på att revisionsbyråerna har kvalitetskontrollsystem. Men hur ser egentligen efterlevnaden av denna standard ut i olika revisionsbyråer? I studien tittar vi närmare på två grupper, Big 4 och smårevisionsbyråer. Syftet med studien är att undersöka och jämföra kvalitetskontrollerna i revisionsprocessen hos revisionsbyråer av olika storlek. För att kunna genomföra detta tittar vi i uppsatsen närmare på likheter och skillnader i byråernas kvalitetskontrollsystem. Detta för att analysera om några eventuella samband finns i termer av kvalitetskontroll mellan de olika byråerna beroende på deras storlek samt försöka förklara vad dessa samband kan bero på. Studien utförs med en induktiv ansats och har subjektiva förtecken. Forskningen bedrivs på ett kvalitativt sätt med hjälp av ett intervjuförfarande av två revisorer från Big 4 och fem revisorer från små byråer som utgör uppsatsens empiri. Det görs också en kompletterande studie av samtliga ur Big 4:s “transparency reports”, vilket är ett dokument där respektive byrå beskrivit sina system för kvalitetssäkring. För ökad förståelse görs en dataanalys på samtliga intervjuer, indelade i en mönsterkodning per grupp. Uppsatsen innefattar även en referensram med tidigare forskning kring olika delar inom ISQC 1 och revisionskvalitet samt kring det teoretiska perspektivet vilket är utifrån den institutionella teorin med inriktning på tvingande-, mimetisk och normativ isomorfismer. I studiens analys och slutsatser framgår att Big 4 har ett egenutvecklat större digitalt system, medan små revisionsbyråer istället har enklare lösningar som policys och kvalitetshandböcker för att uppfylla kraven i ISQC 1. Big 4 har tydliga befattningar avseende ansvar, medan små byråer har mer decentraliserad struktur kring ledningsansvar för kvalitet. Något som är gemensamt för båda grupperna är dock att den påskrivande revisorn har huvudansvaret vid pågående uppdrag. Yrkesetiska krav är något som båda grupper följer, men där Big 4 även interagerar egna etiska värderingar utöver de krav som ställs på IESBA. Oberoendet är något som är otroligt viktigt för samtliga revisorer och som genomsyrar hela deras arbete. Accepterandet av nya kunder är den del som skiljer sig mellan grupperna i denna kategori då Big 4:s acceptans sköts centralt i Stockholm, medan små revisionsbyråer gör det lokalt. Vid bibehållandet av kundrelationer sker det dock lokalt både hos Big 4 och små revisionsbyråer. Big 4 och små revisionsbyråer har liknande krav vid anställande av ny personal och även vid vidareutbildning av befintlig. Hur uppdragen utförs är också likt mellan grupperna då båda har revisionsprogram att följa, även om dessa är inköpta av små revisionsbyråer och egenutvecklade av Big 4. Stora skillnader finns mellan Big 4 och små revisionsbyråer vid övervakning då Big 4 har godkända interna kontroller, varför FAR och RI inte behöver kontrollera enskilda revisorer, medan små revisionsbyråers revisorer blir individuellt kontrollerade av FAR eller RI. Oavsett omfattning på byråns kvalitetskontrollsystem eller storleken på revisionsbyrån uppnås ändå kraven i ISQC 1.
APA, Harvard, Vancouver, ISO, and other styles
22

Moquist, Sundh Ellinor. "Are HiPPOs losing power in organizational decision-making? : An exploratory study on the adoption of Big Data Analytics." Thesis, Internationella Handelshögskolan, Högskolan i Jönköping, IHH, Informatik, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:hj:diva-44556.

Full text
Abstract:
Background: In the past decades, big data (BD) has become a buzzword which is associated with the opportunities of gaining competitive advantage and enhanced business performance. However, data in a vacuum is not valuable, but its value can be harnessed when used to drive decision-making. Consequently, big data analytics (BDA) is required to generate insights from BD. Nevertheless, many companies are struggling in adopting BDA and creating value. Namely, organizations need to deal with the hard work necessary to benefit from the analytics initiatives. Therefore, businesses need to understand how they can effectively manage the adoption of BDA to reach decision-making quality. The study answers the following research questions: What factors could influence the adoption of BDA in decision-making? How can the adoption of BDA affect the quality of decision-making? Purpose: The purpose of this study is to explore the opportunities and challenges of adopting big data analytics in organizational decision-making. Method: Data is collected through interviews based on a theoretical framework. The empirical findings are deductively and inductively analysed to answer the research questions. Conclusion: To harness value from BDA, companies need to deal with several challenges and develop capabilities, leading to decision-maker quality. The major challenges of BDA adoption are talent management, leadership focus, organizational culture, technology management, regulation compliance and strategy alignment. Companies should aim to develop capabilities regarding: knowledge exchange, collaboration, process integration, routinization, flexible infrastructure, big data source quality and decision maker quality. Potential opportunities generated from the adoption of BDA, leading to improved decision-making quality, are: automated decision-making, predictive analytics and more confident decision makers.
APA, Harvard, Vancouver, ISO, and other styles
23

Andersson, Linn, and Elin Österberg. "Resurser i icke Big 4 byråer : En studie ur ett revisionskvalitetsperspektiv." Thesis, Högskolan Kristianstad, Sektionen för hälsa och samhälle, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:hkr:diva-16943.

Full text
Abstract:
Revisionskvalitet är ett återkommande ämne i såväl forskning som i media. Forskare belyser olika faktorer som påverkar revisionskvalitet och den här studien fokuserar på resurser. Majoriteten av tidigare forskning om revisionskvalitet har Big 4 byråer som utgångspunkt. För att skapa bättre förståelse för hela revisionsprofessionen studeras därför en mindre byrå i den här studien. Syftet är att utforska resurser för revisorer på icke Big 4 byråer, utifrån ett revisionskvalitetsperspektiv. Studien har både deduktiva och induktiva inslag där det deduktiva inslaget utgörs av en teoretisk referensram som ligger till grund för studien. Det induktiva inslaget har gjort det möjligt att studera resurser dels utifrån tillgången till, men även bristen på och användningen av resurser i icke Big 4 byråer. Det har även gjort det möjligt att lägga till ett ytterligare perspektiv då empirin påvisade faktorer som inte identifierats innan datainsamlingen påbörjades. Studiens empiri består till stor del av primärdata från deltagande observationer, men även från semistrukturerade intervjuer. Resultatet visar att revisorerna på Revisionsbyrå A har tillgång till, och använder, samtliga resurser som tidigare forskning visat påverka revisionskvalitet positivt. Resultatet visar även tre nyfunna resurser som tillhör det nyfunna kapitalet materiellt kapital, och även de resurserna har en positiv påverkan på revisionskvalitet. Slutsatsen är att revisorerna på Revisionsbyrå A, utifrån dess tillgång till resurser, kan leverera hög revisionskvalitet. Studiens resultat är baserat på empiri från en mindre byrå och det teoretiska bidraget kompletterar tidigare forskning om revisionskvalitet i Big 4 och icke Big 4 byråer. Resultatet av studien är även ett bidrag till revision och revisionskvalitet praktiskt och empiriskt.
Audit quality is a reoccurring field in both research and media. Researchers highlight different factors that affect audit quality and this study focuses on resources. The majority of previous research of audit quality have Big 4 firms as their starting-point. In order to gain a better understanding of the entire audit profession, a smaller firm is the focus of this study. The purpose is to explore resources for auditors in non-Big 4 firms, from an audit quality perspective. The study has both deductive and inductive characteristics; the deductive characteristics are visible through a theoretical framework that functions as a foundation for the study. The inductive characteristics have made it possible to research the resources through both the access to, yet also the lack of, as well as the use of resources in non-Big 4 firms. It has, moreover, made it possible to add another perspective since the empirical data showed factors that had not been identified before the data collection was initiated. The empirical data of this study consists to a large extent of primary data from observations, but also of data from semi-structured interviews. The results show that auditors at Audit Firm A have access to, and use, all the resources that previous research had shown to affect audit quality in a positive way. Moreover, the results show two new resources that belongs to the newly found capital called material capital, and these resources does also have a positive effect on audit quality. The conclusions drawn are that auditors at Audit Firm A can, with respect to their access to resources, deliver high quality audit. The results of the study is based on empirical data from a smaller firm and the theoretical contribution complements previous research of audit quality in Big 4 and non-Big 4 firms. Moreover, the results is also a contribution to audit and audit quality both practical and empirical.
APA, Harvard, Vancouver, ISO, and other styles
24

Dameron-Hager, Irene F. "The contribution of environmental history to the development of a model to aid watershed management a comparative study of the Big Darby Creek and Deer Creek Watersheds in Ohio /." Columbus, Ohio : Ohio State University, 2004. http://rave.ohiolink.edu/etdc/view?acc%5Fnum=osu1078778562.

Full text
Abstract:
Thesis (Ph. D.)--Ohio State University, 2004.
Title from first page of PDF file. Document formatted into pages; contains xiii, 253 p.; also includes graphics (some col.). Includes abstract and vita. Advisor: Earl F. Epstein, Dept. of Natural Resources. Includes bibliographical references (p. 228-238).
APA, Harvard, Vancouver, ISO, and other styles
25

Berg, Marcus. "Evaluating Quality of Online Behavior Data." Thesis, Stockholms universitet, Statistiska institutionen, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:su:diva-97524.

Full text
Abstract:
This thesis has two purposes; emphasizing the importance of data quality of Big Data, and identifying and evaluating potential error sources in JavaScript tracking (a client side on - site online behavior clickstream data collection method commonly used in web analytics). The importance of data quality of Big Data is emphasized through the evaluation of JavaScript tracking. The Total Survey Error framework is applied to JavaScript tracking and 17 nonsampling error sources are identified and evaluated. The bias imposed by these error sources varies from large to small, but the major takeaway is the large number of error sources actually identified. More work is needed. Big Data has much to gain from quality work. Similarly, there is much that can be done with statistics in web analytics.
APA, Harvard, Vancouver, ISO, and other styles
26

Djedaini, Mahfoud. "Automatic assessment of OLAP exploration quality." Thesis, Tours, 2017. http://www.theses.fr/2017TOUR4038/document.

Full text
Abstract:
Avant l’arrivée du Big Data, la quantité de données contenues dans les bases de données était relativement faible et donc plutôt simple à analyser. Dans ce contexte, le principal défi dans ce domaine était d’optimiser le stockage des données, mais aussi et surtout le temps de réponse des Systèmes de Gestion de Bases de Données (SGBD). De nombreux benchmarks, notamment ceux du consortium TPC, ont été mis en place pour permettre l’évaluation des différents systèmes existants dans des conditions similaires. Cependant, l’arrivée de Big Data a complètement changé la situation, avec de plus en plus de données générées de jour en jour. Parallèlement à l’augmentation de la mémoire disponible, nous avons assisté à l’émergence de nouvelles méthodes de stockage basées sur des systèmes distribués tels que le système de fichiers HDFS utilisé notamment dans Hadoop pour couvrir les besoins de stockage technique et le traitement Big Data. L’augmentation du volume de données rend donc leur analyse beaucoup plus difficile. Dans ce contexte, il ne s’agit pas tant de mesurer la vitesse de récupération des données, mais plutôt de produire des séquences de requêtes cohérentes pour identifier rapidement les zones d’intérêt dans les données, ce qui permet d’analyser ces zones plus en profondeur, et d’extraire des informations permettant une prise de décision éclairée
In a Big Data context, traditional data analysis is becoming more and more tedious. Many approaches have been designed and developed to support analysts in their exploration tasks. However, there is no automatic, unified method for evaluating the quality of support for these different approaches. Current benchmarks focus mainly on the evaluation of systems in terms of temporal, energy or financial performance. In this thesis, we propose a model, based on supervised automatic leaming methods, to evaluate the quality of an OLAP exploration. We use this model to build an evaluation benchmark of exploration support sys.terns, the general principle of which is to allow these systems to generate explorations and then to evaluate them through the explorations they produce
APA, Harvard, Vancouver, ISO, and other styles
27

Duško, Petrović. "Uticaj personalnih karakteristika zaposlenih u uslužnim sistemima na performanse organizacije." Phd thesis, Univerzitet u Novom Sadu, Fakultet tehničkih nauka u Novom Sadu, 2020. https://www.cris.uns.ac.rs/record.jsf?recordId=112490&source=NDLTD&language=en.

Full text
Abstract:
Ispitivanje uticaja personalnih karakteristika davaoca usluga (osobine ličnosti imoral) na procenu kvaliteta usluge od strane korisnika usluga i procenu performansipreduzeća (rekreacionih cenrara) pokazalo je da personalne karakteristike davaocausluge direktno utiču na kvalitet usluge i performanse.
An examination of the impact of the personal characteristics of the service provider(personality traits and morale) on the assessment of the quality of service by theservice users and the assessment of the performance of the company (recreationalcenters) showed that the personal characteristics of the service provider directlyaffect the quality of service and performance.
APA, Harvard, Vancouver, ISO, and other styles
28

Mostert, Analene C. "Meat quality of kudu (Tragelaphus strepsiceros) and impala (Aepyceros melampus)." Thesis, Stellenbosch : Stellenbosch University, 2007. http://hdl.handle.net/10019.1/18597.

Full text
Abstract:
Thesis (MScAgric)--University of Stellenbosch, 2007.
ENGLISH ABSTRACT: Although kudu (Tragelaphus strepsiceros) and impala (Aepyceros melampus) are found in the same geographical area, there is variation in their diets as kudu are predominantly browsers, feeding on tree and shrub leaves, while impala are known as mixed feeders as they graze and browse. Therefore this poses the question whether the diet would influence their meat quality. The objective of this investigation was to evaluate the physical measurements and chemical composition of M. longissimus dorsi, M. biceps femoris, M. semimembranosus, M. semitendinosus and M. supraspinatus for kudu and impala, two southern African antelope species. The effects of age (adult and sub-adult) and gender (male and female) were also determined. The sensory characteristics of the M. longissimus dorsi muscle for sub-adult kudu and impala were investigated. Correlations between the various physical measurements and chemical composition of the meat were verified. Physical measurements and chemical composition of the M. longissimus dorsi muscle were tested for correlations with the sensory ratings of the meat. Dressing percentage of impala (59.88%) (n=28) was higher than that of kudu (57.60%) (n=35). The main effects (species, gender and age) showed no differences for drip loss and cooking loss. However, muscles differed in terms of cooking loss with impala M. semitendinosus having the highest (38.28%) value and kudu M. longissimus dorsi having the lowest value (30.77%). For impala, the highest Warner–Bratzler shear (WBS) values were measured for M. semimembranosus (5.90 kg/1.27cmø), followed by M. biceps femoris, M. longissimus dorsi, and M. semitendinosus with the lowest WBS values measured for M. supraspinatus (3.61 kg/1.27cmø). All impala muscles had lower L* values and appeared darker in colour than kudu muscles, except for M. supraspinatus. Adult animals also had lower L* values than the sub-adult group. Kudu had significantly higher a* and b* values (more red) than impala. Chroma values were higher for kudu, thus appearing brighter in colour. The respective muscles of kudu and impala investigated differed significantly in terms of physical characteristics. However, gender and age did not have an effect on the physical measurements. Moisture content was higher in kudu meat (76.46%) than in impala meat (75.28%). Muscles differed for both moisture and fat content. The highest fat was found in M. supraspinatus followed by M. biceps femoris, M. semitendinosus, M. semimembranosus and M. longissimus dorsi. Protein content did not differ between species (kudu: 21.66%; impala: 22.26%), gender (male: 21.98%; female: 21.95%) and age groups (adult: 21.74%; sub-adult: 22.18%). Kudu M. longissimus dorsi (1.62%) had lower fat content than impala M. longissimus dorsi (2.22%) and female animals had a higher fat content than male animals. Sub-adults (1.20 ± 0.02%) had higher ash content than adults (1.10 ± 0.03%). The M. supraspinatus had the lowest protein and also the highest fat content, with M. semimembranosus having the lowest fat content but the highest value for protein. Myoglobin content did not differ between species, although females had higher (6.58 ± 0.20 mg/g) myoglobin content than males (5.11 ± 0.25 mg/g). Glycolitic muscles had the lowest myoglobin content with the highest values found in M. supraspinatus, an oxidative muscle. An interaction was noted between species and muscle for myoglobin content. Myoglobin content in impala M. longissimus dorsi was higher than that in kudu M. longissimus dorsi; however for all other muscles the myoglobin content was lower in impala. Gender did not affect mineral content. Potassium levels were highest for kudu while phosphorus was more prevalent in impala meat. Adult and sub-adult groups differed in terms of potassium, calcium and zinc content. Potassium and calcium content were higher for subadult animals while zinc content was higher in adult animals. In impala meat, stearic acid (22.67%) was the major fatty acid, followed by palmitic acid (16.66%). In contrast, oleic acid (24.35%) was the most profuse fatty acid in kudu, followed by linoleic acid (22.95%). The SFA’s as a percentage of the total fatty acids differed between impala (51.12%) and kudu meat (34.87%). Kudu meat had a higher concentration of total PUFA (38.88%) than impala (34.06%) meat. The PUFA: SFA ratio for kudu meat (1.22) was more favourable than that for impala meat (0.73). The ratio of n-6 PUFA’s to n-3 PUFA’s for kudu and impala were determined as 2.22 and 3.76 respectively. From the current findings it is evident that kudu and impala meat have advantageous fatty acid profiles and can be a healthy substitute for other red meats. Kudu meat (72.62 ± 1.86 mg/100g) had higher cholesterol than impala meat (55.35 ± 1.84 mg/100g). It is recommended that further studies be done in order to confirm the cholesterol content of kudu meat. Within species, no gender differences for any of the sensory characteristics tested were noted. The impala meat had a more intense game aroma than the kudu meat, while kudu meat was found to be more juicy than impala meat. It can therefore be concluded that the marketing of game meat should be species-specific as there are distinct flavour and aroma differences between kudu and impala meat.
AFRIKAANSE OPSOMMING: Alhoewel koedoes (Tragelaphus strepsiceros) en rooibokke (Aepyceros melampus) in dieselfde geografiese area voorkom, is daar variasie in hulle diëte. Koedoes is hoofsaaklik blaarvreters, terwyl rooibokke bekend staan as gemengde vreters aangesien hulle grassowel as blaarvreters is. Die vraag ontstaan dus of die verskil in diëet die kwaliteit van hulle vleis sal beϊnvloed. Die doel van hierdie ondersoek was dus om die fisiese metings en chemiese samestelling van die M. longissimus dorsi, M. biceps femoris, M. semimembranosus, M. semitendinosus en M. supraspinatus vir koedoes en rooibokke te bepaal. Die invloed van ouderdom (volwasse en onvolwasse) en geslag (manlik en vroulik) op hierdie eienskappe is ook geëvalueer. Die sensoriese eienskappe van die M. longissimus dorsi van onvolwasse koedoes en rooibokke is ook ondersoek. Korrelasies tussen die fisiese metings en chemiese samestelling van die vleis is ondersoek. Die fisiese metings en chemiese samestelling van die M. longissimus dorsi is getoets vir korrelasies met die resultate van die sintuiglike evaluering van die vleis. Die gemiddelde uitslagpersentasie van rooibokke (59.88%) (n=28) was hoër as die van koedoes (57.60%) (n=35). Daar was geen verskille in drupverlies en kookverlies vir die hoofeffekte (spesie, geslag en ouderdom) nie. Spiere het wel verskil in terme van kookverlies, met die hoogste waarde gemeet vir rooibok M. semitendinosus (38.28%) en die laagste waarde vir koedoe M. longissimus dorsi (30.77%). In rooibokke was die hoogste Warner- Bratzler skeurkrag waardes gemeet vir M. semimembranosus (5.76 kg/1.27cmø), gevolg deur M. biceps femoris, M. longissimus dorsi, en M. semitendinosus met die laagste Warner- Bratzler skeurkrag waardes gemeet vir M. supraspinatus (3.78 kg/1.27cmø). Alle rooibokspiere het laer L* waardes gehad en was donkerder van kleur as koedoespiere, behalwe vir M. supraspinatus. Laer L* waardes is ook verkry vir volwasse diere in vergelyking met onvolwasse diere. Die a* en b* waardes was hoër in koedoe- as in rooibokvleis, m.a.w. koedoevleis het rooier vertoon. Die onderskeie koedoe- en rooibokspiere het betekenisvol verskil in terme van fisiese eienskappe, terwyl geslag en ouderdom geen effek op die fisiese eienskappe gehad het nie. Voginhoud was hoër in koedoe- (75.52%) as in rooibokvleis (74.52%). Verkille tussen spiere is opgemerk vir beide vog- en vetinhoud. M. supraspinatus het die hoogste vetinhoud gehad, gevolg deur M. biceps femoris, M. semitendinosus, M. semimembranosus en M. longissimus dorsi. Geen verskille is opgemerk tussen spesies (koedoe: 21.66%; rooibok: 22.26%), geslagte (manlik: 21.98%; vroulik: 21.95%) en ouderdomme (volwasse: 21.74%; onvolwasse: 22.18%) in terme van proteϊeninhoud nie. Die vetinhoud van koedoe M. longissimus dorsi (1.62%) was laer as dié van rooibok M. longissimus dorsi (2.22%) en die vetinhoud van vroulike diere was hoër as dié van manlike diere. Onvolwasse diere (1.20 ± 0.02%) het ‘n hoër asinhoud as dié van volwasse diere (1.10 ± 0.03%) getoon. In terme van die onderskeie spiere het M. supraspinatus die laagste proteϊen- en die hoogste vetinhoud gehad, terwyl M. semimembranosus die laagste vet- en die hoogste proteϊeninhoud gehad het. Die mioglobieninhoud was nie beϊnvloed deur spesie nie, terwyl vroulike diere ‘n hoër (6.58 ± 0.20 mg/g) mioglobieninhoud as manlike diere (5.11 ± 0.25 mg/g) gehad het. Die M. supraspinatus, ‘n oksidatiewe spier het die hoogste mioglobieninhoud gehad, terwyl glikolitiese spiere die laagste mioglobieninhoud gehad het. ’n Interaksie tussen spesie en spier was opgemerk vir mioglobieninhoud. Rooibok M. longissimus dorsi het ‘n hoër mioglobieninhoud as koedoe M. longissimus dorsi gehad, terwyl die mioglobieninhoud vir al die ander spiere laer was in rooibokke. Mineraalinhoud was nie deur geslag beϊnvloed nie. Kaliumvlakke was hoër in koedoevleis, terwyl fosforvlakke hoër was in rooibokvleis. Kalium- en kalsiuminhoud was hoër in onvolwasse diere terwyl die sinkinhoud hoër was in volwasse diere. Steariensuur (22.67%), gevolg deur palmitiensuur (16.66%) was die mees algemene vetsure in rooibokvleis. In teenstelling hiermee was oleϊensuur (24.35%), gevolg deur linoleϊensuur (22.95%) die mees algemene vetsure in koedoevleis. Die totale versadigde vetsure was laer in koedoevleis (34.87%) in vergelyking met rooibokvleis (51.12%), terwyl die totale polionversadigde vetsure in koedoevleis (38.88%) hoër was as dié van rooibokvleis (34.06%). Die verhouding van n-6 tot n-3 poli-onversadigde vetsure vir koedoe en rooibok was 2.22 en 3.76 onderskeidelik. Hierdie resultate bevestig dat koedoe- en rooibokvleis oor ‘n vetsuurprofiel beskik wat ’n gesonde alternatief bied tot ander rooivleise. Die cholesterolinhoud van koedoevleis (72.62 ± 1.86 mg/100g) was hoër as dié van rooibokvleis (55.35 ± 1.84 mg/100g). Dit word egter aanbeveel dat verdere studies gedoen word om die cholesterolinhoud van koedoevleis te bevestig. Binne spesies was daar geen geslagsverkille vir enige van die sensoriese eienskappe nie. Rooibokvleis het ‘n meer intense wildsvleis aroma as koedoevleis gehad, terwyl koedoevleis meer sappig was as rooibokvleis. Hierdie resultate dui daarop dat die bemarking van wildsvleis spesie-spesifiiek moet wees aangesien daar defnitiewe geur en aroma verskille tussen koedoe- en rooibokvleis is.
APA, Harvard, Vancouver, ISO, and other styles
29

Kwon, Junhyuk. "The Effect of Value Co-creation and Service Quality on Customer Satisfaction and Commitment in Healthcare Management." Thesis, University of North Texas, 2015. https://digital.library.unt.edu/ark:/67531/metadc804961/.

Full text
Abstract:
Despite much interest in service quality and various other service quality measures, scholars appear to have overlooked the overall concept of quality. More specifically, previous research has yet to integrate the effect of the customer network and customer knowledge into the measurement of quality. In this work, it is posited that the evaluation of quality is based on both the delivered value from the provider as well as the value developed from the relationships among customers and between customers and providers. This research examines quality as a broad and complex issue, and uses the “Big Quality” concept within the context of routine healthcare service. The last few decades have witnessed interest and activities surrounding the subject of quality and value co-creation. These are core features of Service-Dominant (S-D) logic theory. In this theory, the customer is a collaborative partner who co-creates value with the firm. Customers create value through the strength of their relations and network, and they take a central role in value actualization as value co-creator. I propose to examine the relationship between quality and the constructs of value co-creation. As well, due to the pivotal role of the decision-making process in customer satisfaction, I will also operationalize the value co-creation construct. Building upon the “Big Quality” concept, this study suggests a new approach by extending the quality concept to include the value-creation concept in Service Dominant Logic. This study identifies the associated constructs and determinants of Big Quality in routine healthcare management service, and examines the relationship among the associated quality constructs, customer satisfaction, and customer commitment. This study employed an online survey methodology to collect data. In data analysis, I used the variance-based structural equation modeling (PLS-SEM) approach to confirm the factor structure, proposed model, and test the research hypotheses. The results show that the customer’s participation in in-role and extra-role behaviors are positively associated with their perceived quality, satisfaction, and commitment level. The major contribution of this study to decision sciences and the service quality literature is the development of a comprehensive framework explaining the importance of value co-creation within the context of healthcare quality. Finally, this work examines perceived service quality as a key factor of customer satisfaction and the relationship of Big Quality with commitment level in healthcare service management.
APA, Harvard, Vancouver, ISO, and other styles
30

Nesvijevskaia, Anna. "Phénomène Big Data en entreprise : processus projet, génération de valeur et Médiation Homme-Données." Thesis, Paris, CNAM, 2019. http://www.theses.fr/2019CNAM1247.

Full text
Abstract:
Le Big Data, phénomène sociotechnique porteur de mythes, se traduit dans les entreprises par la mise en place de premiers projets, plus particulièrement des projets de Data Science. Cependant, ils ne semblent pas générer la valeur espérée. La recherche-action menée au cours de 3 ans sur le terrain, à travers une étude qualitative approfondie de cas multiples, pointe des facteurs clés qui limitent cette génération de valeur, et notamment des modèles de processus projet trop autocentrés. Le résultat est (1) un modèle ajusté de dispositif projet data (Brizo_DS), ouvert et orienté sur les usages, dont la capitalisation de connaissances, destiné à réduire les incertitudes propres à ces projets exploratoires, et transposable à l’échelle d’une gestion de portefeuille de projets data en entreprise. Il est complété par (2) un outil de documentation de la qualité des données traitées, le Databook, et par (3) un dispositif de Médiation Homme-Données, qui garantissent l’alignement des acteurs vers un résultat optimal
Big Data, a sociotechnical phenomenon carrying myths, is reflected in companies by the implementation of first projects, especially Data Science projects. However, they do not seem to generate the expected value. The action-research carried out over the course of 3 years in the field, through an in-depth qualitative study of multiple cases, points to key factors that limit this generation of value, including overly self-contained project process models. The result is (1) an open data project model (Brizo_DS), orientated on the usage, including knowledge capitalization, intended to reduce the uncertainties inherent in these exploratory projects, and transferable to the scale of portfolio management of corporate data projects. It is completed with (2) a tool for documenting the quality of the processed data, the Databook, and (3) a Human-Data Mediation device, which guarantee the alignment of the actors towards an optimal result
APA, Harvard, Vancouver, ISO, and other styles
31

Jiverud, Lina, and Ida Vikström. "Revisorn och varning om fortsatt drift : En kvantitativ studie om faktorer som påverkar revisionskvaliteten." Thesis, Södertörns högskola, Företagsekonomi, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:sh:diva-35764.

Full text
Abstract:
Efter flertalet revisionsskandaler under åren har det blivit populärt att kritisera revisorn och ifrågasätta kvaliteten på den utförda revisionen. Revisorn har ett ansvar att informera om det finns väsentliga osäkerheter kring ett företags förutsättningar för fortsatt drift och att utfärda en fortlevnadsvarning innebär en hel del tuffa beslut och bedömningar. Dock finns det faktorer som kan påverka revisorns agerande så som revisorns närhet till klienten, byråtillhörighet och risken för rättsliga åtgärder. Tidigare studier har även visat på att det finns skillnader i revisionskvaliteten mellan män och kvinnor. Syftet med studien är att undersöka hur olika faktorer påverkar revisorns utfärdande av fortlevnadsvarning och därmed se vad det har för påverkan på revisionskvaliteten. Syftet är även att undersöka eventuella skillnader som finns i den svenska kontexten jämfört med tidigare studier i andra länder. För att undersöka sambandet genomfördes en kvantitativ studie där årsredovisningar och revisionsberättelser från svenska konkursdrabbade företag studerades. Studiens resultat visar på att det finns ett samband mellan finansiell stress och revisorns utfärdande av fortlevnadsvarning men dock gick det inte att påvisa något samband för variablerna kön, byråtillhörighet och ålder på företag. Resultatet visar även att revisionskvaliteten och utfärdandet av fortlevnadsvarningar i Sverige skiljer sig från andra länder.
After a numerous audit scandals over the years, it has become popular to criticize the auditor and question the audit quality. The auditor has a responsibility to inform if they have significant doubts about the continuity of a company. The issuing of a going-concern involves a lot of tough decisions and assessments for the auditor. However, there are factors that may affect the auditor’s actions such as the auditor’s closeness to the client, audit firm and the risk of legal action. Previous studies have also shown that there’s a difference in audit quality between men and women. The purpose with this paper is to study how different factors affect the auditor’s issuance of a going-concern and thus see how it affects the audit quality. The purpose is also to study any differences that exist in the Swedish context as compared to previous studies in other countries. To examine the relationship between going-concern and the variables, a quantitative study was conducted in which annual reports and audit reports from Swedish bankrupt companies were studied. The results of the study indicate that there is a relation between financial distress and a going-concern, but it was not possible to detect any relation to variables such as gender, audit firm and the age of the client. The result also shows that the audit quality and the issuance of a going-concern in Sweden differs from other countries.
APA, Harvard, Vancouver, ISO, and other styles
32

Ephraim, Ekow Esson, and Sanel Sehic. "The Use of Big Data in Process Management : A Literature Study and Survey Investigation." Thesis, Linköpings universitet, Logistik- och kvalitetsutveckling, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-177921.

Full text
Abstract:
In recent years there has been an increasing interest in understanding how organizations can utilize big data in their process management to create value and improve their processes. This is due to new challenges for process management which have arisen from increasing competition and the complexity of large data sets due to technological advancements. These large data sets have been described by scholars as big data which involves data that are so complex traditional data analysis software are not sufficient in managing or analyzing them. Because of the complexity of handling such great volumes of data there is a big gap in practical examples where organizations have incorporated big data in their process management. Therefore, in order to fill relevant gaps and contribute to advancements in this field, this thesis will explore how big data can contribute to improved process management. Hence, the aim of this thesis entailed investigating how, why and to what extent big data is used in process management. As well as to outline the purpose and challenges of using big data in process management. This was accomplished through a literature review and a survey, respectively, in order to understand how big data had previously been used to create value and improve processes in organizations. From the extensive literature review, an analysis matrix of how big data is used in process management is provided through the intersections between big data and process management dimensions. The analysis matrix showed that most of the instances in which big data was used in process management were in process analysis & improvement and process control & agility. Simply put, organizations used big data in specific activities involved in process management but not in a holistic manner. Furthermore, the limited findings from the survey indicate that the main challenges and purposes of big data use in Swedish organizations are the complexity of handling data and making statistically better decisions, respectively.
APA, Harvard, Vancouver, ISO, and other styles
33

Rangnitt, Eric, and Louise Wiljander. "Tillförlitlighet hos Big Social Data : En fallstudie om upplevd problematik kopplat till beslutfattande i en organisationskontext." Thesis, Uppsala universitet, Institutionen för informatik och media, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-414032.

Full text
Abstract:
Den växande globala användningen av sociala medier skapar enorma mängder social data online, kallat för Big Social Data (BSD). Tidigare forskning lyfter problem med att BSD ofta har bristande tillförlitlighet som underlag vid beslutsfattande och att det är starkt kopplat till dataoch informationskvalitet. Det finns dock en avsaknad av forskning som redogör för praktikers perspektiv på detta. Därför undersökte denna studie vad som upplevs problematiskt kring transformation av BSD till tillförlitlig information för beslutsfattande i en organisationskontext, samt hur detta skiljer sig i teori jämfört med praktik. En fallstudie gjordes av mjukvaruföretaget SAS Institute (SAS). Datainsamlingen genomfördes via intervjuer samt insamling av dokument och resultaten analyserades kvalitativt. Studien gjorde många intressanta fynd gällande upplevda problem kopplat till transformation av BSD, bl.a. hög risk för partisk data och låg analysmognad, samt flera skillnader mellan teori och praktik. Tidigare forskning gör inte heller skillnad mellan begreppen datakvalitet och informationskvalitet, vilket görs i praktiken.
The growing use of social media generates enormous amounts of online social data, called Big Social Data (BSD). Previous research highlights problems with BSD reliability related to decision making, and that reliability is strongly connected to data quality and information quality. However, there is a lack of research with a focus on practitioners’ perspectives on this matter. To address this gap, this study set out to investigate what is perceived as a problem when transforming BSD into reliable information for decision making in an organisational context, and also how this differs in theory compared with practice. A case study was conducted of the software company SAS Institute (SAS). Data collection was done through interviews and gathering of documents, and results were analysed qualitatively. The study resulted in many interesting findings regarding perceived problems connected to the transformation of BSD, e.g. high risk of biased data and low maturity regarding data analysis, as well as several differences between theory and practice. Furthermore, previous research makes no distinction between the terms data quality and information quality, but this is done in practice.
APA, Harvard, Vancouver, ISO, and other styles
34

Agné, Alvin, and Maiju Ruokanen. "Revisionsbyråers kvalité: en studie om Going Concern-varningar och anmärkningar på finansiell stress." Thesis, Mittuniversitetet, Institutionen för samhällsvetenskap, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:miun:diva-14132.

Full text
Abstract:
Träffsäkerheten på Going Concern-varningar (GC-varningar) internationellt sett är låg, 40 procent, men enligt tidigare studier är den lägre i Sverige, under 20 procent. De fyra största revisionsbyråerna i världen, Big four, marknadsför sig med sin branschkännedom och flera studier pekar på att de har högre revisionskvalité än mindre byråer. Samtidigt finns det studier som visar att det inte finns någon skillnad på kvalitén mellan de fyra största och de mindre byråerna. Syftet med denna studie är att pröva om stora byråer har högre kvalité än medelstora och små byråer och om det finns skillnader mellan de enskilda byråerna inom respektive storlekskategori. Revisionskvalité mäter vi i form av GC-varningar och anmärkningar på finansiell stress (eget kapital understiger hälften av aktiekapitalet). Studien är kvantitativ och våra data omfattar 4718 aktiebolag, varav 1809 är finansiellt stressade, som gick i konkurs 2010. Genom Chi 2 test ser vi att träffsäkerheten på GC-varningar ligger på 17 procent och anmärkningar på finansiell stress på 82 procent. På samma sätt men med ytterligare hjälp av sambandsstyrka genom P-värden har vi kommit fram till våra andra resultat. Det finns ett väldigt starkt samband mellan andelen GC-varningar och storleken på byrån. Det finns även ett väldigt starkt samband mellan andelen GC-varningar och byrå inom kategorin stora byråer, ett visst samband inom kategorin medelstora byråer men inget samband inom kategorin små byråer. Det finns inget samband mellan andelen anmärkningar på finansiell stress och storleks-kategori eller mellan byråerna inom de olika kategorierna. Enligt vårt resultat har stora byråer högre revisionskvalité än medelstora och små byråer gällande GC-varningar, men vi finner inga skillnader mellan storlekskategorierna gällande anmärkning på finansiell stress. Vi finner även att kvalité med avseende på GC-varningar inte är homogen mellan de enskilda byråerna inom respektive storlekskategori. Våra slutsatser är att träffsäkerheten på GC-varningar är låg, men skulle kunna öka om revisorn kompletterar svårbedömd information med verifierbar information samt att etablerade metoder för GC-bedömningar infördes. Revisionskvalitén inom stora byråer är inte homogen, vilket kan bero på att de enskilda byråerna inom kategorin använder olika arbetssätt, som medför kvalitetsskillnader.
The international accuracy of Going Concern-warnings (GC-warnings) is as low as 40 percent, although studies show the accuracy is even lower in Sweden, under 20 percent. The four biggest accounting firms in the world advertise their knowledge of industry and a number of studies say that they have higher audit quality than the smaller firms. Meanwhile there are studies that claim that there are no differences in quality between them and the smaller firms. In this study we aim to test if the large firms have higher auditing quality than the medium and small firms and to test if there is any difference within the three categories. We measure quality as GC-warnings and a remark on financial stress (a remark on the equity being lower than 50 percent of the share capital). The study is quantitative and our data consists of 4718 limited companies, and 1809 financially stressed limited companies, which went bankrupt 2010. We measured the GC-warning accuracy to 17 percent and the remark on financial stress to 82 percent by using a Chi square test. We used the same test but added P-value to check the strength on the rest of the results. There is a very strong connection between the share of GC-warnings and the size of the accounting firm. There is also a very strong connection between the share of GC-warnings and the different firms within the large category. However there was only a weak connection between the different firms within the medium size and no connection at all within the small category. There is no connection at all between remarks on financial stress and either the firm size or within the firm size categories. According to our results, large audit firms have higher audit quality than medium and small firms concerning GC-warnings, but there are no differences between and within the categories concerning remarks on financial stress. We also find that quality concerning GC-warnings is not homogeneous between the different audit firms within each category. Our conclusions are that the accuracy on GC-warnings is low but it could increase if the auditors complement imponderable information with verifiable information and if methods for GC-judgments were established. The audit quality within large firms is not homogeneous which may be due to that the different audit firms within the categories use different ways to work, which causes differences in audit quality.
APA, Harvard, Vancouver, ISO, and other styles
35

Ashqar, Huthaifa Issam. "Strategic Design of Smart Bike-Sharing Systems for Smart Cities." Diss., Virginia Tech, 2018. http://hdl.handle.net/10919/97827.

Full text
Abstract:
Traffic congestion has become one of the major challenging problems of modern life in many urban areas. This growing problem leads to negative environmental impacts, wasted fuel, lost productivity, and increased travel time. In big cities, trains and buses bring riders to transit stations near shopping and employment centers, but riders then need another transportation mode to reach their final destination, which is known as the last mile problem. A smart bike-sharing system (BSS) can help address this problem and encourage more people to ride public transportation, thus relieving traffic congestion. At the strategic level, we start with proposing a novel two-layer hierarchical classifier that increases the accuracy of traditional transportation mode classification algorithms. In the transportation sector, researchers can use smartphones to track and obtain information of multi-mode trips. These data can be used to recognize the user's transportation mode, which can be then utilized in several different applications; such as planning new BSS instead of using costly surveys. Next, a new method is proposed to quantify the effect of several factors such as weather conditions on the prediction of bike counts at each station. The proposed approach is promising to quantify the effect of various features on BSSs in cases of large networks with big data. Third, these resulted significant features were used to develop state-of-the-art toolbox algorithms to operate BSSs efficiently at two levels: network and station. Finally, we proposed a quality-of-service (QoS) measurement, namely Optimal Occupancy, which considers the impact of inhomogeneity in a BSS. We used one of toolbox algorithms modeled earlier to estimate the proposed QoS. Results revealed that the Optimal Occupancy is beneficial and outperforms the traditionally-known QoS measurement.
PHD
APA, Harvard, Vancouver, ISO, and other styles
36

Kise, Laura Ann. "AN EXAMINATION OF THE TEIQUE AND ITS ASSOCIATION WITH THE BIG FIVE FACTORS, RELATIONSHIP QUALITY, AND GENERAL MENTAL HEALTH IN AN AMERICAN COLLEGE SAMPLE." OpenSIUC, 2014. https://opensiuc.lib.siu.edu/dissertations/917.

Full text
Abstract:
This study sought to clarify trait EI's relationships with general personality and positive life outcomes, extending previous research suggesting that trait EI is a distinct part of personality and that it positively predicts positive life outcomes. This study examined the discriminant validity of trait EI, (as measured by the TEIQue,) relative to the Big Five personality factors via an exploratory joint factor analysis. Additionally, this study assessed the criterion and incremental validity of the TEIQue with regard to positive life outcomes (positive relations with others, perceived social support, and general mental health) controlling for Big Five factor scores from the International Personality Item Pool. Data from 240 undergraduate students from a Midwestern university were analyzed . Trait EI facets and Big Five factors loaded best onto a 3 factor solution similar to Eysenck's 3-factor model of personality, with factor 1 representing emotional stability, factor 2 indicating social-emotional traits, and factor 3 comprising general agreeableness. While these findings did not provide clear evidence of discriminant validity for the TEIQue, support was observed for its criterion and incremental validity. However, potential issues with the measures and methods are discussed.
APA, Harvard, Vancouver, ISO, and other styles
37

Kjellqvist, Lissie, and Carlqvist Teresia Söderberg. "Revisionskvalitet : Kartläggning och analys av huruvida en revisionsbyrås storlek påverkar revisionskvalitet." Thesis, Linköpings universitet, Företagsekonomi, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-131032.

Full text
Abstract:
Revisorer har idag till uppgift att granska och uttala sig kring ett företags ekonomi, dess redovisning och, i Sverige, företagets förvaltning. Revision handlar delvis om att förstå det granskade företaget och dess verksamhet. Höga krav ställs på revisorns kompetens och oberoende, delvis för att intressenter ska kunna fatta viktiga beslut efter en korrekt utförd revision. Uppsatsen studerar huruvida det finns olika faktorer som har en inverkan på kvalitetsskillnader mellan små och stora revisionsbyråer. Vi vill med studiens syfte kartlägga och analysera huruvida revisionsbyråns storlek påverkas dess revisionskvalitet. Vi försöker i denna studie besvara vårt syfte utifrån två olika delstudier. Uppsatsens första delstudie består av en grundlig litteraturgenomgång där vi presenterar, för uppsatsen, relevant forskning. Efter en genomgripande litteraturgenomgång fann vi fem faktorer som vi valde att bygga vidare vår uppsats på. Dessa faktorer var upptäcka brister, rapportering av brister, klientportfölj, rykte och kontorsstorlek. Slutsatsen i delstudie 1 visar att faktorerna torde ha en betydande inverkan på revisionskvalitet och att de stora byråerna har lättare, än de mindre byråerna, att åstadkomma en revision av hög kvalitet. Några exempel på detta är att mycket kompetens på plats torde öka chansen att upptäcka brister i revisionen och att de stora byråernas rykte kan vara gynnsam för byråns klientselektion. Därefter följer delstudie 2 som är en fördjupning av uppsatsens delstudie 1. I denna del utförde vi en intervjustudie där 13 revisor intervjuades från både små och stora revisionsbyråer. Detta gjorde vi för att få en klarare bild över hur revisionskvalitet uppfattas av dagens revisorer och huruvida de ser någon skillnad mellan byråstorlekarna. Vi lade även till komfortfrågor under dessa intervjuer då vi ansåg det vara av stort intresse att tillföra den individuella revisorn och dennes upplevda komfort och huruvida detta kunde ha en inverkan på revisionens uppnådda kvalitet. Intervjustudiens resultat visade inga specifika mönster på att revisionskvaliteten skulle vara högre hos någon av de respektive byråstorlekarna. Vi fann istället den individuella revisorn och dennes komfort som den mest betydande faktorn för revisionskvalitet. Vårt kunskapsbidrag blir således utifrån denna studie att revisionskvaliteten inte torde skilja sig åt mellan byråstorlekarna. Vår analys åskådliggör däremot att den individuella revisorns engagemang och komfort är den avgörande faktorn till huruvida skillnader inom revisionskvalitet uppstår mellan byråer.
Today auditor´s task is to review and express their opinion on a company's finances, their accounts and, in Sweden, the company's management. Audit is partly about understanding the audited company and its operations. There are high demands on the auditor's competence and independence, partly because stakeholders should be able to make important decisions for a proper and thorough audit. The research involves studying whether there are different factors that have an impact on the quality differences between small and large accounting firms. We found the subject interesting and wanted to identify and analyze whether the auditing firm's size affected its audit quality. In this study we tried to answer our purpose from two different studies. The first sub-study consists of a thorough literature review where we present, for this essay, relevant research. After a thorough literature review, we found five factors that we chose to build our essay on. These factors included the identification of gaps, the shortcomings Report, Client Portfolio, Reputation and Office Size. The conclusion of this study shows that these factors have a significant impact on audit quality, and that the larger firms, from these factors, have it easier to achieve higher quality audits. Some examples are that more expertise in the office should increase the chance of detecting deficiencies in the audit, and that the great reputation of the agencies can be beneficial to the Agency's client selection. Then follows substudy 2, which is a deepening of the essay’s substudy 1. In this part we carried out an interview study, in which 13 auditors were interviewed from both small and large accounting firms. We did this in order to get a clearer picture of how audit quality is perceived by today's accountants, and whether they see any difference between the firm sizes. We also added comfort issues during these interviews because we considered it to be of great interest to supply the individual auditor and their perceived comfort, and whether this could have an impact on audit quality. The results of the interviews showed that no specific pattern on audit quality would be higher in any of the respective firm sizes. Interestingly, we instead found the individual auditor and his comfort as the most significant factor for audit quality. Our knowledge contribution is thus, based on this study, that audit quality is not higher in any particular firm when it comes to agency size. Our analysis illustrates, however, that the individual auditor's engagement and comfort are the deciding factors to whether differences in audit quality arises between agencies.
APA, Harvard, Vancouver, ISO, and other styles
38

WANG, Meixin. "An assessment of dual audit effect and contagious effect on the audit quality of non-Big N CPA firms for Chinese companies in different markets." Digital Commons @ Lingnan University, 2014. https://commons.ln.edu.hk/acct_etd/19.

Full text
Abstract:
External auditor is an independent agent to provide assurance about the validity of financial statements prepared by management to enhance the reliability of information in financial reports. As such, audit quality has long been a concern for all stakeholders and is a topic of on-going research interest. In China, the dual audit requirement for AB share companies and AH share companies started in 2001 was abolished in 2007 and 2010 respectively. This study attempts to examine whether there are dual audit effect and contagious effect on the audit quality of non-Big N audit firms for A share companies in different markets. I focus on non-Big N audit firms since the audit quality of these firms are of greater concern. Using data from 2001 to 2012, I compare the audit quality of A share companies that also have B (or H) shares ((AB/H) with the audit quality of pure A share companies to test whether there is a dual audit effect on the audit quality of A-share financial statements. I also compare AB/H share companies which hire only non-Big N auditors with those ABIH share companies who hire non-Big N domestic auditors and Big N international auditors to test the existence of contagious effect on the audit quality of A-share companies. My findings indicate that dual audit does improve the audit quality of non-Big N audit firms for A share companies. However, there was mixed evidences on the contagious effect using different measures of audit quality. This study contributes to the literature on enhancing our understanding of the determinants of audit quality in China. It can also provide policy makers in emerging economies some useful evidence on ways to improve audit quality.
APA, Harvard, Vancouver, ISO, and other styles
39

Pejčoch, David. "Komplexní řízení kvality dat a informací." Doctoral thesis, Vysoká škola ekonomická v Praze, 2010. http://www.nusl.cz/ntk/nusl-199303.

Full text
Abstract:
This work deals with the issue of Data and Information Quality. It critically assesses the current state of knowledge within tvarious methods used for Data Quality Assessment and Data (Information) Quality improvement. It proposes new principles where this critical assessment revealed some gaps. The main idea of this work is the concept of Data and Information Quality Management across the entire universe of data. This universe represents all data sources which respective subject comes into contact with and which are used under its existing or planned processes. For all these data sources this approach considers setting the consistent set of rules, policies and principles with respect to current and potential benefits of these resources and also taking into account the potential risks of their use. An imaginary red thread that runs through the text, the importance of additional knowledge within a process of Data (Information) Quality Management. The introduction of a knowledge base oriented to support the Data (Information) Quality Management (QKB) is therefore one of the fundamental principles of the author proposed a set of best
APA, Harvard, Vancouver, ISO, and other styles
40

Garefelt, Linus, and Marcus Persson. "Korta vs. långa revisionsuppdrag : Hur ser skillnader ut i termer av revisionsprocesser, oberoende och revisionskvalitet?" Thesis, Umeå universitet, Företagsekonomi, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-160576.

Full text
Abstract:
Sammanfattning   Revision är ett än mer aktuellt ämne nu än på flera decennier då det debatteras vitt och brett kring vad en revisor skall eller bör göra. I och med de tidigare införda byrårotationskraven begränsar det företag av allmänt intresse att inte under mer än maximalt tio år använda samma revisionsbyrå, med chans till förlängning. Vi har därför valt att undersöka hur längden på revisionsuppdragen kan inverka på stora onoterade bolag i en mindre geografisk zon, i vårt fall Umeå kommun med omnejd genom att intervjua tre revisorer och deras klienter.   Baserat på problemdiskussionen har vi bestämt följande problemformulering för studien.      Hur upplever revisionsbyråerna och deras klienter att oberoendet samt revisionskvalitet samspelar med revisionsprocessen vid korta respektive långa revisionsuppdrag?   Syftet med vår studie är att undersöka hur förhållandet mellan klient och revisor upplevs ha en samverkan med kvaliteten i revisionen samt på revisorernas oberoende till deras klienter över tid. Detta kan ligga till grund för framtida diskussioner gällande byrårotation för huruvida oberoende och revisionskvalitet utvecklas över tid under revisionsuppdraget.   Med en kvalitativ ansats har vi genomfört nio intervjuer med tre revisorer samt sex klienter. Klientförhållandena till respektive revisor är ett långvarigt revisionsuppdrag (fler än fem år) och ett kortvarigt revisionsuppdrag (mindre än 5 år sedan övertagandet). Denna unika inblick i revisor-klientförhållandet kopplas till flertalet faktorer samt teorier vilka kan ha en samverkan och förklaring till förändringen i oberoendet samt revisionskvaliteten.    Vi kommer i vår studie fram till att revisionsprocessen ser väldigt lik ut vid korta samt långa revisionsuppdrag där första året är speciellt omfattande, men år 2 och framåt vanligtvis är lika omfattande år efter år. Vad gäller oberoende finner vi att desto längre relation som skapas mellan klient och revisionsbyrå desto mer omfattande blir rådgivningen till klienten vilket kan påverka revisionsbyråns oberoende mot klienten rent finansiellt. Vidare är revisionskvalitet en term som definieras olika, men i teorin hävdar flertalet att detta kan speglas av oberoendet till klienten. I vår studie har definitionen uppfattats olika mellan revisor och klient vilket visar på ett förväntningsgap mellan de båda. Enligt respondenterna är revisionskvalitet något som skulle kunna förbättras av byrårotation när flera anser att långa revisionsuppdrag kan bli för rutinmässiga och att det skulle vara bra om nya individer kommer in och tittar på uppdraget från ett nytt perspektiv.   Förslag till vidare forskning skulle till exempel vara en enkätundersökning med ett mer omfattande geografiskt område där fler företag av större storlek innefattas av populationen för att kunna jämställa dessa med de företag som redan omfattas av reglerna om obligatorisk byrårotation.   Nyckelord: Audit, big four, big seven, audit quality, audit independence, audit firm rotation, auditor rotation, expectation gap, evidential matter, audit procedures och audit tenure.
APA, Harvard, Vancouver, ISO, and other styles
41

Bycroft, Clare. "Genomic data analyses for population history and population health." Thesis, University of Oxford, 2017. https://ora.ox.ac.uk/objects/uuid:c8a76d94-ded6-4a16-b5af-09bbad6292a2.

Full text
Abstract:
Many of the patterns of genetic variation we observe today have arisen via the complex dynamics of interactions and isolation of historic human populations. In this thesis, we focus on two important features of the genetics of populations that can be used to learn about human history: population structure and admixture. The Iberian peninsula has a complex demographic history, as well as rich linguistic and cultural diversity. However, previous studies using small genomic regions (such as Y-chromosome and mtDNA) as well as genome-wide data have so far detected limited genetic structure in Iberia. Larger datasets and powerful new statistical methods that exploit information in the correlation structure of nearby genetic markers have made it possible to detect and characterise genetic differentiation at fine geographic scales. We performed the largest and most comprehensive study of Spanish population structure to date by analysing genotyping array data for ~1,400 Spanish individuals genotyped at ~700,000 polymorphic loci. We show that at broad scales, the major axis of genetic differentiation in Spain runs from west to east, while there is remarkable genetic similarity in the north-south direction. Our analysis also reveals striking patterns of geographically-localised and subtle population structure within Spain at scales down to tens of kilometres. We developed and applied new approaches to show how this structure has arisen from a complex and regionally-varying mix of genetic isolation and recent gene-flow within and from outside of Iberia. To further explore the genetic impact of historical migrations and invasions of Iberia, we assembled a data set of 2,920 individuals (~300,000 markers) from Iberia and the surrounding regions of north Africa, Europe, and sub-Saharan Africa. Our admixture analysis implies that north African-like DNA in Iberia was mainly introduced in the earlier half (860 - 1120 CE) of the period of Muslim rule in Iberia, and we estimate that the closest modern-day equivalents to the initial migrants are located in Western Sahara. We also find that north African-like DNA in Iberia shows striking regional variation, with near-zero contributions in the Basque regions, low amounts (~3%) in the north east of Iberia, and as high as (~11%) in Galicia and Portugal. The UK Biobank project is a large prospective cohort study of ~500,000 individuals from across the United Kingdom, aged between 40-69 at recruitment. A rich variety of phenotypic and health-related information is available on each participant, making the resource unprecedented in its size and scope. Understanding the role that genetics plays in phenotypic variation, and its potential interactions with other factors, provides a critical route to a better understanding of human biology and population health. As such, a key component of the UK Biobank resource has been the collection of genome-wide genetic data (~805,000 markers) on every participant using purpose-designed genotyping arrays. These data are the focus of the second part of this thesis. In particular, we designed and implemented a quality control (QC) pipeline on behalf of the current and future use of this multi-purpose resource. Genotype data on this scale offers novel opportunities for assessing quality issues, although the wide range of ancestral backgrounds in the cohort also creates particular challenges. We also conducted a set of analyses that reveal properties of the genetic data, including population structure and familial relatedness, that can be important for downstream analyses. We find that cryptic relatedness is common among UK Biobank participants (~30% have at least one first cousin relative or closer), and a full range of human population structure is present in this cohort: from world-wide ancestral diversity to subtle population structure at sub-national geographic scales. Finally, we performed a genome-wide association scan on a well-studied and highly polygenic phenotype: standing height. This provided a further test of the effectiveness of our QC, as well as highlighting the potential of the resource to uncover novel regions of association.
APA, Harvard, Vancouver, ISO, and other styles
42

Alsahli, Mohamad, and Hamadou Kandeh. "Effect of Big Data Analytics on Audit : An exploratory qualitative study of data analytics on auditors’ skills and competence, perception of professional judgment, audit efficiency and audit quality." Thesis, Umeå universitet, Företagsekonomi, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-172302.

Full text
Abstract:
Abstract Purpose: The primary goal of this thesis is to provide a deeper understanding of how big data affect professional judgment, audit efficiency, and perceived audit quality. It also aims to explore the effect of Big Data Analytics (BDA) on the skills and competence required by auditors to perform an audit in a big data environment. Theoretical perspectives: Theoretical concepts base on previous research and publications by practitioners and regulators on BDA, professional judgment, audit efficiency, and audit quality. Literature was used to derive the research gap and research questions. Methodology: A qualitative method base exploratory approach. A literature review was conducted to uncover areas of interest that require more research. The effect of data analytics on the audit was identified as a potential area for research; a focus on audit quality was chosen, including key factors that contribute to overall audit quality. The research is based on semi-structured interviews with auditors from big four audit firms in Sweden. Empirical foundation: Empirical evidence was generated through an interview with seven auditors at different levels of the professional hierarchy. Empirical data was analyzed using a thematic data analysis approach. Conclusions: The findings of this research show that using BDA in the audit methodology affect the required skills and competence by auditors to carry out audit engagement activities. More IT related skills and knowledge gaining prominent in the audit field. Implementing data analytics will not be efficient in the early stage but will save time as auditors become more familiar with the tools. Data analytics improve audit quality. Auditors use analytics to gain more insight into the client’s business and communicate such insights to clients. It was found that data analytics generate fact-based audit evidence. The visualization ability enables auditors to visualize and analyze audit evidence to guide their professional judgment and decision making. Key words: Big data, Data analytics, Auditors skills and competence, Audit process, Audit efficiency, Audit quality and Professional judgment.
APA, Harvard, Vancouver, ISO, and other styles
43

Cronholm, Jacob, and Elin Didriksson. "Partners påverkan på revisionskvalité: En andelsfråga? : En studie om hur partnerandelen i de fyra största revisionsbyråerna i Sverige influerar organisationen och på så vis påverkar revisionskvalitén." Thesis, Linnéuniversitetet, Institutionen för ekonomistyrning och logistik (ELO), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-96838.

Full text
Abstract:
Bakgrund: Utvecklingen av revisorsprofessionen kan härledas ända till 1895. En utveckling som grundas i professionalitet men som på senare år blivit allt mer kommersialiserad. Revisionsbyråernas ägare benämns som partners och likt andra organisationstyper har ägarna en stor påverkan på organisationen. Partners skiljer sig dock från många andra ägare i stora företag med sin aktiva närvaro i det dagliga arbetet. Positionen som partner innebär bl.a. makt, status och finansiella fördelar. Incitamenten med partnerskapet bör variera beroende på partnerandel, vilket kan tänkas leda till en dominans av antingen professionell eller kommersiell logik. Syfte: Studiens syfte är att undersöka om partnerandelen påverkar revisionskvalitén inom de fyra största revisionsbyråerna i Sverige. Metod: För att undersöka studiens syfte har en hypotes skapats utifrån agentteorin samt kommersiell och professionell logik. Studien utförs med hjälp av en tvärsnittsdesign och en deduktiv ansats. Studien har utförts på ett urval som består av de fyra största revisionsbyråerna i Sverige och deras klienter på Stockholmsbörsens tre största listor för år 2012 och 2018. Slutsatser: Studien kan från analys och diskussion komma till en slutsats att partnerandelen har en påverkan på revisionskvalitén. Med studien kan det utläsas att auktoriserade revisorer per partner har en positiv relation till revisionskvalitén
Background: The development of the auditor profession originates from 1895. However, the development that is based on professionalism has in recent years become increasingly commercialized. The audit firm’s owners are called partners and like other types of organizations, the owners have a big impact on the organization. However, partners differ from other owners with their presence in the daily business. The position as a partner comes with power, status and financial advantages, which may lead to a dominance of either professional or commercial logic. Purpose: The purpose of the study is to explore whether the partner share affects the quality of auditing within the four biggest audit firms in Sweden. Method: To fulfill the purpose of the study, a hypothesis has been formulated with agency theory and commercial and professional logics. This study uses a crosssectional design with a deductive approach. The selection of data includes the four biggest audit firms in Sweden and their clients on the three biggest lists in the Swedish stock market during 2012 and 2018. Conclusions: The study can conclude from analysis and discussion that the partner share has an impact on audit quality. The study shows that authorized auditors per partner have a positive relationship to audit quality
APA, Harvard, Vancouver, ISO, and other styles
44

Erkki, Robert, and Philip Johnsson. "Quality Data Management in the Next Industrial Revolution : A Study of Prerequisites for Industry 4.0 at GKN Aerospace Sweden." Thesis, Luleå tekniska universitet, Institutionen för ekonomi, teknik och samhälle, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:ltu:diva-69341.

Full text
Abstract:
The so-called Industry 4.0 is by its agitators commonly denoted as the fourth industrial revolution and promises to turn the manufacturing sector on its head. However, everything that glimmers is not gold and in the backwash of hefty consultant fees questions arises: What are the drivers behind Industry 4.0? Which barriers exists? How does one prepare its manufacturing procedures in anticipation of the (if ever) coming era? What is the internet of things and what file sizes’ is characterised as big data? To answer these questions, this thesis aims to resolve the ambiguity surrounding the definitions of Industry 4.0, as well as clarify the fuzziness of a data-driven manufacturing approach. Ergo, the comprehensive usage of data, including collection and storage, quality control, and analysis. In order to do so, this thesis was carried out as a case study at GKN Aerospace Sweden (GAS). Through interviews and observations, as well as a literature review of the subject, the thesis examined different process’ data-driven needs from a quality management perspective. The findings of this thesis show that the collection of quality data at GAS is mainly concerned with explicitly stated customer requirements. As such, the data available for the examined processes is proven inadequate for multivariate analytics. The transition towards a data-driven state of manufacturing involves a five-stage process wherein data collection through sensors is seen as a key enabler for multivariate analytics and a deepened process knowledge. Together, these efforts form the prerequisites for Industry 4.0. In order to effectively start transition towards Industry 4.0, near-time recommendations for GAS includes: capture all data, with emphasize on process data; improve the accessibility of data; and ultimately taking advantage of advanced analytics. Collectively, these undertakings pave the way for the actual improvements of Industry 4.0, such as digital twins, machine cognition, and process self-optimization. Finally, due to the delimitations of the case study, the findings are but generalized for companies with similar characteristics, i.e. complex processes with low volumes.
APA, Harvard, Vancouver, ISO, and other styles
45

Kirchgessner, Martin. "Fouille et classement d'ensembles fermés dans des données transactionnelles de grande échelle." Thesis, Université Grenoble Alpes (ComUE), 2016. http://www.theses.fr/2016GREAM060/document.

Full text
Abstract:
Les algorithmes actuels pour la fouille d’ensembles fréquents sont dépassés par l’augmentation des volumes de données. Dans cette thèse nous nous intéressons plus particulièrement aux données transactionnelles (des collections d’ensembles d’objets, par exemple des tickets de caisse) qui contiennent au moins un million de transactions portant sur au moins des centaines de milliers d’objets. Les jeux de données de cette taille suivent généralement une distribution dite en "longue traine": alors que quelques objets sont très fréquents, la plupart sont rares. Ces distributions sont le plus souvent tronquées par les algorithmes de fouille d’ensembles fréquents, dont les résultats ne portent que sur une infime partie des objets disponibles (les plus fréquents). Les méthodes existantes ne permettent donc pas de découvrir des associations concises et pertinentes au sein d’un grand jeu de données. Nous proposons donc une nouvelle sémantique, plus intuitive pour l’analyste: parcourir les associations par objet, au plus une centaine à la fois, et ce pour chaque objet présent dans les données.Afin de parvenir à couvrir tous les objets, notre première contribution consiste à définir la fouille centrée sur les objets. Cela consiste à calculer, pour chaque objet trouvé dans les données, les k ensembles d’objets les plus fréquents qui le contiennent. Nous présentons un algorithme effectuant ce calcul, TopPI. Nous montrons que TopPI calcule efficacement des résultats intéressants sur nos jeux de données. Il est plus performant que des solutions naives ou des émulations reposant sur des algorithms existants, aussi bien en termes de rapidité que de complétude des résultats. Nous décrivons et expérimentons deux versions parallèles de TopPI (l’une sur des machines multi-coeurs, l’autre sur des grappes Hadoop) qui permettent d’accélerer le calcul à grande échelle.Notre seconde contribution est CAPA, un système permettant d’étudier quelle mesure de qualité des règles d’association serait la plus appropriée pour trier nos résultats. Cela s’applique aussi bien aux résultats issus de TopPI que de jLCM, notre implémentation d’un algorithme récent de fouille d’ensembles fréquents fermés (LCM). Notre étude quantitative montre que les 39 mesures que nous comparons peuvent être regroupées en 5 familles, d’après la similarité des classements de règles qu’elles produisent. Nous invitons aussi des experts en marketing à participer à une étude qualitative, afin de déterminer laquelle des 5 familles que nous proposons met en avant les associations d’objets les plus pertinentes dans leur domaine.Notre collaboration avec Intermarché, partenaire industriel dans le cadre du projet Datalyse, nous permet de présenter des expériences complètes et portant sur des données réelles issues de supermarchés dans toute la France. Nous décrivons un flux d’analyse complet, à même de répondre à cette application. Nous présentons également des expériences portant sur des données issues d’Internet; grâce à la généricité du modèle des ensembles d’objets, nos contributions peuvent s’appliquer dans d’autres domaines.Nos contributions permettent donc aux analystes de découvrir des associations d’objets au milieu de grandes masses de données. Nos travaux ouvrent aussi la voie vers la fouille d’associations interactive à large échelle, afin d’analyser des données hautement dynamiques ou de réduire la portion du fichier à analyser à celle qui intéresse le plus l’analyste
The recent increase of data volumes raises new challenges for itemset mining algorithms. In this thesis, we focus on transactional datasets (collections of items sets, for example supermarket tickets) containing at least a million transactions over hundreds of thousands items. These datasets usually follow a "long tail" distribution: a few items are very frequent, and most items appear rarely. Such distributions are often truncated by existing itemset mining algorithms, whose results concern only a very small portion of the available items (the most frequents, usually). Thus, existing methods fail to concisely provide relevant insights on large datasets. We therefore introduce a new semantics which is more intuitive for the analyst: browsing associations per item, for any item, and less than a hundred associations at once.To address the items' coverage challenge, our first contribution is the item-centric mining problem. It consists in computing, for each item in the dataset, the k most frequent closed itemsets containing this item. We present an algorithm to solve it, TopPI. We show that TopPI computes efficiently interesting results over our datasets, outperforming simpler solutions or emulations based on existing algorithms, both in terms of run-time and result completeness. We also show and empirically validate how TopPI can be parallelized, on multi-core machines and on Hadoop clusters, in order to speed-up computation on large scale datasets.Our second contribution is CAPA, a framework allowing us to study which existing measures of association rules' quality are relevant to rank results. This concerns results obtained from TopPI or from jLCM, our implementation of a state-of-the-art frequent closed itemsets mining algorithm (LCM). Our quantitative study shows that the 39 quality measures we compare can be grouped into 5 families, based on the similarity of the rankings they produce. We also involve marketing experts in a qualitative study, in order to discover which of the 5 families we propose highlights the most interesting associations for their domain.Our close collaboration with Intermarché, one of our industrial partners in the Datalyse project, allows us to show extensive experiments on real, nation-wide supermarket data. We present a complete analytics workflow addressing this use case. We also experiment on Web data. Our contributions can be relevant in various other fields, thanks to the genericity of transactional datasets.Altogether our contributions allow analysts to discover associations of interest in modern datasets. We pave the way for a more reactive discovery of items' associations in large-scale datasets, whether on highly dynamic data or for interactive exploration systems
APA, Harvard, Vancouver, ISO, and other styles
46

Johnsen, Sofia, and Sarah Felldin. "Improving Knowledge of Truck Fuel Consumption Using Data Analysis." Thesis, Linköpings universitet, Reglerteknik, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-130047.

Full text
Abstract:
The large potential of big data and how it has brought value into various industries have been established in research. Since big data has such large potential if handled and analyzed in the right way, revealing information to support decision making in an organization, this thesis is conducted as a case study at an automotive manufacturer with access to large amounts of customer usage data of their vehicles. The reason for performing an analysis of this kind of data is based on the cornerstones of Total Quality Management with the end objective of increasing customer satisfaction of the concerned products or services. The case study includes a data analysis exploring how and if patterns about what affects fuel consumption can be revealed from aggregated customer usage data of trucks linked to truck applications. Based on the case study, conclusions are drawn about how a company can use this type of analysis as well as how to handle the data in order to turn it into business value. The data analysis reveals properties describing truck usage using Factor Analysis and Principal Component Analysis. Especially one property is concluded to be important as it appears in the result of both techniques. Based on these properties the trucks are clustered using k-means and Hierarchical Clustering which shows groups of trucks where the importance of the properties varies. Due to the homogeneity and complexity of the chosen data, the clusters of trucks cannot be linked to truck applications. This would require data that is more easily interpretable. Finally, the importance for fuel consumption in the clusters is explored using model estimation. A comparison of Principal Component Regression (PCR) and the two regularization techniques Lasso and Elastic Net is made. PCR results in poor models difficult to evaluate. The two regularization techniques however outperform PCR, both giving a higher and very similar explained variance. The three techniques do not show obvious similarities in the models and no conclusions can therefore be drawn concerning what is important for fuel consumption. During the data analysis many problems with the data are discovered, which are linked to managerial and technical issues of big data. This leads to for example that some of the parameters interesting for the analysis cannot be used and this is likely to have an impact on the inability to get unanimous results in the model estimations. It is also concluded that the data was not originally intended for this type of analysis of large populations, but rather for testing and engineering purposes. Nevertheless, this type of data still contains valuable information and can be used if managed in the right way. From the case study it can be concluded that in order to use the data for more advanced analysis a big-data plan is needed at a strategic level in the organization. The plan summarizes the suggested solution for the managerial issues of the big data for the organization. This plan describes how to handle the data, how the analytic models revealing the information should be designed and the tools and organizational capabilities needed to support the people using the information.
APA, Harvard, Vancouver, ISO, and other styles
47

Fransson, Oliver, and Simon Sleman. "The absolution of non-audit services – unravelling a nexus of research : A quantitative study of non-audit services’ impact on financial reporting quality among private firms in Sweden." Thesis, Internationella Handelshögskolan, Jönköping University, IHH, Företagsekonomi, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:hj:diva-48577.

Full text
Abstract:
Non-audit services provided by audit firms have been a popular scientific topic within the fields of audit and accounting research over the past decades. Numerous researchers have attempted to provide a theoretical contribution by examining different ways of measuring the concepts of audit quality and financial reporting quality. The resulting consequences are mixed results and a lack of consensus among researchers from both research fields. The two, in other situations, rather distinctive research fields of audit quality and financial reporting quality, have, in several cases, been confounded without analytical reflection regarding their differences. In parallel to the scientific progress, regulatory bodies have noticed the increasing trend of non-audit services and how they constitute larger portions of the audit firm’s annual revenues. Their responses have been legal restrictions, both in the US and Europe, in order to cease the trend. The purpose of this thesis is to make a pronounced investigation regarding the relationship between non-audit services and financial reporting quality in Swedish private firms. Furthermore, it will also be of interest to examine if this proposed relationship is moderated by the presence of the four global market-leading audit firms or not. The study is based on a deductive approach and a quantitative research strategy, to collect and analyze data from annual reports. To fulfill the purpose of the study, the data is analyzed by conducting binary and multinomial logistic regression tests. The results suggest that there is an association between certain types of non-audit services and financial reporting quality. Specifically, services that are unrelated to tax have proven to be statistically significant positively correlated with financial reporting quality. No evidence was found supporting a moderating effect by the characteristics of audit firms, suggesting that the choice of an audit firm is irrelevant for attaining high financial reporting quality when purchasing non-audit services. The study’s theoretical contribution is the novelty arising from the combination of studying non-audit services’ impact on financial reporting quality within a Swedish setting on private firms. The study also provides empirical contribution by using a proxy for financial reporting quality rarely used in previous research. The findings are of practical importance since they suggest that firms potentially benefit in their financial reporting by purchasing these kinds of services, which contradicts past actions made by regulatory bodies.
APA, Harvard, Vancouver, ISO, and other styles
48

Zaidi, Houda. "Amélioration de la qualité des données : correction sémantique des anomalies inter-colonnes." Thesis, Paris, CNAM, 2017. http://www.theses.fr/2017CNAM1094/document.

Full text
Abstract:
La qualité des données présente un grand enjeu au sein d'une organisation et influe énormément sur la qualité de ses services et sur sa rentabilité. La présence de données erronées engendre donc des préoccupations importantes autour de cette qualité. Ce rapport traite la problématique de l'amélioration de la qualité des données dans les grosses masses de données. Notre approche consiste à aider l'utilisateur afin de mieux comprendre les schémas des données manipulées, mais aussi définir les actions à réaliser sur celles-ci. Nous abordons plusieurs concepts tels que les anomalies des données au sein d'une même colonne, et les anomalies entre les colonnes relatives aux dépendances fonctionnelles. Nous proposons dans ce contexte plusieurs moyens de pallier ces défauts en nous intéressons à la performance des traitements ainsi opérés
Data quality represents a major challenge because the cost of anomalies can be very high especially for large databases in enterprises that need to exchange information between systems and integrate large amounts of data. Decision making using erroneous data has a bad influence on the activities of organizations. Quantity of data continues to increase as well as the risks of anomalies. The automatic correction of these anomalies is a topic that is becoming more important both in business and in the academic world. In this report, we propose an approach to better understand the semantics and the structure of the data. Our approach helps to correct automatically the intra-column anomalies and the inter-columns ones. We aim to improve the quality of data by processing the null values and the semantic dependencies between columns
APA, Harvard, Vancouver, ISO, and other styles
49

Varotto, Luís Fernando. "Franchisor-franchisee relationship and performance: influence of personality traits, entrepreneurial drive, and time of relationship." reponame:Repositório Institucional do FGV, 2015. http://hdl.handle.net/10438/13601.

Full text
Abstract:
Submitted by Luís Fernando Varotto (lvarotto@ig.com.br) on 2015-03-30T14:56:01Z No. of bitstreams: 1 TESE Luís F Varotto - Franchisor- franchisee relationship and performance.pdf: 5577752 bytes, checksum: c2c0a5c8e14cab676f16eb65bfeaa9b3 (MD5)
Approved for entry into archive by PAMELA BELTRAN TONSA (pamela.tonsa@fgv.br) on 2015-03-30T14:56:46Z (GMT) No. of bitstreams: 1 TESE Luís F Varotto - Franchisor- franchisee relationship and performance.pdf: 5577752 bytes, checksum: c2c0a5c8e14cab676f16eb65bfeaa9b3 (MD5)
Made available in DSpace on 2015-03-30T14:57:27Z (GMT). No. of bitstreams: 1 TESE Luís F Varotto - Franchisor- franchisee relationship and performance.pdf: 5577752 bytes, checksum: c2c0a5c8e14cab676f16eb65bfeaa9b3 (MD5) Previous issue date: 2015-02-27
Literature in franchise has virtually ignored the role of psychological aspects on firm interorganizational results, despite its influence on firm level results and relationship quality. Therefore, the present study aims to examine the influence of franchisees’ personality and entrepreneurial drive on franchisor-franchisee relationship quality and financial performance over time. The study also investigated the role of the time of relationship on the relationship quality and financial performance. This study used a self-report survey conducted by mail to collect data from a sample of 342 franchisees selected from 3 franchise networks. Personality was represented by the Big-Five personality traits (IPIP-B5 scales): extraversion, agreeableness, conscientiousness, emotional stability, and imagination. Entrepreneur drive was represented by the Carland Entrepreneurship Index (CEI). Relationship quality was conceptualized through a 23-item second-order construct (incorporating trust, commitment, and relationship satisfaction), while financial performance was represented by using a scale measuring sales growth and profitability. Time of relationship was measured by the months of relationship between franchisee and franchisor. A Partial Least Squares (PLS) structural equation model, mean analysis, and regression analysis were conducted to test the hypothesized relationships. Three of the five personality dimensions produced the predicted effect on the outcome variables of relationship quality – agreeableness (positively), emotional stability (positively), and imagination (positively). Financial performance was affected as predicted by conscientiousness (positively), emotional stability (positively), and imagination (positively). As expected, relationship quality presented a positive and significant effect on financial performance. Entrepreneurial drive showed the predicted positive effect only on performance. Time of relationship presented the positive predicted effect on the franchisor-franchisee relationship as regards relationship quality and financial performance; however, the hypothesized shape of the relationship phases could only partially be confirmed, since only between two phases (routine and stabilization) mean analysis showed significant differences. Results indicate that personality does in fact influence relationship quality and performance, but the manner in which this occurs differs from the Brazilian context where this research was conducted to the findings of research conducted in Australia, suggesting that factors such as culture and market stability may have influence on the relationship between personality traits and both relationship quality and financial performance. Entrepreneurial drive appears to positively influence franchisee performance, but its influence proved not to produce a significant impact on relationship quality. The present study’s results also indicate the importance of the time of relationship needed to foster relationship quality and performance. Moreover, long-term relationships are related to better franchisee relationship quality and financial performance assessments. Limitations of this work and suggestions for future studies are also discussed.
A literatura em franchising tem virtualmente ignorado o papel de aspectos psicologicos nos resultados interorganizacionais das empresas, a despeito de sua influencia nos resultados das organizações e da qualidade de relacionamento. Este estudo, portanto, tem por objetivo analisar a influência da personalidade e do potencial empreendedor na qualidade de relacionamento e desempenho financeiro na relação franqueador-franqueado, ao longo do tempo, sob a perspectiva dos franqueados. Este estudo analisa também o papel do tempo de relacionamento sobre a qualidade de relacionamento e o desempenho financeiro. Foi utilizado neste estudo um questionário de auto-preenchimento, enviado por e-mail, com o objetivo de recolher dados de uma amostra de 342 franqueados de 3 redes de franquias. A personalidade foi mensurada por meio dos “Cinco Grandes” traços de personalidade (escalas IPIP-B5): extroversão, agradabilidade, consciencia, estabilidade emocional e imaginação. O potencial empreendedor foi mensurado por meio do índice CEI (Carland Entrepreneurship Index). A qualidade do relacionamento foi estruturada como um constructo de segunda ordem, composto por 23 itens (incorporando confiança, comprometimento e satisfação com o relacionamento), e o desempenho financeiro foi representado por meio de uma escala de mensuração de crescimento de vendas e de rentabilidade. O tempo de relacionamento foi medido por meio dos meses de relacionamento entre franqueado e franqueador. As hipoteses foram testadas por meio de modelagem por equações estruturais, com a utilização do método de mínimos quadrados parciais (PLS), análise de regressão e análise de médias. Três das cinco dimensões da personalidade apresentaram o efeito previsto sobre as variáveis qualidade do relacionamento – agradabilidade (positivamente), estabilidade emocional (positivamente), e imaginação (positivamente). O desempenho financeiro foi influenciado, como previsto por consciência (positivamente), estabilidade emocional (positivamente), e imaginação (positivamente). Como esperado, a qualidade do relacionamento apresentou efeito positivo e significativo em relação ao desempenho financeiro. O potencial empreendedor apresentou o efeito positivo previsto apenas sobre desempenho. O tempo de relacionamento teve o efeito positivo esperado sobre o relacionamento franqueador-franqueado, em relação à qualidade do relacionamento e o desempenho financeiro, mas as diferenças entre as fases de relacionamento propostas foram apenas parcialmente confirmadas, uma vez que em somente duas fases (rotina e estabilização) a análise de médias mostrou diferenças significativas. Os resultados indicam que a personalidade influencia a qualidade de relacionamento e o desempenho, mas a meneira pela qual isso ocorre é diferente no contexto brasileiro, onde esta pesquisa foi realizada, dos achados da pesquisa conduzida na Austrália, sugerindo que fatores como cultura e estabilidade de mercado podem ter influencia sobre a relação entre traços de personalidade e qualidade de relacionamento, e traços de personalidade e desempenho financeiro. O potencial empreendedor parece influenciar positivamente o desempenho do franqueado, mas a sua influência não foi significativa em relação à qualidade do relacionamento. Os resultados também indicam a importância do tempo no desenvolvimento da qualidade de relacionamento e desempenho. Além disso, os relacionamentos de longo prazo estão relacionados a melhores avaliações de qualidade de relacionamento e desempenho financeiros por parte dos franqueados. As limitações do trabalho e sugestões para estudos futuros também são discutidos.
APA, Harvard, Vancouver, ISO, and other styles
50

NASCIMENTO, FILHO Dimas Cassimiro do. "Reduzindo custos da deduplicação de dados utilizando heurísticas e computação em nuvem." Universidade Federal de Campina Grande, 2017. http://dspace.sti.ufcg.edu.br:8080/jspui/handle/riufcg/559.

Full text
Abstract:
Submitted by Lucienne Costa (lucienneferreira@ufcg.edu.br) on 2018-05-02T21:20:23Z No. of bitstreams: 1 DIMAS CASSIMIRO DO NASCIMENTO FILHO – TESE (PPGCC) 2017.pdf: 1879329 bytes, checksum: bda72914ec66d17611d9d0ab5b9ec6d5 (MD5)
Made available in DSpace on 2018-05-02T21:20:23Z (GMT). No. of bitstreams: 1 DIMAS CASSIMIRO DO NASCIMENTO FILHO – TESE (PPGCC) 2017.pdf: 1879329 bytes, checksum: bda72914ec66d17611d9d0ab5b9ec6d5 (MD5) Previous issue date: 2017-11-10
Na era de Big Data, na qual a escala dos dados provê inúmeros desafios para algoritmos clássicos, a tarefa de avaliar a qualidade dos dados pode se tornar custosa e apresentar tempos de execução elevados. Por este motivo, gerentes de negócio podem optar por terceirizar o monitoramento da qualidade de bancos de dados para um serviço específico, usualmente baseado em computação em nuvem. Neste contexto, este trabalho propõe abordagens para redução de custos da tarefa de deduplicação de dados, a qual visa detectar entidades duplicadas em bases de dados, no contexto de um serviço de qualidade de dados em nuvem. O trabalho tem como foco a tarefa de deduplicação de dados devido a sua importância em diversos contextos e sua elevada complexidade. É proposta a arquitetura em alto nível de um serviço de monitoramento de qualidade de dados que emprega o provisionamento dinâmico de recursos computacionais por meio da utilização de heurísticas e técnicas de aprendizado de máquina. Além disso, são propostas abordagens para a adoção de algoritmos incrementais de deduplicação de dados e controle do tamanho de blocos gerados na etapa de indexação do problema investigado. Foram conduzidos quatro experimentos diferentes visando avaliar a eficácia dos algoritmos de provisionamento de recursos propostos e das heurísticas empregadas no contexto de algoritmos incrementais de deduplicação de dados e de controle de tamanho dos blocos. Os resultados dos experimentos apresentam uma gama de opções englobando diferentes relações de custo e benefício, envolvendo principalmente: custo de infraestrutura do serviço e quantidade de violações de SLA ao longo do tempo. Outrossim, a avaliação empírica das heurísticas propostas para o problema de deduplicação incremental de dados também apresentou uma série de padrões nos resultados, envolvendo principalmente o tempo de execução das heurísticas e os resultados de eficácia produzidos. Por fim, foram avaliadas diversas heurísticas para controlar o tamanho dos blocos produzidos em uma tarefa de deduplicação de dados, cujos resultados de eficácia são bastante influenciados pelos valores dos parâmetros empregados. Além disso, as heurísticas apresentaram resultados de eficiência que variam significativamente, dependendo da estratégia de poda de blocos adotada. Os resultados dos quatro experimentos conduzidos apresentam suporte para demonstrar que diferentes estratégias (associadas ao provisionamento de recursos computacionais e aos algoritmos de qualidade de dados) adotadas por um serviço de qualidade de dados podem influenciar significativamente nos custos do serviço e, consequentemente, os custos repassados aos usuários do serviço.
In the era of Big Data, in which the scale of the data provides many challenges for classical algorithms, the task of assessing the quality of datasets may become costly and complex. For this reason, business managers may opt to outsource the data quality monitoring for a specific cloud service for this purpose. In this context, this work proposes approaches for reducing the costs generated from solutions for the data deduplication problem, which aims to detect duplicate entities in datasets, in the context of a service for data quality monitoring. This work investigates the deduplication task due to its importance in a variety of contexts and its high complexity. We propose a high-level architecture of a service for data quality monitoring, which employs provisioning algorithms that use heuristics and machine learning techniques. Furthermore, we propose approaches for the adoption of incremental data quality algorithms and heuristics for controlling the size of the blocks produced in the indexing phase of the investigated problem. Four different experiments have been conducted to evaluate the effectiveness of the proposed provisioning algorithms, the heuristics for incremental record linkage and the heuristics to control block sizes for entity resolution. The results of the experiments show a range of options covering different tradeoffs, which involves: infrastructure costs of the service and the amount of SLA violations over time. In turn, the empirical evaluation of the proposed heuristics for incremental record linkage also presented a number of patterns in the results, which involves tradeoffs between the runtime of the heuristics and the obtained efficacy results. Lastly, the evaluation of the heuristics proposed to control block sizes have presented a large number of tradeoffs regarding execution time, amount of pruning approaches and the obtained efficacy results. Besides, the efficiency results of these heuristics may vary significantly, depending of the adopted pruning strategy. The results from the conducted experiments support the fact that different approaches (associated with cloud computing provisioning and the employed data quality algorithms) adopted by a data quality service may produce significant influence over the generated service costs, and thus, the final costs forwarded to the service customers.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography