To see the other types of publications on this topic, follow the link: Benchmark.

Journal articles on the topic 'Benchmark'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 journal articles for your research on the topic 'Benchmark.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse journal articles on a wide variety of disciplines and organise your bibliography correctly.

1

Dao, Loc Xuan, and Hung Manh Chu. "ESTIMATION OF BENCHMARK'S STABILITY BY ADJUSTMENT OF FREE GEODETIC ELEVATION BASE NETWORK." Science and Technology Development Journal 12, no. 18 (December 15, 2009): 69–75. http://dx.doi.org/10.32508/stdj.v12i18.2385.

Full text
Abstract:
The paper presents a procedure to estimate benchmark's stability by adjustment of free geodetic elevation base network. The changes of adjusted heigh by software " CBSCI" for free networks consisted 3,4,5 benchmarks are 0.1-0.2 mm when to compare this method with a method one benchmark is unchanged in height and method the mean of heigh for benchmark group is unchanged. So, we can use this method for estimation of benchmark's stability when to measure settlement for engineering constructions.
APA, Harvard, Vancouver, ISO, and other styles
2

Blume-Kohout, Robin, and Kevin C. Young. "A volumetric framework for quantum computer benchmarks." Quantum 4 (November 15, 2020): 362. http://dx.doi.org/10.22331/q-2020-11-15-362.

Full text
Abstract:
We propose a very large family of benchmarks for probing the performance of quantum computers. We call them volumetric benchmarks (VBs) because they generalize IBM's benchmark for measuring quantum volume \cite{Cross18}. The quantum volume benchmark defines a family of square circuits whose depth d and width w are the same. A volumetric benchmark defines a family of rectangular quantum circuits, for which d and w are uncoupled to allow the study of time/space performance trade-offs. Each VB defines a mapping from circuit shapes — (w,d) pairs — to test suites C(w,d). A test suite is an ensemble of test circuits that share a common structure. The test suite C for a given circuit shape may be a single circuit C, a specific list of circuits {C1…CN} that must all be run, or a large set of possible circuits equipped with a distribution Pr(C). The circuits in a given VB share a structure, which is limited only by designers' creativity. We list some known benchmarks, and other circuit families, that fit into the VB framework: several families of random circuits, periodic circuits, and algorithm-inspired circuits. The last ingredient defining a benchmark is a success criterion that defines when a processor is judged to have ``passed'' a given test circuit. We discuss several options. Benchmark data can be analyzed in many ways to extract many properties, but we propose a simple, universal graphical summary of results that illustrates the Pareto frontier of the d vs w trade-off for the processor being benchmarked.
APA, Harvard, Vancouver, ISO, and other styles
3

Adibekyan, V., S. G. Sousa, N. C. Santos, P. Figueira, C. Allende Prieto, E. Delgado Mena, J. I. González Hernández, et al. "Benchmark stars, benchmark spectrographs." Astronomy & Astrophysics 642 (October 2020): A182. http://dx.doi.org/10.1051/0004-6361/202038793.

Full text
Abstract:
Context. Gaia benchmark stars are selected to be calibration stars for different spectroscopic surveys. Very high-quality and homogeneous spectroscopic data for these stars are therefore required. We collected ultrahigh-resolution ESPRESSO spectra for 30 of the 34 Gaia benchmark stars and made them public. Aims. We quantify the consistency of the results that are obtained with different high- (R ~ 115 000), and ultrahigh- (R ~ 220 000) resolution spectrographs. We also comprehensively studied the effect of using different spectral reduction products of ESPRESSO on the final spectroscopic results. Methods. We used ultrahigh- and high-resolution spectra obtained with the ESPRESSO, PEPSI, and HARPS spectrographs to measure spectral line characteristics (line depth; line width; and equivalent width, EW) and determined stellar parameters and abundances for a subset of 11 Gaia benchmark stars. We used the ARES code for automatic measurements of the spectral line parameters. Results. Our measurements reveal that the same individual spectral lines measured from adjacent 2D (spectrum in the wavelength-order space) echelle orders of ESPRESSO spectra differ slightly in line depth and line width. When a long list of spectral lines is considered, the EW measurements based on the 2D and 1D (the final spectral product) ESPRESSO spectra agree very well. The EW spectral line measurements based on the ESPRESSO, PEPSI, and HARPS spectra also agree to within a few percent. However, we note that the lines appear deeper in the ESPRESSO spectra than in PEPSI and HARPS. The stellar parameters derived from each spectrograph by combining the several available spectra agree well overall. Conclusions. We conclude that the ESPRESSO, PEPSI, and HARPS spectrographs can deliver spectroscopic results that are sufficiently consistent for most of the science cases in stellar spectroscopy. However, we found small but important differences in the performance of the three spectrographs that can be crucial for specific science cases.
APA, Harvard, Vancouver, ISO, and other styles
4

Papiani, Mark, Anthony J. G. Hey, and Roger W. Hockney. "The Graphical Benchmark Information Service." Scientific Programming 4, no. 4 (1995): 219–27. http://dx.doi.org/10.1155/1995/270294.

Full text
Abstract:
Unlike single-processor benchmarks, multiprocessor benchmarks can yield tens of numbers for each benchmark on each computer, as factors such as the number of processors and problem size are varied. A graphical display of performance surfaces therefore provides a satisfactory way of comparing results. The University of Southampton has developed the Graphical Benchmark Information Service (GBIS) on the World Wide Web (WWW) to display interactively graphs of user-selected benchmark results from the GENESIS and PARKBENCH benchmark suites.
APA, Harvard, Vancouver, ISO, and other styles
5

Dieu Dang, Huong. "KiwiSaver fund performance and asset allocation policy." Pacific Accounting Review 31, no. 2 (April 1, 2019): 232–57. http://dx.doi.org/10.1108/par-06-2018-0044.

Full text
Abstract:
Purpose This paper aims to examine the performance and benchmark asset allocation policy of 70 KiwiSaver funds catergorised as growth, balanced or conservative over the period October 2007-June 2016. The study focuses on the sources for returns variability across time and returns variation among funds. Design/methodology/approach Each fund is benchmarked against a portfolio of eight indices representing eight invested asset classes. Three measures were used to examine the after-fee benchmark-adjusted performance of each fund: excess return, cumulative abnormal return and holding period returns difference. Tracking error and active share were used to capture manager’s benchmark deviation. Findings On average, funds underperform their respective benchmarks, with the mean quarterly excess return (after management fees) of −0.15 per cent (growth), −0.63 per cent (balanced) and −0.83 per cent (conservative). Benchmark returns variability, on average, explains 43-78 per cent of fund’s across-time returns variability, and this is primarily driven by fund’s exposures to global capital markets. Differences in benchmark policies, on average, account for 18.8-39.3 per cent of among-fund returns variation, while differences in fees and security selection may explain the rest. About 61 per cent of balanced and 47 per cent of Growth funds’ managers make selection bets against their benchmarks. There is no consistent evidence that more actively managed funds deliver higher after-fee risk-adjusted performance. Superior performance is often due to randomness. Originality/value This study makes use of a unique data set gathered directly from KiwiSaver managers and captures the long-term strategic asset allocation target which underlines the investment management process in reality. The study represents the first attempt to examine the impact of benchmark asset allocation policy on KiwiSaver fund’s returns variability across time and returns variation among funds.
APA, Harvard, Vancouver, ISO, and other styles
6

Szárnyas, Gábor, Jack Waudby, Benjamin A. Steer, Dávid Szakállas, Altan Birler, Mingxi Wu, Yuchen Zhang, and Peter Boncz. "The LDBC Social Network Benchmark." Proceedings of the VLDB Endowment 16, no. 4 (December 2022): 877–90. http://dx.doi.org/10.14778/3574245.3574270.

Full text
Abstract:
The Social Network Benchmark's Business Intelligence workload (SNB BI) is a comprehensive graph OLAP benchmark targeting analytical data systems capable of supporting graph workloads. This paper marks the finalization of almost a decade of research in academia and industry via the Linked Data Benchmark Council (LDBC). SNB BI advances the state-of-the art in synthetic and scalable analytical database benchmarks in many aspects. Its base is a sophisticated data generator, implemented on a scalable distributed infrastructure, that produces a social graph with small-world phenomena, whose value properties follow skewed and correlated distributions and where values correlate with structure. This is a temporal graph where all nodes and edges follow lifespan-based rules with temporal skew enabling realistic and consistent temporal inserts and (recursive) deletes. The query workload exploiting this skew and correlation is based on LDBC's "choke point"-driven design methodology and will entice technical and scientific improvements in future (graph) database systems. SNB BI includes the first adoption of "parameter curation" in an analytical benchmark, a technique that ensures stable runtimes of query variants across different parameter values. Two performance metrics characterize peak single-query performance (power) and sustained concurrent query throughput. To demonstrate the portability of the benchmark, we present experimental results on a relational and a graph DBMS. Note that these do not constitute an official LDBC Benchmark Result - only audited results can use this trademarked term.
APA, Harvard, Vancouver, ISO, and other styles
7

Erdelt, Patrick K., and Jascha Jestel. "DBMS-Benchmarker: Benchmark and Evaluate DBMS in Python." Journal of Open Source Software 7, no. 79 (November 2, 2022): 4628. http://dx.doi.org/10.21105/joss.04628.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Vreven, Thom, Iain H. Moal, Anna Vangone, Brian G. Pierce, Panagiotis L. Kastritis, Mieczyslaw Torchala, Raphael Chaleil, et al. "Updates to the Integrated Protein–Protein Interaction Benchmarks: Docking Benchmark Version 5 and Affinity Benchmark Version 2." Journal of Molecular Biology 427, no. 19 (September 2015): 3031–41. http://dx.doi.org/10.1016/j.jmb.2015.07.016.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Gökçe, Özge, and Veysel Harun Şahin. "ABench2020: A WCET Benchmark Suite in Ada Programming Language." Academic Perspective Procedia 3, no. 1 (October 25, 2020): 509–18. http://dx.doi.org/10.33793/acperpro.03.01.99.

Full text
Abstract:
Validation is an important part of the development process of real-time systems. During validation, it should be proved that, the system meets its timing constraints. Therefore, worst-case execution time (WCET) analysis is performed. Currently, several WCET analysis tools are being developed. Researchers who develop new WCET analysis tools, need benchmarks to evaluate and compare their tools with alternatives. These benchmarks are called WCET benchmarks. In this paper a new WCET benchmark suite named ABench2020 is introduced. Its main focus is to provide benchmark programs in Ada programming language for WCET research. Therefore, ABench2020 includes several benchmark programs which were written in Ada programming language. The benchmark programs implement different program structures and properties to help researchers test their systems from different aspects. ABench2020 was published as open source. It is freely available over the Internet.
APA, Harvard, Vancouver, ISO, and other styles
10

Berry, Michael W., Jack J. Dongarra, Brian H. Larose, and Todd A. Letsche. "PDS: A Performance Database Server." Scientific Programming 3, no. 2 (1994): 147–56. http://dx.doi.org/10.1155/1994/391804.

Full text
Abstract:
The process of gathering, archiving, and distributing computer benchmark data is a cumbersome task usually performed by computer users and vendors with little coordination. Most important, there is no publicly available central depository of performance data for all ranges of machines from personal computers to supercomputers. We present an Internet-accessible performance database server (PDS) that can be used to extract current benchmark data and literature. As an extension to the X-Windows-based user interface (Xnetlib) to the Netlib archival system, PDS provides an on-line catalog of public domain computer benchmarks such as the LINPACK benchmark, Perfect benchmarks, and the NAS parallel benchmarks. PDS does not reformat or present the benchmark data in any way that conflicts with the original methodology of any particular benchmark; it is thereby devoid of any subjective interpretations of machine performance. We believe that all branches (research laboratories, academia, and industry) of the general computing community can use this facility to archive performance metrics and make them readily available to the public. PDS can provide a more manageable approach to the development and support of a large dynamic database of published performance metrics.
APA, Harvard, Vancouver, ISO, and other styles
11

Lee, Ho Min, Donghwi Jung, Ali Sadollah, Do Guen Yoo, and Joong Hoon Kim. "Generation of Benchmark Problems for Optimal Design of Water Distribution Systems." Water 11, no. 8 (August 8, 2019): 1637. http://dx.doi.org/10.3390/w11081637.

Full text
Abstract:
Engineering benchmark problems with specific characteristics have been used to compare the performance and reliability of metaheuristic algorithms, and water distribution system design benchmarks are also widely used. However, only a few benchmark design problems have been considered in the research community. Due to the limited set of previous benchmarks, it is challenging to identify the algorithm with the best performance and the highest reliability among a group of algorithms. Therefore, in this study, a new water distribution system design benchmark problem generation method is proposed considering problem size and complexity modifications of a reference benchmark. The water distribution system design benchmark problems are used for performance and reliability comparison among several reported metaheuristic optimization algorithms. The optimal design results are able to quantify the performance and reliability of the compared algorithms which shows each metaheuristic algorithm has its own strengths and weaknesses. Finally, using the proposed method in this study, guidelines are derived for selecting an appropriate metaheuristic algorithm for water distribution system design.
APA, Harvard, Vancouver, ISO, and other styles
12

Jan Hendriks, Gert. "Stichting Benchmark GGZ creëert een enorme bureaucratie, maar benchmarkt niet!" Psychopraktijk 5, no. 4 (August 2013): 25–26. http://dx.doi.org/10.1007/s13170-013-0056-1.

Full text
APA, Harvard, Vancouver, ISO, and other styles
13

Drexler, Richard, Roman Rotermund, Timothy Smith, John Kilgallon, Jürgen Honegger, Isabella Nasi-Kordhishti, Paul Gardner, et al. "QLTI-09. DEFINING GLOBAL BENCHMARK OUTCOMES FOR TRANSSPHENOIDAL SURGERY OF PITUITARY ADENOMAS: A MULTICENTER ANALYSIS OF 2862 CASES." Neuro-Oncology 24, Supplement_7 (November 1, 2022): vii236. http://dx.doi.org/10.1093/neuonc/noac209.911.

Full text
Abstract:
Abstract Benchmarks are important to measure and aid in improve outcomes for surgical procedures. However, best achievable results that have been validated internationally for transsphenoidal surgery are not available. Therefore, we aimed to establish robust, standardized outcome benchmarks for transsphenoidal surgery of pituitary adenomas. A total of 2862 transsphenoidal tumor resections from 12 high-volume centers in 4 continents were analyzed. Patients were risk stratified and the median values of each center’s outcomes were established. The outcome benchmark was defined as the 75th percentile of all median values for a particular outcome as defined by Staiger et al. Out of 2862 patients, 1201 (41.9%) defined the benchmark cohort. The proportion of benchmark cases contributing to the final cohort ranged across centers between 22.1% to 59.7%. Within the benchmark cases, 928 (73.3%) patients underwent microscopic (MTS) and 263 (21.9%) patients endoscopic endonasal resection (EES). The overall postoperative complication rate was 18.9% with an in-hospital mortality between 0.0-0.8%. Benchmark cutoffs were ≤ 3.3% for reoperation rate, ≤ 4.6% for cerebrospinal fluid leak requiring intervention, and ≤ 15.3% for transient diabetes insipidus. At 6 months follow-up, benchmark cutoffs were calculated as follows: readmission rate: ≤ 7.1%, new hypopituitarism ≤ 15.5%, new neurological deficit ≤ 1.2%, tumor remnant ≤ 25.5%. This analysis defines benchmark values for transsphenoidal resection of pituitary adenomas targeting morbidity, mortality, surgical and tumor-related outcomes. The benchmark cutoffs can be used to assess different centers, patients’ populations, and novel surgical techniques.
APA, Harvard, Vancouver, ISO, and other styles
14

Quetschlich, Nils, Lukas Burgholzer, and Robert Wille. "MQT Bench: Benchmarking Software and Design Automation Tools for Quantum Computing." Quantum 7 (July 20, 2023): 1062. http://dx.doi.org/10.22331/q-2023-07-20-1062.

Full text
Abstract:
Quantum software tools for a wide variety of design tasks on and across different levels of abstraction are crucial in order to eventually realize useful quantum applications. This requires practical and relevant benchmarks for new software tools to be empirically evaluated and compared to the current state of the art. Although benchmarks for specific design tasks are commonly available, the demand for an overarching cross-level benchmark suite has not yet been fully met and there is no mutual consolidation in how quantum software tools are evaluated thus far. In this work, we propose the MQT Bench benchmark suite (as part of the Munich Quantum Toolkit, MQT) based on four core traits: (1) cross-level support for different abstraction levels, (2) accessibility via an easy-to-use web interface (https://www.cda.cit.tum.de/mqtbench/) and a Python package, (3) provision of a broad selection of benchmarks to facilitate generalizability, as well as (4) extendability to future algorithms, gate-sets, and hardware architectures. By comprising more than 70,000 benchmark circuits ranging from 2 to 130 qubits on four abstraction levels, MQT Bench presents a first step towards benchmarking different abstraction levels with a single benchmark suite to increase comparability, reproducibility, and transparency.
APA, Harvard, Vancouver, ISO, and other styles
15

Chambers, Robert D., and Nathanael C. Yoder. "FilterNet: A Many-to-Many Deep Learning Architecture for Time Series Classification." Sensors 20, no. 9 (April 28, 2020): 2498. http://dx.doi.org/10.3390/s20092498.

Full text
Abstract:
In this paper, we present and benchmark FilterNet, a flexible deep learning architecture for time series classification tasks, such as activity recognition via multichannel sensor data. It adapts popular convolutional neural network (CNN) and long short-term memory (LSTM) motifs which have excelled in activity recognition benchmarks, implementing them in a many-to-many architecture to markedly improve frame-by-frame accuracy, event segmentation accuracy, model size, and computational efficiency. We propose several model variants, evaluate them alongside other published models using the Opportunity benchmark dataset, demonstrate the effect of model ensembling and of altering key parameters, and quantify the quality of the models’ segmentation of discrete events. We also offer recommendations for use and suggest potential model extensions. FilterNet advances the state of the art in all measured accuracy and speed metrics when applied to the benchmarked dataset, and it can be extensively customized for other applications.
APA, Harvard, Vancouver, ISO, and other styles
16

Jiang, John (Xuefeng). "Beating Earnings Benchmarks and the Cost of Debt." Accounting Review 83, no. 2 (March 1, 2008): 377–416. http://dx.doi.org/10.2308/accr.2008.83.2.377.

Full text
Abstract:
Prior research documents that firms tend to beat three earnings benchmarks—zero earnings, last year's earnings, and analyst's forecasted earnings—and that there are both equity market and compensation-related benefits associated with beating these benchmarks. This study investigates whether and under what conditions beating these three earnings benchmarks reduces a firm's cost of debt. I use two proxies for a firm's cost of debt: credit ratings and initial bond yield spread. Results suggest that firms beating earnings benchmarks have a higher probability of rating upgrades and a smaller initial bond yield spread. Additional analyses indicate that (1) the benefits of beating earnings benchmarks are more pronounced for firms with high default risk; (2) beating the zero earnings benchmark generally provides the biggest reward in terms of a lower cost of debt; and (3) the reduction in the cost of debt is attenuated but does not disappear for firms beating benchmarks through earnings management. In sum, results suggest that there are benefits associated with beating earnings benchmarks in the debt market. These benefits vary by benchmark, firm default risk, and method utilized to beat the benchmark. Among other implications, this evidence suggests that the relative importance of specific benchmarks differs across the equity and bond markets.
APA, Harvard, Vancouver, ISO, and other styles
17

Couto, Braulio Roberto Gonçalves Marinho, and Carlos Ernesto Ferreira Starling. "823. How to Compare standardized Healthcare-associated Infection (HAI) Rates? Benchmark 2D and 3D." Open Forum Infectious Diseases 7, Supplement_1 (October 1, 2020): S453—S454. http://dx.doi.org/10.1093/ofid/ofaa439.1012.

Full text
Abstract:
Abstract Background External benchmarking involves comparing standardized data on HAI rates in one hospital or healthcare facility in relation to others. Here we present two epidemiological graphical tools, 2D and 3D benchmarks, which summarize the efficiency in preventing main infections in a Medical/Surgical Intensive Care Unit (MSICU). Methods The 3D benchmark graph considers the incidence density rate of ventilator-associated pneumonias (VAP cases per 1,000 ventilator-days) as the X-Axis, the incidence density rate of central line-associated primary bloodstream infections (CLABSI cases per 1,000 central line-days) as the Y-Axis, and the incidence density rate of urinary catheter-associated urinary tract infections (CAUTI per 1,000 urinary catheter-days) as the Z-Axis. Efficiency in preventing infection (e) considers the zero rate to be 100% efficient (e=100%) and the highest available benchmark rate to be “zero” efficiency (RMax: e=0%). From this definition, the efficiency of any MSICU (0% ≤ e ≤ 100%) is obtained using a linear interpolation function, from the rate observed in the MSICU under evaluation (Rx): e = 100x(RMax – Rx)/RMax. If Rx > RMax, then RMax = Rx. The 3D benchmark is build by calculating the preventing infection (e) for each infection (VAP, CLABSI, and CAUTI) for all benchmarks and for the MSICU under evaluation. In the 3D Benchmark, three control volumes are created: “Infection Control Urgency” volume, “Infection Control Excellence” volume, “Infection Prevention Opportunity” volume. Benchmark 2D considers only the VAP density rate as X-Axis, and the CLABSI density rate as Y-Axis. In this graph, five control regions are created: 1=excellence in the control of VAP+CLABSI; 2=excellence in VAP control and opportunity for CLABSI prevention; 3=excellence in CLABSI control and opportunity to prevent VAP; 4=opportunity to prevent VAP+CLABSI; 5=urgency in infection control. Results Graph parameters were based on NHSN data from the device-associated module, NOIS Project, Anahp, CQH, and GVIMS/GGTES/ANVISA (Brazilian benchmarks), and El-Saed et al. benchmarks. We applied the 2D/3D benchmarks to several Brazilian ICUs. 2D benchmark for the MSICUs from Lifecenter Hospital, Brazil, Jan-Dez/2019: UCO & UTI 19 =excellence in CLABSI control and opportunity to prevent VAP; UTI 20=excellence in VAP control and opportunity for CLABSI prevention; UTI 18=opportunity to prevent VAP+CLABSI. 3D benchmark for the MSICUs from Lifecenter Hospital, Brazil, Jan-Dez/2019 2D benchmark for the MSICUs from Vera Cruz Hospital, Brazil, Jan-Dez/2019: CTI 3.o Andar =excellence in the control of VAP+CLABSI; CTI 1.o Andar=excellence in VAP control and opportunity for CLABSI prevention. Conclusion 2D and 3D benchmarks are easy to understand and summarize the efficiency in prevention the mains infections of MSICU. Disclosures All Authors: No reported disclosures
APA, Harvard, Vancouver, ISO, and other styles
18

LEHMANN, BRUCE N., and DAVID M. MODEST. "Mutual Fund Performance Evaluation: A Comparison of Benchmarks and Benchmark Comparisons." Journal of Finance 42, no. 2 (June 1987): 233–65. http://dx.doi.org/10.1111/j.1540-6261.1987.tb02566.x.

Full text
APA, Harvard, Vancouver, ISO, and other styles
19

Platen, Eckhard. "A class of complete benchmark models with intensity-based jumps." Journal of Applied Probability 41, no. 1 (March 2004): 19–34. http://dx.doi.org/10.1239/jap/1077134665.

Full text
Abstract:
This paper proposes a class of complete financial market models, the benchmark models, with security price processes that exhibit intensity-based jumps. The benchmark or reference unit is chosen to be the growth-optimal portfolio. Primary security account prices, when expressed in units of the benchmark, turn out to be local martingales. In the proposed framework an equivalent risk-neutral measure need not exist. Benchmarked fair derivative prices are obtained as conditional expectations of future benchmarked prices under the real-world probability measure. This concept of fair pricing generalizes the classical risk-neutral approach and the actuarial present-value pricing methodology.
APA, Harvard, Vancouver, ISO, and other styles
20

Platen, Eckhard. "A class of complete benchmark models with intensity-based jumps." Journal of Applied Probability 41, no. 01 (March 2004): 19–34. http://dx.doi.org/10.1017/s0021900200014017.

Full text
Abstract:
This paper proposes a class of complete financial market models, the benchmark models, with security price processes that exhibit intensity-based jumps. The benchmark or reference unit is chosen to be the growth-optimal portfolio. Primary security account prices, when expressed in units of the benchmark, turn out to be local martingales. In the proposed framework an equivalent risk-neutral measure need not exist. Benchmarked fair derivative prices are obtained as conditional expectations of future benchmarked prices under the real-world probability measure. This concept of fair pricing generalizes the classical risk-neutral approach and the actuarial present-value pricing methodology.
APA, Harvard, Vancouver, ISO, and other styles
21

Zahorec, Pavol, Juraj Papčo, Peter Vajda, Filippo Greco, Massimo Cantarero, and Daniele Carbone. "Refined prediction of vertical gradient of gravity at Etna volcano gravity network (Italy)." Contributions to Geophysics and Geodesy 48, no. 4 (December 1, 2018): 299–317. http://dx.doi.org/10.2478/congeo-2018-0014.

Full text
Abstract:
Abstract Predicted values of the vertical gradient of gravity (VGG) on benchmarks of Etna’s monitoring system, based on calculation of the topographic contribution to the theoretical free-air gradient, are compared with VGG values observed in situ. The verification campaign indicated that improvements are required when predicting the VGGs at such networks. Our work identified the following factors to be resolved: (a) accuracy of the benchmark position; (b) gravitational effect of buildings and roadside walls adjacent to benchmarks; (c) accuracy of the digital elevation model (DEM) in the proximity of benchmarks. Benchmark positions were refined using precise geodetic methods. The gravitational effects of the benchmark-adjacent walls and buildings were modeled and accounted for in the prediction. New high-resolution DEMs were produced in the innermost zone at some benchmarks based on drone-flown photogrammetry to improve the VGG prediction at those benchmarks. The three described refinements in the VGG prediction improved the match between predicted and in situ observed VGGs at the network considerably. The standard deviation of differences between the measured and predicted VGG values decreased from 36 to 13 μGal/m.
APA, Harvard, Vancouver, ISO, and other styles
22

Orogat, Abdelghny, and Ahmed El-Roby. "SmartBench." Proceedings of the VLDB Endowment 15, no. 12 (August 2022): 3662–65. http://dx.doi.org/10.14778/3554821.3554869.

Full text
Abstract:
In recent years, a significant number of question answering (QA) systems that retrieve answers to natural language questions from knowledge graphs (KG) have been introduced. However, finding a benchmark that accurately evaluates the quality of a question answering system is a difficult task because of (1) the high degree of variations with respect to the fine-grained properties among the available benchmarks, (2) the static nature of the available benchmarks versus the evolving nature of KGs, and (3) the limited number of KGs targeted by existing benchmarks, which hinders the usability of QA systems in real deployment over KGs that are different from those which the QA system was evaluated using. In this demonstration, we introduce SmartBench, an automatic benchmark generating system for QA over any KG. The benchmark generated by SmartBench is guaranteed to cover all the properties of the natural language questions and queries that were encountered in the literature as long as the targeted KG includes these properties.
APA, Harvard, Vancouver, ISO, and other styles
23

Harmon, Thomas C., Robyn L. Smyth, Sudeep Chandra, Daniel Conde, Ramesh Dhungel, Jaime Escobar, Natalia Hoyos, et al. "Socioeconomic and Environmental Proxies for Comparing Freshwater Ecosystem Service Threats across International Sites: A Diagnostic Approach." Water 10, no. 11 (November 4, 2018): 1578. http://dx.doi.org/10.3390/w10111578.

Full text
Abstract:
In this work, we develop and test proxy-based diagnostic tools for comparing freshwater ecosystem services (FWES) risks across an international array of freshwater ecosystems. FWES threats are increasing rapidly under pressure from population, climate change, pollution, land use change, and other factors. We identified spatially explicit FWES threats estimates (referred to as threat benchmarks) and extracted watershed-specific values for an array of aquatic ecosystems in the Western Hemisphere (Ramsar sites). We compared these benchmark values to values extracted for sites associated with an international FWES threat investigation. The resulting benchmark threats appeared to provide a meaningful context for the diagnostic assessment of study site selection by revealing gaps in coverage of the underlying socio-environmental problem. In an effort to simplify the method, we tested regularly updated environmental and socioeconomic metrics as potential proxies for the benchmark threats using regression analysis. Three category proxies, aggregated from (i) external (global to regional, climate-related), (ii) internal (watershed management-related), and (iii) socioeconomic and governance related proxies produced strong relationships with water supply threat benchmarks, but only weak relationships with biodiversity-related and nutrient regulation benchmark threats. Our results demonstrate the utility of advancing global FWES status and threat benchmarks for organizing coordinated research efforts and prioritizing decisions with regard to international socio-environmental problems.
APA, Harvard, Vancouver, ISO, and other styles
24

Su, Tzu Ching, and Hsien Te Lin. "Electricity Consumption Benchmark and Ranking System for Office Buildings in Taiwan." Applied Mechanics and Materials 71-78 (July 2011): 2352–56. http://dx.doi.org/10.4028/www.scientific.net/amm.71-78.2352.

Full text
Abstract:
Energy use intensity (EUI) is a significant index for energy management in the building sector; however, previous research that determined an EUI benchmark for a type of building resulted in serious problems. Such earlier EUI benchmarks cannot fairly evaluate office buildings with different proportions of parking areas and different numbers of floors, becoming an obstacle for energy management. Therefore, this study proposes an area-weighted office building EUI benchmark that calculates according to office areas and parking areas. This study subsequently surveys the electricity consumption of 58 office buildings in Taiwan to determine the office area EUI benchmark that depends on the number of floors in a building, and suggests a reasonable parking area EUI benchmark. Finally, this study promotes a fairer EUI benchmark and ranking system for office buildings in Taiwan. Such a benchmark and ranking system act as a reference for the promotion of the building energy certificate system in Taiwan.
APA, Harvard, Vancouver, ISO, and other styles
25

Macías, Jorge, Cipriano Escalante, and Manuel J. Castro. "Multilayer-HySEA model validation for landslide-generated tsunamis – Part 2: Granular slides." Natural Hazards and Earth System Sciences 21, no. 2 (February 26, 2021): 791–805. http://dx.doi.org/10.5194/nhess-21-791-2021.

Full text
Abstract:
Abstract. The final aim of the present work is to propose a NTHMP-benchmarked numerical tool for landslide-generated tsunami hazard assessment. To achieve this, the novel Multilayer-HySEA model is validated using laboratory experiment data for landslide-generated tsunamis. In particular, this second part of the work deals with granular slides, while the first part, in a companion paper, considers rigid slides. The experimental data used have been proposed by the US National Tsunami Hazard and Mitigation Program (NTHMP) and were established for the NTHMP Landslide Benchmark Workshop, held in January 2017 at Galveston (Texas). Three of the seven benchmark problems proposed in that workshop dealt with tsunamis generated by rigid slides and are collected in the companion paper (Macías et al., 2021). Another three benchmarks considered tsunamis generated by granular slides. They are the subject of the present study. The seventh benchmark problem proposed the field case of Port Valdez, Alaska, 1964 and can be found in Macías et al. (2017). In order to reproduce the laboratory experiments dealing with granular slides, two models need to be coupled: one for the granular slide and a second one for the water dynamics. The coupled model used consists of a new and efficient hybrid finite-volume–finite-difference implementation on GPU architectures of a non-hydrostatic multilayer model coupled with a Savage–Hutter model. To introduce the multilayer model more fluidly, we first present the equations of the one-layer model, Landslide-HySEA, with both strong and weak couplings between the fluid layer and the granular slide. Then, a brief description of the multilayer model equations and the numerical scheme used is included. The dispersive properties of the multilayer model can be found in the companion paper. Then, results for the three NTHMP benchmark problems dealing with tsunamis generated by granular slides are presented with a description of each benchmark problem.
APA, Harvard, Vancouver, ISO, and other styles
26

Pyotr Skorobogaty. "CRIMEAN BENCHMARK." Current Digest of the Post-Soviet Press, The 68, no. 012 (March 21, 2016): 10–11. http://dx.doi.org/10.21557/dsp.46460581.

Full text
APA, Harvard, Vancouver, ISO, and other styles
27

Podro, Michael, and Eric Fernie. "WORMALD'S BENCHMARK." Art History 8, no. 2 (June 1985): 247–48. http://dx.doi.org/10.1111/j.1467-8365.1985.tb00164.x.

Full text
APA, Harvard, Vancouver, ISO, and other styles
28

Ponder, Carl G. "Benchmark semantics." ACM SIGPLAN Notices 23, no. 2 (February 1988): 44–48. http://dx.doi.org/10.1145/43908.43913.

Full text
APA, Harvard, Vancouver, ISO, and other styles
29

Ponder, Carl G. "Benchmark semantics." ACM SIGPLAN Notices 23, no. 6 (June 1988): 157–61. http://dx.doi.org/10.1145/44546.44563.

Full text
APA, Harvard, Vancouver, ISO, and other styles
30

Ponder, Carl G. "Benchmark semantics." ACM SIGMETRICS Performance Evaluation Review 18, no. 4 (March 1991): 20–24. http://dx.doi.org/10.1145/122289.122291.

Full text
APA, Harvard, Vancouver, ISO, and other styles
31

Conte, T. M., and W. M. W. Hwu. "Benchmark characterization." Computer 24, no. 1 (January 1991): 48–56. http://dx.doi.org/10.1109/2.67193.

Full text
APA, Harvard, Vancouver, ISO, and other styles
32

Saavedra, Rafael H., and Alan J. Smith. "Analysis of benchmark characteristics and benchmark performance prediction." ACM Transactions on Computer Systems 14, no. 4 (November 1996): 344–84. http://dx.doi.org/10.1145/235543.235545.

Full text
APA, Harvard, Vancouver, ISO, and other styles
33

Poovey, Jason A., Thomas M. Conte, Markus Levy, and Shay Gal-On. "A Benchmark Characterization of the EEMBC Benchmark Suite." IEEE Micro 29, no. 5 (September 2009): 18–29. http://dx.doi.org/10.1109/mm.2009.74.

Full text
APA, Harvard, Vancouver, ISO, and other styles
34

Klauck, Michaela, Marcel Steinmetz, Jörg Hoffmann, and Holger Hermanns. "Bridging the Gap Between Probabilistic Model Checking and Probabilistic Planning: Survey, Compilations, and Empirical Comparison." Journal of Artificial Intelligence Research 68 (May 19, 2020): 247–310. http://dx.doi.org/10.1613/jair.1.11595.

Full text
Abstract:
Markov decision processes are of major interest in the planning community as well as in the model checking community. But in spite of the similarity in the considered formal models, the development of new techniques and methods happened largely independently in both communities. This work is intended as a beginning to unite the two research branches. We consider goal-reachability analysis as a common basis between both communities. The core of this paper is the translation from Jani, an overarching input language for quantitative model checkers, into the probabilistic planning domain definition language (PPDDL), and vice versa from PPDDL into Jani. These translations allow the creation of an overarching benchmark collection, including existing case studies from the model checking community, as well as benchmarks from the international probabilistic planning competitions (IPPC). We use this benchmark set as a basis for an extensive empirical comparison of various approaches from the model checking community, variants of value iteration, and MDP heuristic search algorithms developed by the AI planning community. On a per benchmark domain basis, techniques from one community can achieve state-ofthe-art performance in benchmarks of the other community. Across all benchmark domains of one community, the performance comparison is however in favor of the solvers and algorithms of that particular community. Reasons are the design of the benchmarks, as well as tool-related limitations. Our translation methods and benchmark collection foster crossfertilization between both communities, pointing out specific opportunities for widening the scope of solvers to different kinds of models, as well as for exchanging and adopting algorithms across communities.
APA, Harvard, Vancouver, ISO, and other styles
35

Dorval, Eric. "Validation of the Serpent 2 Monte Carlo code for reactor dosimetry applications." EPJ Web of Conferences 278 (2023): 02002. http://dx.doi.org/10.1051/epjconf/202327802002.

Full text
Abstract:
The verification and validation (V&V) of the Serpent 2 Monte Carlo code for 3-D reactor dosimetry applications is being carried out at VTT. Two code-to-code computational benchmark cases were calculated by MCNP and Serpent 2. The computational efficiency of different variance reduction techniques was appraised. The validation part currently includes two experimental benchmarks from the SINBAD database: the Pool Critical Assembly-Pressure Vessel Facility (PCA); and the H.B. Robinson-2 (HBR-2) Pressure Vessel Dosimetry Benchmark. All simulations replicate the benchmark source specifications at pin level thanks to the development of a flexible, geometry-independent, fixed source definition for reactor dosimetry applications. Good agreement was found in all benchmark cases.
APA, Harvard, Vancouver, ISO, and other styles
36

Adedeji, Gbadebo, Akande Joseph, Yoshifumi Harada, Orraya Suwanno, and Adekunle Ahmed. "Benchmark Beating and Earnings Manipulation in Nigerian Firms." Asian Economic and Financial Review 13, no. 1 (December 20, 2022): 49–73. http://dx.doi.org/10.55493/5002.v13i1.4694.

Full text
Abstract:
Earnings management among firms remains a central focus for academics, auditors and regulatory bodies. Benchmark-motivated earnings management occurs when managers engage in opportunistic activities, including flexible use of accounting standards to misrepresent the information in a firm’s financial reports. Academic research has focused on how firms manage earnings to beat benchmarks, but the evidence regarding firms in emerging African stock markets is scarce and none is available for Nigeria. We applied both accruals quality and discretionary accruals models to detect whether firms that beat earnings benchmarks report earnings differently from others. Using 161 firms listed on the Nigerian Stock Exchange from 2002 to 2019, the study verifies how benchmark beaters manage earnings under the framework of two earnings thresholds – earnings (level) and positive earnings changes. Earnings persistence tests were carried out to verify whether benchmark beaters are consistent manipulators relative to non-beaters. The findings indicate that positive earnings benchmarks differ among the dichotomized groups. The evidence is not sufficient to validate that the change in earnings benchmarks motivates earnings discretions. However, the evidence may improve for larger samples. The study offers insights for informed decisions on the expectation of investment returns for investors, creditors, and other market partakers that require earnings information.
APA, Harvard, Vancouver, ISO, and other styles
37

Gao, Lijie, Xiaoqi Shang, Fengmei Yang, and Longyu Shi. "A Dynamic Benchmark System for Per Capita Carbon Emissions in Low-Carbon Counties of China." Energies 14, no. 3 (January 25, 2021): 599. http://dx.doi.org/10.3390/en14030599.

Full text
Abstract:
As the most basic unit of the national economy and administrative management, the low-carbon transformation of the vast counties is of great significance to China’s overall greenhouse gas emission reduction. Although the low-carbon evaluation (LCE) indicator system and benchmarks have been extensively studied, most benchmarks ignore the needs of the evaluated object at the development stage. When the local economy develops to a certain level, it may be restricted by static low-carbon target constraints. This study reviews the relevant research on LCE indicator system and benchmarks based on convergence. The Environmental Kuznets Curve (EKC), a dynamic benchmark system for per capita carbon emissions (PCCEs), is proposed for low-carbon counties. Taking Changxing County, Zhejiang Province, China as an example, a dynamic benchmark for PCCEs was established by benchmarking the Carbon Kuznets Curve (CKC) of best practices. Based on the principles of best practice, comparability, data completeness, and the CKC hypothesis acceptance, the best practice database is screened, and Singapore is selected as a potential benchmark. By constructing an econometric model to conduct an empirical study on Singapore’s CKC hypothesis, the regression results of the least squares method support the CKC hypothesis and its rationality as a benchmark. The result of the PCCE benchmarks of Changxing County show that when the per capita income of Changxing County in 2025, 2030, and 2035 reaches USD 19,172.92, USD 24,483.01, and USD 29,366.11, respectively, the corresponding benchmarks should be 14.95 tons CO2/person, 14.70 tons CO2/person, and 13.55 tons CO2/person. For every 1% increase in the county’s per capita income, the PCCE allowable room for growth is 17.6453%. The turning point is when the per capita gross domestic product (PCGDP) is USD 20,843.23 and the PCCE is 15.03 tons of CO2/person, which will occur between 2025 and 2030. Prior to this, the PCCE benchmark increases with the increase of PCGDP. After that, the PCCE benchmark decreases with the increase of PCGDP. The system is economically sensitive, adaptable to different development stages, and enriches the methodology of low-carbon indicator evaluation and benchmark setting at the county scale. It can provide scientific basis for Chinese county decision makers to formulate reasonable targets under the management idea driven by evaluation indicators and emission reduction targets and help counties explore the coordinated paths of economic development and emission reduction in different development stages. It has certain reference significance for other developing regions facing similar challenges of economic development and low-carbon transformation to Changxing County to formulate scientific and reasonable low-carbon emission reduction targets.
APA, Harvard, Vancouver, ISO, and other styles
38

Huang, Eric, and Richard Korf. "Optimal Rectangle Packing on Non-Square Benchmarks." Proceedings of the AAAI Conference on Artificial Intelligence 24, no. 1 (July 3, 2010): 83–88. http://dx.doi.org/10.1609/aaai.v24i1.7538.

Full text
Abstract:
The rectangle packing problem consists of finding an enclosing rectangle of smallest area that can contain a given set of rectangles without overlap. We propose two new benchmarks, one where the orientation of the rectangles is fixed and one where it is free, that include rectangles of various aspect ratios. The new benchmarks avoid certain properties of easy instances, which we identify as instances where rectangles have dimensions in common or where a few rectangles occupy most of the area. Our benchmarks are much more difficult for the previous state-of-the-art solver, requiring orders of magnitude more time, compared to similar-sized instances from a popular benchmark consisting only of squares. On the new benchmarks, we improve upon the previous strategy used to handle dominance conditions, we define a variable order over non-square rectangles that generalizes previous strategies, and we present a way to adjust the sizes of intervals of values for each rectangle's x-coordinates. Using these techniques together, we can solve the new oriented benchmark about 500 times faster, and the new unoriented benchmark about 40 times faster than the previous state-of-the-art.
APA, Harvard, Vancouver, ISO, and other styles
39

Rafik, Abdur, and Salsabila Annisa Azmi. "Long-Run Underperformance on Seasoned Equity Offerings: An Evidence from Indonesia." Binus Business Review 10, no. 1 (March 31, 2019): 41–49. http://dx.doi.org/10.21512/bbr.v10i1.5403.

Full text
Abstract:
The research aimed to analyze the underperformance phenomenon following Seasoned Equity Offering (SEO)in Indonesian context. Samples were all firms listed on the Indonesia Stock Exchange with the right issue in the observed periods, were chosen by purposive sampling with several criteria. In total, there were 109 issuing firms from 2009-2014 that were analyzed using performance benchmarking approaches. The approaches consisted of market-based, size-based, growth-based, and industry-based benchmarks. The market-based was constructed using a market return. Then, the size and the growth benchmarks were constructed on the basis of closest market capitalization and closest price-to-book value respectively. Then, the industry benchmark was based on a closest combined of market capitalization and Price to Book Value (PBV) of matched firms within the first-second digit of Standard Industry Classification (SIC). The test was conducted using standard t and Wilcoxon tests by examining the benchmark-based abnormal returns over various spans ranging from 3 to 36 months following the right offerings. Like several findings in developed countries, the results also confirm the underperformance phenomenon following right offerings in Indonesia. The negative abnormal returns are found for all benchmark performances, but they are only significant for the market benchmark and partially significant for the size and the growth benchmarks. Behavioral explanations need to be modeled to reveal the intuitions behind the results.
APA, Harvard, Vancouver, ISO, and other styles
40

Rebel, P., C. Poblete-Echeverría, J. G. van Zyl, J. P. B. Wessels, C. Coetzer, and A. McLeod. "Determining Mancozeb Deposition Benchmark Values on Apple Leaves for the Management of Venturia inaequalis." Plant Disease 104, no. 1 (January 2020): 168–78. http://dx.doi.org/10.1094/pdis-04-19-0873-re.

Full text
Abstract:
Apple scab, caused by Venturia inaequalis, is the most common fruit and foliar disease in commercial apple production worldwide. Early in the production season, preventative contact fungicide sprays are essential for protecting highly susceptible continuously unfolding and expanding young leaves. In South Africa, mancozeb is a key contact fungicide used for controlling apple scab early in the season. The current study developed deposition benchmarks indicative of the biological efficacy of mancozeb against apple scab, using a laboratory-based apple seedling model system. The model system employed a yellow fluorescent pigment that is known to be an effective tracer of mancozeb deposition. A concentration range of mancozeb (0.15 to 1 times the registered dosage) and fluorescent pigment concentrations was sprayed onto seedling leaves, which yielded various fluorescent particle coverage (FPC%) levels. Modeling of the FPC% values versus percent disease control yielded different benchmark values when disease quantification was conducted using two different methods. Thermal infrared imaging (TIRI) disease quantification resulted in a benchmark model where 0.40%, 0.79%, and 1.35 FPC% yielded 50, 75, and 90% apple scab control, respectively. These FPC% values were higher than the benchmarks (0.10, 0.20, and 0.34 FPC%, respectively) obtained with quantitative real-time PCR (qPCR) disease quantification. The qPCR benchmark model is recommended as a guideline for evaluating the efficacy of mancozeb sprays on leaves in apple orchards since the TIRI benchmark model underestimated disease control. The TIRI benchmark model yielded 68% disease control at the lowest mancozeb dosage, yet no visible lesion developed at this dosage. Both benchmark models showed that mancozeb yielded high levels of disease control at very low concentrations; for the qPCR benchmark model the FPC% value of the FPC90 (90% control) corresponded to 0.15 times that of the registered mancozeb concentration in South Africa, i.e., 85% lower than the registered dosage.
APA, Harvard, Vancouver, ISO, and other styles
41

Abusam, A., K. J. Keesman, H. Spanjers, G. van Straten, and K. Meinema. "Evaluation of control strategies using an oxidation ditch benchmark." Water Science and Technology 45, no. 4-5 (February 1, 2002): 151–58. http://dx.doi.org/10.2166/wst.2002.0574.

Full text
Abstract:
This paper presents validation and implementation results of a benchmark developed for a specific full-scale oxidation ditch wastewater treatment plant. A benchmark is a standard simulation procedure that can be used as a tool in evaluating various control strategies proposed for wastewater treatment plants. It is based on model and performance criteria development. Testing of this benchmark, by comparing benchmark predictions to real measurements of the electrical energy consumptions and amounts of disposed sludge for a specific oxidation ditch WWTP, has shown that it can (reasonably) be used for evaluating the performance of this WWTP. Subsequently, the validated benchmark was then used in evaluating some basic and advanced control strategies. Some of the interesting results obtained are the following: (i) influent flow splitting ratio, between the first and the fourth aerated compartments of the ditch, has no significant effect on the TN concentrations in the effluent, and (ii) for evaluation of long-term control strategies, future benchmarks need to be able to assess settlers' performance.
APA, Harvard, Vancouver, ISO, and other styles
42

Baek, Doehyun, Jakob Getz, Yusung Sim, Daniel Lehmann, Ben L. Titzer, Sukyoung Ryu, and Michael Pradel. "Wasm-R3: Record-Reduce-Replay for Realistic and Standalone WebAssembly Benchmarks." Proceedings of the ACM on Programming Languages 8, OOPSLA2 (October 8, 2024): 2156–82. http://dx.doi.org/10.1145/3689787.

Full text
Abstract:
WebAssembly (Wasm for short) brings a new, powerful capability to the web as well as Edge, IoT, and embedded systems. Wasm is a portable, compact binary code format with high performance and robust sandboxing properties. As Wasm applications grow in size and importance, the complex performance characteristics of diverse Wasm engines demand robust, representative benchmarks for proper tuning. Stopgap benchmark suites, such as PolyBenchC and libsodium, continue to be used in the literature, though they are known to be unrepresentative. Porting of more complex suites remains difficult because Wasm lacks many system APIs and extracting real-world Wasm benchmarks from the web is difficult due to complex host interactions. To address this challenge, we introduce Wasm-R3, the first record and replay technique for Wasm. Wasm-R3 transparently injects instrumentation into Wasm modules to record an execution trace from inside the module, then reduces the execution trace via several optimizations, and finally produces a replay module that is executable standalone without any host environment-on any engine. The benchmarks created by our approach are (i) realistic, because the approach records real-world web applications, (ii) faithful to the original execution, because the replay benchmark includes the unmodified original code, only adding emulation of host interactions, and (iii) standalone, because the replay benchmarks run on any engine. Applying Wasm-R3 to web-based Wasm applications in the wild demonstrates the correctness of our approach as well as the effectiveness of our optimizations, which reduce the recorded traces by 99.53% and the size of the replay benchmark by 9.98%. We release the resulting benchmark suite of 27 applications, called Wasm-R3-Bench, to the community, to inspire a new generation of realistic and standalone Wasm benchmarks.
APA, Harvard, Vancouver, ISO, and other styles
43

Valentine, Timothy, Maria Avramova, Michael Fleming, Mathieu Hursin, Kostadin Ivanov, Alessandro Petruzzi, Upendra Rohatgi, and Kiril Velkov. "OVERVIEW OF THE OECD-NEA EXPERT GROUP ON MULTI-PHYSICS EXPERIMENTAL DATA, BENCHMARKS AND VALIDATION." EPJ Web of Conferences 247 (2021): 06048. http://dx.doi.org/10.1051/epjconf/202124706048.

Full text
Abstract:
The OECD Nuclear Energy Agency (NEA) Nuclear Science Committee (NSC) established the Expert Group on Multi-physics Experimental Data, Benchmarks and Validation (EGMPEBV) in 2014 to bridge the gap between advanced, multi-physics simulation capabilities and the relatively low availability of dedicated, high-fidelity experimental data and benchmarks specifically for multi-physics modelling and simulation tools. The EGMPEBV was mandated to establish mechanisms for the certification of experimental data and benchmark models and to establish the processes and procedures for the validation of multi-physics modelling and simulation tools. The EGMPEBV oversees three task forces, covering (1) experimental data qualification and benchmark evaluation, (2) validation guidelines and needs and (3) example application of validation experiments. These have generated numerous reports surveying the state-of-the-art in multi-physics validation, challenge areas and recommendations for the evaluation of multi-physics benchmarks, while in parallel developing the specifications for multi-physics benchmarks. Three benchmark specifications are in active development, including a reactivity compensation scenario in the Rostov Unit 2 VVER-1000, multi-cycle depletion of the TVA Watts Bar Unit 1 and study of pellet cladding mechanical interaction within ramp tests performed at the Studsvik R2 reactor. We provide an overview of the recent progress in these areas and a summary of the future activities of the EGMPEBV in establishing international multi-physics benchmarks.
APA, Harvard, Vancouver, ISO, and other styles
44

de Goede, Marieke, Bert Enserink, Ignaz Worm, and Jan Peter van der Hoek. "Drivers for performance improvement originating from the Dutch drinking water benchmark." Water Policy 18, no. 5 (March 30, 2016): 1247–66. http://dx.doi.org/10.2166/wp.2016.125.

Full text
Abstract:
The Dutch drinking water sector has been benchmarked every 3 years since 1997, and the sector has significantly improved performance since then. Based on interviews with CEOs and financial managers of drinking water companies five drivers for improvement as a result of this benchmark are identified: ‘learning effect’, ‘enhanced transparency’, ‘managed competition’, ‘avoidance of negative consequences' and ‘personal honour of director’. Different developments have caused stagnation of further improvement: the variation on the benchmarked performance indicators has decreased, participation in the benchmark became mandatory for all Dutch drinking water supply organizations, it lacks a focus on the future, and participating organizations experience high financial pressure. These developments decrease the influence of the drivers. Four possible new impulses for the benchmark are identified and their influence on the effect of the drivers is analysed. The two most promising new impulses are to make the benchmark adaptive and to involve consumers in the process of benchmarking, both have a positive influence on the effect of almost all drivers. This study contributes to the understanding of how benchmarking leads to improvement and to the analysis of the impact of design choices, leading to well-founded decisions for re-design of the Dutch drinking water benchmark.
APA, Harvard, Vancouver, ISO, and other styles
45

Zachar, P., K. Bakuła, and W. Ostrowski. "CENAGIS-ALS BENCHMARK - NEW PROPOSAL FOR DENSE ALS BENCHMARK BASED ON THE REVIEW OF DATASETS AND BENCHMARKS FOR 3D POINT CLOUD SEGMENTATION." International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences XLVIII-1/W3-2023 (October 19, 2023): 227–34. http://dx.doi.org/10.5194/isprs-archives-xlviii-1-w3-2023-227-2023.

Full text
Abstract:
Abstract. Benchmarking is an essential tool for scientific and technological progress. This article reviews the benchmarks for 3D point cloud segmentation and classification. Based on the analysis of the articles and the knowledge gathered, it can be concluded that there has been an increase in the number of benchmarks, allowing to compare research results against specific performance metrics independently. However, benchmarks vary regarding the number of classes, spatial size, nomenclature, and class division. In this article, we introduce a new annotated 3D dataset - CENAGIS-ALS Benchmark. We propose a benchmark of highly dense lidar point clouds acquired by Leica CityMapper-2 for the Centre of Warsaw, Poland. The area covers 2 km2, and the data has a density of 275 pts/m2. The dataset consists of a number of classes that are distinguishable for this type of data. In addition to the basic classes, more specialized classes, important from the perspective of urban space, are also distinguished. Moreover, the division of classes consists of three levels of detail from coarse (e.g., a building) to refined elements (e.g., roofs, chimneys, and other rooftop objects). This benchmark can contribute to geospatial societies, considering the large spatial size of the study area with unified data quality and the higher number of classes with the hierarchical division compared to other benchmarking data.
APA, Harvard, Vancouver, ISO, and other styles
46

Ayers, Benjamin C., John (Xuefeng) Jiang, and P. Eric Yeung. "Discretionary Accruals and Earnings Management: An Analysis of Pseudo Earnings Targets." Accounting Review 81, no. 3 (May 1, 2006): 617–52. http://dx.doi.org/10.2308/accr.2006.81.3.617.

Full text
Abstract:
We investigate whether the positive associations between discretionary accrual proxies and beating earnings benchmarks hold for comparisons of groups segregated at other points in the distributions of earnings, earnings changes, and analystsbased unexpected earnings. We refer to these points as “pseudo” targets. Results suggest that the positive association between discretionary accruals and beating the profit benchmark extends to pseudo targets throughout the earnings distribution. We find similar results for the earnings change distribution. In contrast, we find few positive associations between discretionary accruals and beating pseudo targets derived from analysts-based unexpected earnings. We develop an additional analysis that accounts for the systematic association between discretionary accruals and earnings and earnings changes. Results suggest that the positive association between discretionary accruals and earnings intensifies around the actual profit benchmark (i.e., where earnings management incentives may be more pronounced). We find similar effects around the actual earnings increase benchmark. However, analogous patterns exist for cash flows around the profit and earnings increase benchmarks. In sum, we are unable to eliminate other plausible explanations for the associations between discretionary accruals and beating the profit and earnings increase benchmarks.
APA, Harvard, Vancouver, ISO, and other styles
47

Katz, Michael, Junkyu Lee, and Shirin Sohrabi. "Generating SAS+ Planning Tasks of Specified Causal Structure." Proceedings of the International Symposium on Combinatorial Search 16, no. 1 (July 2, 2023): 29–37. http://dx.doi.org/10.1609/socs.v16i1.27280.

Full text
Abstract:
Recent advances in data-driven approaches in AI planning demand more and more planning tasks. The supply, however, is somewhat limited. Past International Planning Competitions (IPCs) have introduced the de-facto standard benchmarks with the domains written by domain experts. The few existing methods for sampling random planning tasks severely limit the resulting problem structure. In this work we show a method for generating planning tasks of any requested causal graph structure, alleviating the shortage in existing planning benchmarks. We present an algorithm for constructing random SAS+ planning tasks given an arbitrary causal graph and offer random task generators for the well-explored causal graph structures in the planning literature. We further allow to generate a planning task equivalent in causal structure to an input SAS+ planning task. We generate two benchmark sets: 26 collections for select well-explored causal graph structures and 42 collections for existing IPC domains. We evaluate both benchmark sets with the state-of-the-art optimal planners, showing the adequacy for adopting them as benchmarks in cost-optimal classical planning. The benchmark sets and the task generator code are publicly available at https://github.com/IBM/fdr-generator.
APA, Harvard, Vancouver, ISO, and other styles
48

Church, Kenneth Ward. "Benchmarks and goals." Natural Language Engineering 26, no. 5 (August 10, 2020): 579–92. http://dx.doi.org/10.1017/s1351324920000418.

Full text
Abstract:
AbstractBenchmarks can be a useful step toward the goals of the field (when the benchmark is on the critical path), as demonstrated by the GLUE benchmark, and deep nets such as BERT and ERNIE. The case for other benchmarks such as MUSE and WN18RR is less well established. Hopefully, these benchmarks are on a critical path toward progress on bilingual lexicon induction (BLI) and knowledge graph completion (KGC). Many KGC algorithms have been proposed such as Trans[DEHRM], but it remains to be seen how this work improves WordNet coverage. Given how much work is based on these benchmarks, the literature should have more to say than it does about the connection between benchmarks and goals. Is optimizing P@10 on WN18RR likely to produce more complete knowledge graphs? Is MUSE likely to improve Machine Translation?
APA, Harvard, Vancouver, ISO, and other styles
49

Kwas, Marek, and Michał Rubaszek. "Forecasting Commodity Prices: Looking for a Benchmark." Forecasting 3, no. 2 (June 19, 2021): 447–59. http://dx.doi.org/10.3390/forecast3020027.

Full text
Abstract:
The random walk, no-change forecast is a customary benchmark in the literature on forecasting commodity prices. We challenge this custom by examining whether alternative models are more suited for this purpose. Based on a literature review and the results of two out-of-sample forecasting experiments, we draw two conclusions. First, in forecasting nominal commodity prices at shorter horizons, the random walk benchmark should be supplemented by futures-based forecasts. Second, in forecasting real commodity prices, the random walk benchmark should be supplemented, if not substituted, by forecasts from the local projection models. In both cases, the alternative benchmarks deliver forecasts of comparable and, in many cases, of superior accuracy.
APA, Harvard, Vancouver, ISO, and other styles
50

Sanjay Kumar N V. "Performance Benchmarking of Redpanda an event streaming platform using SBK and Open Messaging Frameworks." Journal of Electrical Systems 20, no. 6s (April 29, 2024): 1148–61. http://dx.doi.org/10.52783/jes.2846.

Full text
Abstract:
Redpanda is an event-driven, publish/subscribe based messaging system. The two main performance benchmarking frameworks, Storage Benchmark Kit (SBK) and open messaging benchmark are used here to test the performance benchmarking of redpanda’s event processing system. SBK is a software framework for measuring any storage device or client's performance benchmarking and open messaging benchmark framework which supports benchmarking for a wide variety of messaging platforms. Well formulated fine tuning system for benchmarks is still an open research problem. In this paper, SBK and open messaging frameworks have been experimentally compared with respect to their messaging capabilities (Message sending and receiving throughput) on redpanda’s event processing system. A number of experimental test scenarios have been conducted on data streaming workload, the performance evaluation results revealed that SBK gives maximum throughput performance compared to open messaging benchmark.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography