472 research outputs found
Registered reports: an early example and analysis
© 2019 Wiseman et al.The recent ‘replication crisis’ in psychology has focused attention on ways of increasing methodological rigor within the behavioral sciences. Part of this work has involved promoting ‘Registered Reports’, wherein journals peer review papers prior to data collection and publication. Although this approach is usually seen as a relatively recent development, we note that a prototype of this publishing model was initiated in the mid-1970s by parapsychologist Martin Johnson in the European Journal of Parapsychology (EJP). A retrospective and observational comparison of Registered and non-Registered Reports published in the EJP during a seventeen-year period provides circumstantial evidence to suggest that the approach helped to reduce questionable research practices. This paper aims both to bring Johnson’s pioneering work to a wider audience, and to investigate the positive role that Registered Reports may play in helping to promote higher methodological and statistical standards.Peer reviewe
Self-other differences in student drinking norms research: the role of impression management, self-deception and measurement methodology
Background: Data-driven student drinking norms interventions are based on reported normative overestimation of the extent and approval of an average student’s drinking. Self-reported differences between personal and perceived normative drinking behaviors and attitudes are taken at face value as evidence of actual levels of overestimation. This study investigates whether commonly used data collection methods and socially desirable responding may inadvertently impede establishing 'objective' drinking norms.
Methods: UK students [N=421; 69% female; Mean age 20.22 years (SD = 2.5)] were randomly assigned to one of three versions of a drinking norms questionnaire: The standard multi-target questionnaire assessed respondents' drinking attitudes and behaviors (frequency of consumption, heavy drinking, units on a typical occasion) as well as drinking attitudes and behaviors for an ‘average student’. Two deconstructed versions of this questionnaire assessed identical behaviors and attitudes for participants themselves or an 'average student'. The Balanced Inventory of Desirable Responding was also administered.
Results: Students who answered questions about themselves and peers reported more extreme perceived drinking attitudes for the average student compared with those reporting solely on the ‘average student’. Personal and perceived reports of drinking behaviors did not differ between multi- and single-target versions of the questionnaire. Among those who completed the multi-target questionnaire, after controlling for demographics and weekly drinking, socially desirable responding was related positively with the magnitude of difference between students’ own reported behaviors/attitudes and those perceived for the average student.
Conclusions: Standard methodological practices and socially desirable responding may be sources of bias in peer norm overestimation research
False positives and other statistical errors in standard analyses of eye movements in reading
In research on eye movements in reading, it is common to analyze a number of canonical dependent measures to study how the effects of a manipulation unfold over time. Although this gives rise to the well-known multiple comparisons problem, i.e. an inflated probability that the null hypothesis is incorrectly rejected (Type I error), it is accepted standard practice not to apply any correction procedures. Instead, there appears to be a widespread belief that corrections are not necessary because the increase in false positives is too small to matter. To our knowledge, no formal argument has ever been presented to justify this assumption. Here, we report a computational investigation of this issue using Monte Carlo simulations. Our results show that, contrary to conventional wisdom, false positives are increased to unacceptable levels when no corrections are applied. Our simulations also show that counter-measures like the Bonferroni correction keep false positives in check while reducing statistical power only moderately. Hence, there is little reason why such corrections should not be made a standard requirement. Further, we discuss three statistical illusions that can arise when statistical power is low, and we show how power can be improved to prevent these illusions. In sum, our work renders a detailed picture of the various types of statistical errors than can occur in studies of reading behavior and we provide concrete guidance about how these errors can be avoided
Using Bayes Factors to evaluate evidence for no effect: examples from the SIPS project
Aims: To illustrate how Bayes Factors are important for determining the effectiveness of
interventions.
Method: We consider a case where inappropriate conclusions were publicly drawn based on
significance testing, namely the SIPS Project (Screening and Intervention Programme for Sensible
drinking), a pragmatic, cluster-randomized controlled trial in each of two healthcare settings and in
the criminal justice system. We showhow Bayes Factors can disambiguate the non-significant findings
from the SIPS Project and thus determine whether the findings represent evidence of absence or
absence of evidence. We show how to model the sort of effects that could be expected, and how to
check the robustness of the Bayes Factors.
Results: The findings from the three SIPS trials taken individually are largely uninformative but, when
data from these trials are combined, there is moderate evidence for a null hypothesis (H0) and thus
for a lack of effect of brief intervention compared with simple clinical feedback and an alcohol
information leaflet (B = 0.24, p = 0.43).
Conclusion: Scientists who find non-significant results should suspend judgment – unless they
calculate a Bayes Factor to indicate either that there is evidence for a null hypothesis (H0) over a (welljustified)
alternative hypothesis (H1), or else that more data are needed
CMS distributed computing workflow experience
The vast majority of the CMS Computing capacity, which is organized in a tiered hierarchy, is located away from CERN. The 7 Tier-1 sites archive the LHC proton-proton collision data that is initially processed at CERN. These sites provide access to all recorded and simulated data for the Tier-2 sites, via wide-area network (WAN) transfers. All central data processing workflows are executed at the Tier-1 level, which contain re-reconstruction and skimming workflows of collision data as well as reprocessing of simulated data to adapt to changing detector conditions. This paper describes the operation of the CMS processing infrastructure at the Tier-1 level. The Tier-1 workflows are described in detail. The operational optimization of resource usage is described. In particular, the variation of different workflows during the data taking period of 2010, their efficiencies and latencies as well as their impact on the delivery of physics results is discussed and lessons are drawn from this experience. The simulation of proton-proton collisions for the CMS experiment is primarily carried out at the second tier of the CMS computing infrastructure. Half of the Tier-2 sites of CMS are reserved for central Monte Carlo (MC) production while the other half is available for user analysis. This paper summarizes the large throughput of the MC production operation during the data taking period of 2010 and discusses the latencies and efficiencies of the various types of MC production workflows. We present the operational procedures to optimize the usage of available resources and we the operational model of CMS for including opportunistic resources, such as the larger Tier-3 sites, into the central production operation
Participant Nonnaiveté and the reproducibility of cognitive psychology
Many argue that there is a reproducibility crisis in psychology. We investigated nine well-known effects from the cognitive psychology literature—three each from the domains of perception/action, memory, and language, respectively—and found that they are highly reproducible. Not only can they be reproduced in online environments, but they also can be reproduced with nonnaïve participants with no reduction of effect size. Apparently, some cognitive tasks are so constraining that they encapsulate behavior from external influences, such as testing situation and prior recent experience with the experiment to yield highly robust effects
Analytic philosophy for biomedical research: the imperative of applying yesterday's timeless messages to today's impasses
The mantra that "the best way to predict the future is to invent it" (attributed to the computer scientist Alan Kay) exemplifies some of the expectations from the technical and innovative sides of biomedical research at present. However, for technical advancements to make real impacts both on patient health and genuine scientific understanding, quite a number of lingering challenges facing the entire spectrum from protein biology all the way to randomized controlled trials should start to be overcome. The proposal in this chapter is that philosophy is essential in this process. By reviewing select examples from the history of science and philosophy, disciplines which were indistinguishable until the mid-nineteenth century, I argue that progress toward the many impasses in biomedicine can be achieved by emphasizing theoretical work (in the true sense of the word 'theory') as a vital foundation for experimental biology. Furthermore, a philosophical biology program that could provide a framework for theoretical investigations is outlined
- …
