Skip to main content
🔬

Methodology & Theory

A curated collection of research papers focusing on methodology & theory. Explore the methodology, key findings, and ongoing debates in this field.

Total Papers 70
Year Range 1978 – 2025
Top Contributors
Kennedy, J.EKennedy, James EIoannidis, John P.A

Recent Publications

Paranormal belief, conspiracy endorsement, and positive wellbeing: a network analysis

Dagnall, Neil; Drinkwater, Kenneth Graham; Denovan, Andrew; Escolá Gascón, Alex 2025 Frontiers in Psychology

Using network analysis (EBICglasso) on a cross-sectional survey of 1,667 UK adults, interrelationships between paranormal belief (PB), conspiracy theory endorsement (CT), schizotypy, and positive wellbeing factors were examined. PB and Self-Esteem emerged as the two most central nodes. PB linked most strongly with CT (r = 0.60), cognitive-perceptual schizotypy (r = 0.56), search for meaning (r = 0.42), and avoidant coping (r = 0.40). Though correlated, PB and CT related differently to wellbeing: PB correlated positively with presence of meaning while CT did not, and CT correlated negatively with life satisfaction while PB did not. PB appeared to mediate relationships between schizotypy, meaning-seeking, and avoidant coping, while self-esteem bridged coping, meaning in life, and life satisfaction.

#paranormal_belief #conspiracy_theory #network_analysis #wellbeing #schizotypy

Planning Falsifiable Confirmatory Research

Kennedy, James E 2024 Psychological Methods

Falsifiable research requires that study designs can provide evidence a hypothesis is false as well as true. This article integrates power analysis, equivalence testing, Bayesian operating characteristics, and preregistration into a framework for falsifiable confirmatory research. Power >= .95 for a prespecified minimum effect size is optimal; .90 is good. If any nonzero effect is considered meaningful, the hypothesis is unfalsifiable. For d = 0.20, sample sizes of 858-1,084 are needed at adequate power. The alternative hypothesis can be rejected via noncentral t distributions when power is high. Preregistrations should specify criteria for evidence the hypothesis is true, false, or inconclusive. Retrospective meta-analyses are exploratory; prospective meta-analysis is preferred.

#falsifiability #statistical_power #preregistration #confirmatory_research #effect_size_estimation

Addressing Researcher Fraud: Retrospective, Real-Time, and Preventive Strategies — Including Legal Points and Data Management That Prevents Fraud

Kennedy, James E 2024 Frontiers in Research Metrics and Analytics

Policy and practice review analyzing three strategies for addressing researcher fraud: (1) retrospective investigations after allegations — common but vulnerable to post-hoc bias, low evidentiary standards, and inability to identify perpetrators; (2) real-time sting operations — conclusive when feasible, but rarely practicable; (3) preventive data management practices from FDA-regulated clinical trials — largely absent from academic research. Eight preventive practices are described: archiving raw data, audit trails, restricted data access, software validation, quality control, blinding with preregistered analysis programs, research audits, and public raw data. Proposes an 'error-controlled data management' badge for studies meeting these standards. AI is identified as a looming escalation of fraud sophistication.

#research_fraud #data_management #research_misconduct #open_science #replication_crisis

Quantum Aspects of the Brain-Mind Relationship: A Hypothesis with Supporting Evidence

Kauffman, Stuart A; Radin, Dean 2023 BioSystems

Proposes a non-substance dualism in which reality consists of ontologically real Possibles (Res potentia, following Heisenberg's 1958 concept) that do not obey Aristotle's law of the excluded middle, and ontologically real Actuals (Res extensa) that do. Mind acausally converts Possibles into Actuals through measurement, breaking the causal closure of classical physics and allowing consciousness an active role. This single hypothesis is claimed to explain five quantum mechanics mysteries. Reviews supporting evidence from ganzfeld experiments (30.6% hit rate, 8.1σ), forced-choice precognition (11σ), presentiment (5.3–6.9σ), dice PK (19σ), RNG experiments (6.8σ), and double-slit consciousness studies (cumulative p < 4 × 10⁻¹¹). Argues that quantum mind provides a basis for responsible free will.

#quantum_consciousness #res_potentia #mind_matter_interaction #non_substance_dualism #free_will

Paranormal beliefs and cognitive function: A systematic review and assessment of study quality across four decades of research

Dean, Charlotte E; Akhtar, Shazia; Gale, Tim M; Irvine, Karen; Grohmann, Dominique; Laws, Keith R 2022 PLoS ONE

Systematic review of 71 studies (N=20,993) examining associations between paranormal beliefs and cognitive functioning published 1980-2021. Study quality was assessed using the AXIS tool. Results show 75% of studies rated good-to-strong quality, with quality improving over time. Most consistent findings: paranormal belief associated with increased intuitive thinking style (8/8 studies), increased confirmatory bias, poorer conditional reasoning, and reduced perception of randomness. Two-thirds of studies document poorer cognitive performance in believers. Major methodological weaknesses identified: only 7% included a-priori power analyses, 3% preregistered, 63% used undergraduate samples, 17% corrected for multiple comparisons. Authors propose a fluid-executive model suggesting disparate cognitive deficits may reflect common underlying executive dysfunction.

#paranormal_beliefs #cognitive_function #systematic_review #study_quality #cognitive_differences

Experimental evidence of non-classical brain functions

Kerskens, Christian Matthias; López Pérez, David 2022 Journal of Physics Communications

Using proposals from quantum gravity that entanglement between two known quantum systems can witness non-classicality in a mediating system, this study applied zero quantum coherence (ZQC) NMR protocols to 40 human volunteers at rest in a 3T MRI scanner. Evoked signal bursts resembling heartbeat-evoked potentials appeared in most brain regions with up to 15% signal increases. These signals showed definitive ZQC characteristics (magic angle dependency R²=0.9958, optimal flip angle at 45° R²=0.9964, MTC immunity), had no classical NMR contrast correlates, surpassed classical bounds, and depended on conscious awareness—declining during sleep. The authors conclude these constitute experimental evidence of non-classical (quantum) brain functions.

#quantum_coherence #nmr_mri #consciousness #entanglement_witness #heartbeat_evoked_potentials

Self-Ascribed Paranormal Ability: Reflexive Thematic Analysis

Drinkwater, Kenneth Graham; Dagnall, Neil; Walsh, Stephen; Sproson, Lisa; Peverell, Matthew; Denovan, Andrew 2022 Frontiers in Psychology

Investigated how 12 individuals with self-ascribed paranormal abilities (mediums, psychics, sensitives, clairvoyants) perceive and make sense of their professed powers through semi-structured interviews analyzed with reflexive thematic analysis (RTA). Four themes emerged: Formative Influences (childhood experiences and gifted family members), (Inter)Subjective Paranormal Experience (transcendental/mystic encounters and ESP), Embodied Processes (sense of control over abilities), and Perception of Reality (self-awareness and surreal perceptions). These themes mapped onto lifeworld fractions of temporality, inter-subjectivity, embodiment, and spatiality. Self-ascription operates as an attributional process where individuals construct coherent narratives to contextualize and validate their claimed powers, with belief reciprocally reinforcing paranormal interpretation of anomalous experiences.

#qualitative_research #reflexive_thematic_analysis #self_ascribed_ability #paranormal_belief #phenomenology

The Community Assessment of Psychic Experiences-Positive scale (CAPE-P15) accurately classifies and differentiates psychotic experience levels in adolescents from the general population

Núñez, D; Godoy, M. I; Gaete, J; Faúndez, M. J; Campos, S; Fresno, A; Spencer, R 2021 PLoS ONE

Psychometric validation of the CAPE-P15—a 15-item self-report scale measuring psychosis-spectrum experiences (paranoid ideation, bizarre experiences, perceptual anomalies)—in 1,594 Chilean adolescents (ages 12–19). Using Item Response Theory, item discrimination parameters ranged from moderate (α=1.05) to very high (α=2.44); scale reliability was high (α=0.94, ω=0.81). Both unidimensional and hierarchical models fit acceptably; the test information function was most accurate between −1 and +2.5 SD. Items assessing auditory/visual hallucinations and being controlled by external forces showed highest discriminative power. Higher PE scores correlated significantly with depression, anxiety, suicidal ideation, defeat, entrapment, and rumination.

#psychometric_validation #psychosis_screening #cape_p15 #item_response_theory #subclinical_psychosis

Why Most Research Findings About Psi Are False: The Replicability Crisis, the Psi Paradox and the Myth of Sisyphus

Rabeyron, Thomas 2020 Frontiers in Psychology

Argues that psi research faces a fundamental epistemological paradox: if psi exists, it violates the observer-independence assumption underlying the scientific method, meaning that proving psi retroactively invalidates the experimental framework used to prove it. Using the Ganzfeld protocol and Bem's (2011) precognition experiments as case studies, traces an infinite logical loop where successful psi demonstrations undermine their own evidential basis. Connects this paradox to the broader replication crisis in psychology, the decline effect, and proposes Lucadou's Model of Pragmatic Information as a framework for moving beyond classical experimental approaches.

#replicability_crisis #psi_paradox #epistemology #decline_effect #questionable_research_practices

Measuring Extraordinary Experiences and Beliefs: A Validation and Reliability Study

Wahbeh, Helané; Yount, Garret; Vieten, Cassandra; Radin, Dean; Delorme, Arnaud 2019 F1000Research

Development and psychometric validation of the 20-item Noetic Experience and Belief Scale (NEBS), which separately measures paranormal belief (10 items) and paranormal experience (10 items) on 0–100 visual analog scales. Study 1 administered the NEBS to 361 U.S. general-population adults (96 retested at one month), demonstrating strong internal consistency (Cronbach's α = 0.90 Belief, 0.93 Experience), high test-retest reliability (Belief r = 0.83, Experience r = 0.77), convergent validity with three established paranormal belief scales, and a good-fitting two-factor confirmatory model (RMSEA = 0.060, CFI = 0.94). Study 2 confirmed the factor structure in 646 IONS Discovery Lab participants and established divergent validity with negligible correlations to health, personality, and affect measures. The NEBS is the first validated instrument treating belief and experience as distinct but correlated constructs (r = 0.77 general population, r = 0.64 IONS sample).

#psychometrics #paranormal_belief #paranormal_experience #scale_validation #confirmatory_factor_analysis

False-Positive Effect in the Radin Double-Slit Experiment on Observer Consciousness as Determined With the Advanced Meta-Experimental Protocol

Walleczek, Jan; von Stillfried, Nikolaus 2019 Frontiers in Psychology

A commissioned conceptual replication of the Radin double-slit experiment on observer consciousness employed the Advanced Meta-Experimental Protocol (AMP), which pairs true-experiments (with test subjects) against sham-experiments (without test subjects) across eight pre-specified test categories (10,000 total trials). The replication failed to confirm the original anomalous consciousness effect. The sham-experiment identified a statistically significant false-positive effect (p = 0.021, sigma = -2.02, N = 1,250) in exactly the test category predicted for a true-positive consciousness effect. The false-positive effect size (0.016%) was approximately 10 times larger than the claimed consciousness effect (0.001%), and its statistical significance was comparable to that reported in the original study.

#double_slit_experiment #false_positive_detection #systematic_methodological_error #confirmatory_research_design #sham_experiment

Future Directions in Meditation Research: Recommendations for Expanding the Field of Contemplative Science

Vieten, C; Wahbeh, H; Cahn, B.R; MacLean, K; Estrada, M; Mills, P; Murphy, M; Shapiro, S; Radin, D.I; Josipovic, Z; Presti, D.E; Sapiro, M; Bays, J.C; Russell, P; Vago, D; Travis, F; Walsh, R; Delorme, A 2018 PLOS ONE

A cross-sectional online survey of 1,120 meditators (from 66 countries, mean 14.7 years practice) investigating the prevalence of extraordinary, mystical, and anomalous experiences during or related to meditation. Developed by a 4-year IONS task force of meditation researchers and teachers. Using the Revised Mystical Experience Questionnaire (MEQ30) and newly developed items, found high prevalence of mystical experiences (MEQ30 subscales 3.26-3.71/5.0) and extraordinary phenomena: altered awareness (91%), synchronicities (82%), clairvoyance/telepathy (56%), external physical phenomena (31%), and disturbing emotions (32%). Clairvoyance/telepathy showed the strongest correlation with meditation practice length (r = .30). Provides recommendations for expanding meditation research into 6 under-studied domains: mystical/transcendent experiences, social/relational aspects, physical/perceptual phenomena, spatial/temporal phenomena, extended perception, and difficult experiences.

#meditation #anomalous_experiences #survey #mystical_experiences #clairvoyance

Classic Hallucinogens and Mystical Experiences: Phenomenology and Neural Correlates

Barrett, Frederick S; Griffiths, Roland R 2018 Current Topics in Behavioral Neurosciences

Reviews the phenomenology, measurement, and neural correlates of mystical experiences occasioned by classic hallucinogens. Under double-blind conditions, psilocybin (30 mg/70 kg) produced complete mystical experiences in 57-67% of participants, with effects persisting at 14-month follow-up. The psychometrically validated 30-item Mystical Experience Questionnaire (MEQ30) captures four factors: Mystical, Positive Mood, Transcendence of Time and Space, and Ineffability. Mystical experience scores mediated therapeutic outcomes in addiction and cancer-related distress. Proposes a functional neural model in which disruption of the default mode network — particularly decreased activity in the medial prefrontal cortex, posterior cingulate cortex, and angular gyrus — underlies the unity and timelessness central to mystical experience, drawing parallels with meditation neuroimaging.

#psilocybin #mystical_experience #default_mode_network #meq30_validation #psychedelic_therapy

Commentary: Reproducibility in Psychological Science: When Do Psychological Phenomena Exist?

Heino, Matti T. J; Fried, Eiko I; LeBel, Etienne P 2017 Frontiers in Psychology

Short commentary responding to Iso-Ahola (2017) who argued that falsifiability and replication are of secondary importance to scientific progress and that psychological phenomena are inherently not fully reproducible because humans are irreducibly complex. Heino, Fried, and LeBel counter that complexity should motivate more sophisticated study designs (person-level time series, within-person repeated measures) rather than abandoning replicability standards. They argue that falsification is what makes science self-correcting, and that without rigorous direct replications psychology turns into astrology. They cite successful complexity science in ecology, biology, and physics as evidence that complexity and replicability are compatible.

#replication #falsifiability #open_science #complex_systems #replicability_crisis

Experimenter Fraud: What Are Appropriate Methodological Standards?

Kennedy, J.E 2017 Journal of Parapsychology

Drawing on firsthand experience exposing the W. J. Levy fraud at J. B. Rhine's lab and two decades of work in FDA-regulated pharmaceutical research, Kennedy argues that standard post hoc investigations of fraud are uniquely ineffective in parapsychology because statistical anomalies indicative of data manipulation can be explained as psi effects. The Levy case required direct covert detection during an ongoing experiment — a 'sting operation' — to produce compelling evidence. Kennedy advocates adopting the pharmaceutical research standard that fraud by one experimenter should be very difficult or impossible, implemented through duplicate data records, dual-experimenter oversight, and documented software validation. This systematic prevention approach would eliminate the vast majority of fraud cases and is preferable to reliance on after-the-fact accusations and investigations.

#experimenter_fraud #research_integrity #levy_case #fraud_prevention #confirmatory_research

Equivalence Tests: A Practical Primer for t Tests, Correlations, and Meta-Analyses

Lakens, Daniël 2017 Social Psychological and Personality Science

Scientists should be able to provide support for the absence of a meaningful effect, but nonsignificant p-values cannot establish this. This tutorial introduces the two one-sided tests (TOST) procedure for equivalence testing: researchers specify upper and lower equivalence bounds based on the smallest effect size of interest (SESOI), then test whether observed effects fall within this range. Formulas and worked examples are provided for independent/dependent t-tests, correlations, and meta-analyses. An accompanying spreadsheet and R package (TOSTER) enable psychologists to perform equivalence tests and power analyses. Adopting equivalence testing prevents misinterpreting nonsignificant results as evidence for the null, enables replication studies to test for absence of meaningful effects, and encourages researchers to specify which effect sizes they find theoretically worthwhile.

#equivalence_testing #TOST #statistical_inference #power_analysis #SESOI

The Varieties of Contemplative Experience: A Mixed-Methods Study of Meditation-Related Challenges in Western Buddhists

Lindahl, Jared R; Fisher, Nathan E; Cooper, David J; Rosen, Rochelle K; Britton, Willoughby B 2017 PLoS ONE

Mixed-methods investigation of meditation-related experiences in Western Buddhist practitioners, with deliberate focus on underreported challenging, difficult, distressing, or functionally impairing effects. Qualitative interviews with 60 practitioners (20 each from Theravāda, Zen, Tibetan traditions) and 32 experts yielded a taxonomy of 59 meditation-related experiences across 7 domains (cognitive, perceptual, affective, somatic, conative, sense of self, social) and 26 influencing factors across 4 domains (practitioner-level, practice-level, relationships, health behaviors). Valence ranged from very positive to very negative; 73% reported moderate-severe impairment, 17% suicidality, 17% hospitalization. Causality assessment met WHO/FDA criteria (mean 4.2/6). Findings challenge assumptions that meditation difficulties only occur in vulnerable populations or intensive practice contexts.

#meditation_adverse_effects #contemplative_experience #mixed_methods #buddhist_meditation #phenomenology_taxonomy

Decline Effects: Types, Mechanisms, and Personal Reflections

Protzko, John; Schooler, Jonathan W 2017 Psychological Science Under Scrutiny: Recent Challenges and Proposed Solutions (Wiley)

A taxonomy of four types of declining effect sizes in science: false positive (no true effect; e.g., Mozart effect), inflated (true effect exaggerated by small N and selective reporting), under-specified (true effect but boundary conditions unknown), and genuinely decreasing (true effect diminishes over time with variables held constant). General mechanisms include underpowered studies, publication bias, and selective reporting. Schooler reports his verbal overshadowing effect declined from ~25% impairment to markedly smaller effects in a 30-lab registered replication. Protzko argues most genuinely decreasing effects are Type I errors of a mega-analytic framework. Both advocate pre-registration and prospective multi-site replication.

#decline_effect #replication_crisis #meta_science #verbal_overshadowing #observer_effects

Options for Prospective Meta-Analysis and Introduction of Registration-Based Prospective Meta-Analysis

Watt, Caroline A; Kennedy, James E 2017 Frontiers in Psychology

Retrospective meta-analyses resemble exploratory rather than confirmatory research because inclusion criteria, statistical methods, and moderator decisions are made after analysts know study outcomes. Three options for prospective meta-analysis are compared: (1) preregistered plans, still allowing retrospective decisions about unanticipated variations; (2) preplanned multi-center projects, optimal but requiring funding rarely available in behavioral science; (3) registration-based prospective meta-analysis, where inclusion/exclusion of independently-initiated studies is decided at pre-registration, before data collection. An exemplar applying Option 3 to ganzfeld ESP studies at the Koestler Parapsychology Unit illustrates the approach.

#prospective_meta_analysis #study_registration #confirmatory_research #ganzfeld_exemplar #researcher_bias

Testing for Questionable Research Practices in a Meta-Analysis: An Example from Experimental Parapsychology

Bierman, Dick J; Spottiswoode, James P; Bijl, Aron 2016 PLOS ONE

Using Monte Carlo simulations and a genetic algorithm, a method was developed to quantify the impact of Questionable Research Practices (QRPs) on meta-analytic results. Applied to 78 post-1985 Ganzfeld telepathy experiments (3,494 trials, mean hit rate 31% vs. 25% chance), seven QRPs were modeled at prevalence rates from published surveys of psychologists. With realistic QRP parameters and no anomalous effect, simulations failed to reproduce the empirical database (F=10.15, p<0.05). Allowing a 2% excess hit rate yielded acceptable fit (F=1.79, p=0.47). QRPs explain approximately 60% of the reported effect size, but a residual effect remains significant (p=0.003).

#questionable_research_practices #ganzfeld_telepathy #monte_carlo_simulation #publication_bias #genetic_algorithm

Is the Methodological Revolution in Psychology Over or Just Beginning?

Kennedy, J.E 2016 Journal of Parapsychology

Prompted by Bem's (2011) 'feeling the future' experiments, psychology recognized the need for pre-registered confirmatory research. Eight additional methodological deficiencies remain unaddressed in both fields: deficient study registration, bias from dropouts and incomplete data, lack of software validation, absence of fraud prevention measures, inappropriate statistical methods for confirmatory studies, overlooked Bayesian inferential errors, weaknesses of retrospective meta-analysis versus prospective meta-analysis, and statistical dependence problems in outcome variables. Drawing on regulated medical research standards (FDA clinical trial guidance), specific practices are recommended for each deficiency. When confronted with the choice between psi and methodological flaws, psychologists will inevitably choose reform — making proactive adoption of these standards the efficient path forward.

#research_methodology #confirmatory_research #pre_registration #statistical_power #software_validation

Integrated Information Theory: From Consciousness to Its Physical Substrate

Tononi, Giulio; Boly, Melanie; Massimini, Marcello; Koch, Christof 2016 Nature Reviews Neuroscience

Integrated information theory (IIT) starts from five phenomenological axioms — intrinsic existence, composition, information, integration, and exclusion — and derives postulates about the physical substrate of consciousness (PSC). The PSC is the complex of neural elements specifying a conceptual structure with maximum integrated information (Phi_max). IIT explains why the cortex supports consciousness but the cerebellum (with 4x more neurons) does not, why consciousness fades during slow-wave sleep despite continued neural activity, and predicts that consciousness can split in split-brain patients. The perturbational complexity index (PCI), validated across sleep, anesthesia, and brain damage, serves as a practical proxy for Phi_max.

#integrated_information_theory #consciousness_theory #neural_correlates #perturbational_complexity_index #phi_max

Estimating the Reproducibility of Psychological Science

Open Science Collaboration 2015 Science

A collaborative effort by 270 researchers replicated 100 experimental and correlational studies from three leading psychology journals (2008 issues of Psychological Science, JPSP, and JEP:LMC) using pre-registered, high-powered designs with original materials. While 97% of original studies reported significant results, only 36% of replications achieved significance. Mean replication effect sizes (r = 0.197) were half the original magnitudes (r = 0.403). Cognitive psychology findings replicated better (50%) than social psychology (25%). Replication success correlated with original evidence strength rather than replication team characteristics, implicating publication bias and analytic flexibility as likely contributors to inflated original effects.

#replication_crisis #reproducibility #open_science #pre_registration #publication_bias

Lessons from the First Two Years of Operating a Study Registry

Watt, Caroline; Kennedy, James E 2015 Frontiers in Psychology

Opinion article on the first two years of the Koestler Parapsychology Unit study registry at the University of Edinburgh (opened fall 2012). Modeled on clinical trial standards (ICMJE 2005), the registry requires public, prospective, irreversibly public registration with independent review. Key recommendations: classify each hypothesis as exploratory or confirmatory; pre-specify all analysis decisions for confirmatory research; independently review registrations for completeness. Virtually all initial submissions had deficiencies. Critiques OSF registration for allowing experimenters to keep or revert registrations to private after viewing results, undermining the file-drawer protection that registration is meant to provide.

#study_registration #pre_registration #confirmatory_research #exploratory_research #kpu_registry

Consciousness: here, there and everywhere?

Tononi, Giulio; Koch, Christof 2015 Philosophical Transactions of the Royal Society B: Biological Sciences

Integrated Information Theory (IIT 3.0) is presented as a principled framework for understanding consciousness. Starting from five phenomenological axioms — intrinsic existence, composition, information, integration, and exclusion — IIT derives postulates about the physical substrates required for experience. The theory identifies consciousness with maximally irreducible integrated information (Phi_max), explaining why the cerebral cortex supports consciousness while the cerebellum, despite having four times as many neurons, does not. IIT predicts that feed-forward networks are unconscious regardless of computational sophistication, that digital simulations of conscious brains would not themselves be conscious, and that consciousness is graded and widespread among biological organisms. The perturbational complexity index (PCI), inspired by IIT, reliably tracks consciousness across sleep, anaesthesia, and brain injury.

#integrated_information_theory #hard_problem_consciousness #neural_correlates_consciousness #panpsychism #phi_measure

Small Telescopes: Detectability and the Evaluation of Replication Results

Simonsohn, Uri 2015 Psychological Science

Introduces the "small telescopes" approach for evaluating replication results: instead of asking whether a replication effect is significantly different from zero or from the original estimate, the method tests whether it is significantly smaller than d33% — the effect size giving the original study only 33% power. Applied to three replication disputes, the approach shows that underpowered replications (e.g., Gamez et al. with 44% power) can be uninformatively noisy, while large well-powered replications can reject the original design's adequacy even when the replication finds a significant effect. A simple rule emerges: replications need approximately 2.5x the original sample size for 80% power to reject d33%.

#replication_evaluation #statistical_power #effect_size_estimation #publication_bias #detectability

Bayesian and Classical Hypothesis Testing: Practical Differences for a Controversial Area of Research

Kennedy, J.E 2014 Journal of Parapsychology

Conceptual analysis comparing Bayesian and classical (frequentist) hypothesis testing for controversial research areas such as parapsychology. Describes the philosophical differences (objective vs subjective probability), mathematical models, and practical implications of each approach. Demonstrates that a uniform Bayesian prior can bias against small effects: 5100 hits in 10000 trials yields classical p=.046 but Bayesian BF=10.8 for the null. Recommends the FDA guidance on Bayesian clinical trials as a framework for confirmatory psi experiments: specify priors, acceptance criteria, operating characteristics (Type I error and power), and relative roles of priors vs data prospectively. Argues that both approaches are valid when properly applied, but psychology and parapsychology lack confirmatory methodology, undermining both statistical traditions.

#bayesian_analysis #confirmatory_research #hypothesis_testing #inferential_errors #prior_probability

Registered Reports: A Method to Increase the Credibility of Published Results

Nosek, Brian A; Lakens, Daniël 2014 Social Psychology

This editorial introduces Registered Reports, a novel publishing format that incorporates peer review and preregistration of designs before data collection. The authors present the first known journal issue in any discipline consisting exclusively of pre-registered replication studies: a special issue of Social Psychology with 15 articles. The process involved issuing calls for proposals (36 received), peer review of full proposals (24 encouraged, 14 accepted), and OSF registration prior to data collection with publication guaranteed irrespective of results. This approach shifts incentives to evaluate methodological quality rather than results, addressing the publication bias against replications and negative results that undermines scientific credibility.

#registered_reports #pre_registration #replication #methodology_reform #open_science

Parapsychological Phenomena as Examples of Generalized Nonlocal Correlations—A Theoretical Framework

Walach, Harald; von Lucadou, Walter; Römer, Hartmann 2014 Journal of Scientific Exploration

Argues that parapsychological phenomena lack scientific acceptance not merely due to insufficient evidence but because no adequate theoretical framework connects them to mainstream science. Proposes Generalized Quantum Theory (GQT), which extends quantum formalism to any system requiring incompatible/complementary observables. Derives three conditions for generalized entanglement correlations (GET): a system with identifiable subsystems, a global observable complementary to local observables, and an entangled state. Applies the framework to telepathy, healing, clairvoyance, psychokinesis, and precognition. The No-Transmission (NT) axiom predicts that classical experimental designs will fail in exact replications because they attempt to code nonlocal correlations as signals — explaining the decline effect in PK meta-analyses and repeated replication failures.

#generalized_quantum_theory #nonlocal_correlations #entanglement #nt_axiom #decline_effect

Anomalous Experiences, Psi, and Functional Neuroimaging

Acunzo, David J; Evrard, Renaud; Rabeyron, Thomas 2013 Frontiers in Human Neuroscience

Opinion article reviewing six functional neuroimaging studies testing the psi hypothesis: Standish et al. (2003), Richards et al. (2005), Achterberg et al. (2005), Venkatasubramanian et al. (2008), Bierman & Scholte (2002), and Moulton & Kosslyn (2008). Five of six reported psi-consistent results; only Moulton & Kosslyn (2008) reported null results. Identifies four categories of methodological weaknesses: inadequate counter-balancing, improper trial randomization, insufficient information shielding, and small sample sizes. Concludes that only Bierman & Scholte (2002) and Moulton & Kosslyn (2008) are methodologically sound, and no firm conclusions about psi can be drawn from the neuroimaging corpus.

#neuroimaging #fmri #methodology_critique #counter_balancing #randomization

Power failure: why small sample size undermines the reliability of neuroscience

Button, Katherine S; Ioannidis, John P.A; Mokrysz, Claire; Nosek, Brian A; Flint, Jonathan; Robinson, Emma S. J; Munafò, Marcus R 2013 Nature Reviews Neuroscience

Low statistical power in neuroscience studies reduces both the chance of detecting true effects and the probability that significant findings reflect genuine effects (positive predictive value). Analysis of 49 neuroscience meta-analyses (730 studies published in 2011) reveals median power of 21%, falling to 18% when high-power neurological outliers are excluded. Brain volume studies show median power of just 8%. An excess significance test confirms more significant results than expected (349 vs. 254, p < 0.0001), indicating reporting biases. The winner's curse further inflates initial effect estimates by 25-50% at typical power levels. These problems undermine reproducibility and waste resources, including animal lives in preclinical research. Recommendations include a priori power calculations, pre-registration, transparent reporting, data sharing, and large-scale collaborative replication.

#statistical_power #replication_crisis #effect_size_inflation #publication_bias #positive_predictive_value

Can Parapsychology Move Beyond the Controversies of Retrospective Meta-Analyses?

Kennedy, J.E 2013 Journal of Parapsychology

Argues that retrospective meta-analyses have failed to resolve parapsychological controversies because they aggregate substantially underpowered studies. Compiles replication rates across major meta-analyses (Table 1), finding only 20-33% of well-conducted studies obtain significant results versus the 0.80 standard. A ganzfeld experiment needs N=201 for adequate power, yet the median is 40 trials (power 0.22). RNG meta-analyses consistently show z independent of sample size, contrary to statistical theory. Monte Carlo simulations confirm small-study effects in early ganzfeld data. Proposes goal-oriented psi experimenter effects as a parsimonious explanation and recommends prospective registration, adequate power, and best-evidence synthesis.

#meta_analysis_critique #statistical_power #replication_rates #experimenter_effects #goal_oriented_psi

The CEMI Field Theory: Closing the Loop

McFadden, Johnjoe 2013 Journal of Consciousness Studies

Synchronous neuronal firing is the strongest known correlate of attention, awareness, and consciousness, yet when EM field theories of consciousness were first proposed around 2002 there was no direct evidence that synchrony was functional or that endogenous EM fields influenced neuron firing. Three key studies have since closed that evidential gap: Fujisawa et al. (2004) showed externally applied 40 Hz fields modulate hippocampal neuron firing; Frohlich & McCormick (2010) demonstrated that endogenous fields as weak as 0.25 mV/mm entrain neocortical oscillations; and Anastassiou et al. (2011) proved ephaptic coupling causes phase-locked synchronous firing. Together these establish a feedback loop — neurons generate coherent EM fields that in turn recruit more neurons into synchrony — supporting the CEMI field theory's claim that the brain's EM field is the physical substrate of consciousness.

#consciousness_theory #electromagnetic_field #neural_synchrony #binding_problem #communication_through_coherence

Harnessing Repetitive Behaviours to Engage Attention and Learning in a Novel Therapy for Autism: An Exploratory Analysis

Chen, Grace Megumi; Yoder, Keith Jonathon; Ganzel, Barbara Lynn; Goodwin, Matthew S; Belmonte, Matthew Kenneth 2012 Frontiers in Psychology

Video-coded analysis of RPM (Rapid Prompting Method) therapy sessions with 9 non-speaking autistic children (ages 8-14, CARS 42.5-50). Coders blind to session order rated middle 10-minute segments of sessions 1, 2, 4, and 8 for repetitive/stereotypic behaviours, gaze, response rate, choice complexity, and prompting. Mixed-effects model controlling for age found RSB incidence declined across sessions (b=-0.011, p=0.045). Direct gaze was negatively correlated with task success in 8/9 subjects. Therapist prompt rate strongly predicted response rate (b=0.480, p=0.00004). Choice complexity increased across sessions while success rate did not decline. Validity of communications was explicitly not tested.

#rapid_prompting_method #autism_therapy #repetitive_behaviors #attention_gaze #facilitated_communication_alternative

Measuring the Prevalence of Questionable Research Practices With Incentives for Truth Telling

John, Leslie K; Loewenstein, George; Prelec, Drazen 2012 Psychological Science

Survey of 5,964 academic psychologists (N=2,155 respondents, 36% response rate) measured prevalence of questionable research practices (QRPs) using Bayesian Truth Serum (BTS) incentives for truthful disclosure. Admissions were surprisingly high: 94% of BTS respondents admitted at least one QRP, including failing to report all dependent measures (66.5%), collecting more data after checking significance (58%), and selectively reporting studies (50%). BTS incentives raised admissions most for less-defensible practices. Geometric-mean estimates suggest ~1 in 10 psychologists has falsified data. Items formed approximate Guttman scale (reproducibility=0.80). Findings suggest QRPs may constitute the de facto scientific norm, with researchers rationalizing borderline behaviors as defensible.

#questionable_research_practices #survey_methodology #scientific_integrity #replication_crisis #p_hacking

Editors' Introduction to the Special Section on Replicability in Psychological Science: A Crisis of Confidence?

Pashler, Harold; Wagenmakers, Eric-Jan 2012 Perspectives on Psychological Science

An editorial introduction to the Perspectives on Psychological Science special section on replicability, chronicling the 'crisis of confidence' that unfolded in 2011-2012. Three catalysts are identified: the Stapel fraud case, Bem's (2011) ESP publication followed by widespread mockery, and Simmons et al.'s (2011) demonstration that flexible data analysis produces false-positive rates far exceeding 5%. The editors note 2012 brought further evidence of the problem: QRP prevalence surveys, suspicious clustering of p-values just below .05, and acrimonious disputes over failed social priming replications. The special section's 15 articles span diagnosis and treatment, from replication failures to Bayesian reanalyses, pre-registration, and data sharing.

#replication_crisis #publication_bias #questionable_research_practices #bem_controversy #editorial

Replication Unreliability in Psychology: Elusive Phenomena or "Elusive" Statistical Power?

Tressoldi, Patrizio E 2012 Frontiers in Psychology

Analyzes whether the unreliability of replication in controversial psychological phenomena stems from insufficient statistical power rather than non-existence of the effects. Retrospective power analysis of meta-analyses covering four phenomena — subliminal semantic priming (d = 0.47–0.80), incubation effect (d = 0.29), unconscious thought theory (d = 0.22), and non-local perception across three protocols (d = 0.011–0.16) — reveals that except for semantic priming on categorization (power = 0.96), the typical study has power far below 0.90. Forced-choice ESP studies (d = 0.011) would require N = 3,450 participants for adequate power, yet average N = 128. Recommends alternatives to NHST including confidence intervals, equivalence testing, and Bayesian approaches.

#statistical_power #replication #nhst #effect_size #non_local_perception

An Agenda for Purely Confirmatory Research

Wagenmakers, Eric-Jan; Wetzels, Ruud; Borsboom, Denny; van der Maas, Han L. J; Kievit, Rogier A 2012 Perspectives on Psychological Science

Argues that the pervasive confusion between exploratory and confirmatory research threatens psychological science. Proposes preregistration as the primary remedy: researchers publicly post a detailed analysis plan before testing any participants, and only pre-specified analyses qualify as confirmatory with valid statistical inference. Illustrates the proposal with a preregistered Bayesian replication of Bem's (2011) precognition experiment (N=100 women, two sessions of 60 forced-choice trials each). The replication found BF01 = 16.6 (default prior) and BF01 = 6.2 (Bem's knowledge-based prior), both strongly favoring the null hypothesis of no precognition.

#preregistration #confirmatory_vs_exploratory #bayesian_hypothesis_testing #replication_crisis #bem_replication

Neural correlates of the psychedelic state as determined by fMRI studies with psilocybin

Carhart-Harris, Robin L; Erritzoe, David; Williams, Tim; Stone, James M; Reed, Laurence J; Colasanti, Alessandro; Tyacke, Robin J; Leech, Robert; Malizia, Andrea L; Murphy, Kevin; Hobden, Peter; Evans, John; Feilding, Amanda; Wise, Richard G; Nutt, David J 2012 Proceedings of the National Academy of Sciences

Using arterial spin labeling perfusion and BOLD fMRI, two groups of 15 healthy volunteers were scanned during IV infusion of psilocybin (2 mg) versus saline placebo in a task-free protocol. Psilocybin produced only decreases in cerebral blood flow and BOLD signal, with no increases in any region. Decreases were maximal in default mode network hub regions — thalamus, anterior and posterior cingulate cortex, and medial prefrontal cortex. Greater ACC blood flow decreases predicted more intense subjective effects (r = −0.55, p = 0.017). A pharmaco-physiological interaction analysis revealed significantly decreased positive coupling between mPFC and PCC. The authors conclude that psychedelic consciousness arises from decreased activity and connectivity in connector hubs, enabling unconstrained cognition — consistent with Huxley's 'reducing valve' metaphor.

#fmri #psilocybin #default_mode_network #altered_states_of_consciousness #cerebral_blood_flow

Why Science Is Not Necessarily Self-Correcting

Ioannidis, John P.A 2012 Perspectives on Psychological Science

Self-correction is widely assumed to be a defining hallmark of science, but how often does it actually occur? Reviewing empirical evidence from psychology and biomedicine, Ioannidis argues that self-correction requires active replication effort — yet only ~1% of psychology papers are replications, fewer than 0.2% are independent direct replications, and most yield confirming results (perpetuated fallacies). With average power of 35% and modest bias, unchallenged fallacies may constitute 30–95% of published significant findings. A taxonomy of six discovery–replication paradigms quantifies the problem. Thirteen impediments to self-correction are catalogued, including publication bias, selective reporting, underpowered studies, and editorial bias against replication. Proposed reforms each carry unintended risks unless pursuit of truth remains the overriding priority.

#replication_crisis #self_correction #publication_bias #scientific_credibility #questionable_research_practices

Scientific Utopia: II. Restructuring Incentives and Practices to Promote Truth Over Publishability

Nosek, Brian A; Spies, Jeffrey R; Motyl, Matt 2012 Perspectives on Psychological Science

Publication norms in academic science emphasize novel, positive results, creating incentives that inflate false-positive rates and discourage replication. Drawing on the authors' own failed replication of a provocative embodied-cognition finding (original p = .01, N = 1,979; replication p = .59, N = 1,300), the paper catalogues nine common practices that increase publishability at the expense of accuracy, including optional stopping, selective reporting, HARKing, and avoidance of direct replication. Existing remedies (negative-results journals, education campaigns, reviewer vigilance) are judged insufficient. The proposed solutions restructure incentives around open data, open methods, open workflow with pre-registration, post-publication review, and Replication Value metrics, arguing that transparency and accountability can make the abstract accuracy motive competitive with the concrete publication motive.

#publication_bias #questionable_research_practices #open_science #replication_crisis #incentive_reform

Must Psychologists Change the Way They Analyze Their Data?

Bem, Daryl J; Utts, Jessica; Johnson, Wesley O 2011 Journal of Personality and Social Psychology

A reply to Wagenmakers et al. (2011), who argued that a Bayesian reanalysis of Bem's (2011) nine precognition experiments yields no evidence for psi. Bem, Utts, and Johnson contend that Wagenmakers et al.'s diffuse Cauchy prior is unrealistic, placing 57% probability on effect sizes >= 0.8 and triggering the Lindley-Jeffreys paradox. Using a knowledge-based normal prior (90th percentile of |d| at 0.5, informed by known psychological and psi effect sizes), the combined Bayes factor across all nine experiments is 13,669 with posterior P(H0) = 7.3 x 10^-5 — extreme evidence for psi. The authors argue Bayesian methods are valuable but contain hidden traps when priors are poorly specified.

#bayesian_statistics #prior_specification #lindley_jeffreys_paradox #bayes_factor #frequentist_vs_bayesian

Temporal patterns of happiness and information in a global social network: Hedonometrics and Twitter

Dodds, Peter Sheridan; Harris, Kameron Decker; Kloumann, Isabel M; Bliss, Catherine A; Danforth, Christopher M 2011 PLoS ONE

A tunable, real-time, text-based 'hedonometer' was constructed using 10,222 English words rated for happiness (1-9 scale) by 50 Amazon Mechanical Turk evaluators each, validated against the established ANEW word set (Spearman r_s = 0.944, p < 10^-10). Applied to approximately 46 billion words from 4.6 billion tweets by over 63 million users across 33 months, the instrument revealed robust temporal patterns: a weekly cycle with Saturday happiest (h_avg ~ 6.06) and Tuesday least happy (h_avg ~ 6.03), a daily cycle peaking at 5-6 am (h_avg ~ 6.12), and sensitivity to major events (Christmas consistently happiest; Osama Bin Laden's death lowest overall). Happiness and information content (Simpson lexical size) were found to be statistically independent (r_s = -0.038, p ~ 0.71).

#sentiment_analysis #hedonometer #twitter_social_media #collective_mood #computational_methods

Fearing the Future of Empirical Psychology: Bem's (2011) Evidence of Psi as a Case Study of Deficiencies in Modal Research Practice

LeBel, Etienne P; Peters, Kurt R 2011 Review of General Psychology

Bem's (2011) nine-experiment report of evidence for psi is used as a diagnostic case study for three systemic deficiencies in modal research practice (MRP) in empirical psychology: (a) overemphasis on conceptual rather than close replication, which allows failed extensions to be filed away while successful ones count as replications; (b) failure to independently verify measurement instrument integrity, with no reliability estimates reported for any dependent variable; and (c) flawed NHST implementation that tests against a nil hypothesis virtually guaranteed to be false, making rejection contingent only on sample size (e.g., a 51.7% hit rate reaching p < .05 at N = 150). These deficiencies produce an 'interpretation bias' that buffers any theory from falsification. Conservatism in theory choice favors revising beliefs about MRP over revising beliefs about causality and time.

#replication_crisis #bem_critique #nhst #file_drawer #interpretation_bias

Extraordinary Claims Require Extraordinary Evidence: The Case of Non-Local Perception, A Classical and Bayesian Review of Evidences

Tressoldi, Patrizio E 2011 Frontiers in Psychology

This mini-review presents quantitative evidence supporting non-local perception (NLP) — the hypothesis that human perceptual abilities may extend beyond space-time constraints of sensory organs. Tressoldi analyzes seven databases covering six experimental protocols: ganzfeld free-response, remote viewing, dream-ESP, presentiment/anticipatory responses, implicit precognition, and forced-choice ESP. Using both frequentist meta-analysis (weighted effect sizes) and Bayesian meta-analysis (Bayes factors), the paper evaluates whether cumulative evidence meets 'extraordinary evidence' standards. Frequentist analysis rejects the null for all protocols (effect sizes d=0.007-0.28). Bayesian analysis shows strong evidence for H1 in three protocols: ganzfeld (BF=18.8 million), remote viewing (BF=25.4 billion), and anticipatory responses (BF=2.89×10^13). Normal consciousness protocols favored the null. Quality-effect size correlations were modestly positive (r=0.05-0.36), contradicting methodological artifact claims.

#meta_analysis #bayesian #frequentist #non_local_perception #methodology

Meditation Experience Is Associated with Differences in Default Mode Network Activity and Connectivity

Brewer, Judson A; Worhunsky, Patrick D; Gray, Jeremy R; Tang, Yi-Yuan; Weber, Jochen; Kober, Hedy 2011 Proceedings of the National Academy of Sciences

Experienced mindfulness meditators (N=12, mean >10,000 h practice) were compared to 12 matched meditation-naive controls using fMRI during Concentration, Loving-Kindness, and Choiceless Awareness meditation. Main default mode network (DMN) nodes — the medial prefrontal cortex (mPFC) and posterior cingulate cortex (PCC) — showed relative deactivation in meditators across all conditions. Functional connectivity analyses revealed stronger PCC coupling with the dorsal anterior cingulate cortex (dACC) and dorsolateral prefrontal cortex (dlPFC), regions implicated in cognitive control, in both meditation and resting-state baseline. Meditators reported significantly less mind-wandering (F(1,22)=7.93, P=0.010). These findings suggest meditation may transform the resting-state default mode toward sustained present-centered awareness.

#fmri #default_mode_network #mindfulness_meditation #functional_connectivity #mind_wandering

Consciousness in the Universe: Neuroscience, Quantum Space-Time Geometry and Orch OR Theory

Penrose, Roger; Hameroff, Stuart 2011 Journal of Cosmology

Orchestrated objective reduction (Orch OR) proposes that consciousness arises from quantum computations in microtubule protein polymers within brain neurons, terminated by an objective reduction of quantum states linked to quantum gravity and spacetime geometry. Building on Penrose's argument that human understanding requires non-computable processes (via Gödel's theorem), and Hameroff's models of microtubules as molecular automata, the theory posits that quantum superpositions of tubulin states evolve until reaching a Diósi–Penrose gravitational threshold (τ ≈ ℏ/EG), producing discrete moments of conscious awareness at roughly gamma synchrony frequencies (~40 Hz). The authors review and respond to criticisms including Tegmark's decoherence calculations, and cite emerging evidence for warm quantum biological processes in photosynthesis, bird navigation, and preliminary ballistic conductance in microtubules.

#quantum_consciousness #microtubules #objective_reduction #quantum_gravity #gamma_synchrony

Extrasensory Perception and Quantum Models of Cognition

Tressoldi, Patrizio E; Storm, Lance; Radin, Dean 2010 NeuroQuantology

Reviewing 108 ganzfeld telepathy experiments conducted from 1974 to 2008 across laboratories worldwide, six independent meta-analyses all show significantly positive hit rates above the 25% chance baseline. The overall effect across 4,196 trials (after outlier removal) was 31.5% hits, corresponding to ES π = 0.58 (95% CI .56–.60, z = 9.9, p = 1.0 × 10⁻¹¹). Ganzfeld and other altered-state experiments yielded comparable effects (π = 0.57–0.60), while non-ASC experiments produced null results. The authors argue that quantum-inspired cognitive models—already shown to outperform classical models for conjunction fallacy, decision-making, and categorization—provide a natural theoretical framework for accommodating nonlocal information transfer.

#ganzfeld_telepathy #quantum_cognition #nonlocal_perception #mental_entanglement #altered_states_of_consciousness

How Many Scientists Fabricate and Falsify Research? A Systematic Review and Meta-Analysis of Survey Data

Fanelli, Daniele 2009 PLoS ONE

Systematic review and meta-analysis of 21 surveys asking scientists about research misconduct, with 18 surveys pooled quantitatively. A random-effects meta-analysis found that 1.97% (95%CI: 0.86–4.45) of scientists admitted fabricating, falsifying, or modifying data at least once, while up to 33.7% admitted questionable research practices. Surveys about colleagues yielded higher rates: 14.12% (95%CI: 9.91–19.72) reported observing falsification and up to 72% observed QRPs. Three methodological factors—self vs. non-self reports, mailed vs. handed surveys, and use of explicit fraud terminology—explained 82% of between-study variance. Medical researchers reported significantly more misconduct. The author concludes these figures are conservative lower bounds.

#research_misconduct #scientific_fraud #questionable_research_practices #survey_data #publication_bias

A Practical Solution to the Pervasive Problems of p Values

Wagenmakers, Eric-Jan 2007 Psychonomic Bulletin & Review

Three fundamental problems with p value null-hypothesis significance testing (NHST) are reviewed with concrete examples. First, p values depend on data never observed, violating the conditionality principle. Second, p values depend on the researcher's subjective sampling intentions — identical data yield p = 0.146 under binomial but p = 0.033 under negative binomial sampling. Third, the 'p postulate' (equal p values = equal evidence) is false: Bayesian analysis shows that for data with p = .05, posterior probability of H₀ rises from ~0.69 at n = 400 to ~0.92 at n = 10,000. The Bayesian information criterion (BIC) is proposed as a practical alternative, approximating Bayesian hypothesis testing without requiring prior specification.

#p_values #bayesian_inference #bayesian_information_criterion #optional_stopping #nhst

Synchronistic Phenomena as Entanglement Correlations in Generalized Quantum Theory

von Lucadou, Walter; Römer, Hartmann; Walach, Harald 2007 Journal of Consciousness Studies

Synchronistic and psi phenomena are reinterpreted as entanglement correlations within Generalized Quantum Theory (GQT), a framework extending standard quantum mechanics to non-physical systems while preserving complementarity and entanglement. From the Non-Transmission (NT) axiom — that entanglement correlations cannot transmit information — three empirically observed features of psi research are derived: the decline effect (effect sizes diminish with replication), displacement (effects migrate to unmonitored variables), and elusiveness. PEAR PK replication data show effect sizes declining from E=0.024 (1981) to E=0.001 (2000), consistent with the predicted 1/√n decay. Correlation-matrix experiments yield significantly elevated psycho-physical correlations without any detectable signal, as the model predicts.

#generalized_quantum_theory #entanglement_correlations #decline_effect #non_signal_transfer #synchronicity

The Mental Universe

Henry, Richard Conn 2005 Nature

Arguing that quantum mechanics resolved the nature of the Universe as fundamentally mental, physicist Richard Conn Henry contends that observations — not material objects — are the only reality. Drawing on Newton's rejection of absolute material existence, Jeans's vision of a 'great thought,' and Bohr's observer-dependent reality, Henry maintains that decoherence and other attempts to preserve material realism produce no new physics. He cites Renninger-type experiments, where wave function collapse occurs without physical detection, as evidence that conscious observation is irreducible. The essay concludes that the Universe is 'immaterial — mental and spiritual.'

#quantum_mechanics #consciousness_primacy #idealism #philosophy_of_physics #copenhagen_interpretation

Why Most Published Research Findings Are False

Ioannidis, John P.A 2005 PLoS Medicine

Mathematical modeling using 2x2 contingency tables proves that most published research findings are false. Positive predictive value (PPV) depends on statistical power (1-beta), pre-study odds (R), Type I error rate (alpha), bias (u), and number of competing teams (n). The core formula PPV = (1-beta)R/(R - betaR + alpha) shows a finding is more likely true than false only when power times pre-study odds exceeds 0.05. Six corollaries identify conditions reducing PPV: small studies, small effects, multiple testing, analytical flexibility, conflicts of interest, and competitive fields. Simulations show adequately powered RCTs achieve PPV = 85%, underpowered exploratory research achieves 12-23%, and discovery-oriented genomics achieves < 0.2%.

#positive_predictive_value #false_positives #statistical_power #replication_crisis #research_bias

Scientists behaving badly

Martinson, Brian C; Anderson, Melissa S; de Vries, Raymond 2005 Nature

An anonymous mail survey of 3,247 NIH-funded US scientists (1,768 mid-career R01 grantees and 1,479 early-career postdoctoral trainees) asked respondents to self-report engagement in 16 specific questionable research practices over the past three years. While falsification and plagiarism were reported at under 2%, broader misbehaviours were strikingly common: 33% admitted to at least one of the ten most serious behaviours, including changing study design under funding pressure (15.5%), dropping data on gut feeling (15.3%), and inadequate record keeping (27.5%). Mid-career scientists reported significantly higher rates than early-career (38% vs 28%, P<0.001). The authors argue these are conservative estimates and call for attention to systemic features of the research environment that may foster misbehaviour.

#research_integrity #scientific_misconduct #self_report_survey #questionable_research_practices #nih

Mindless Statistics

Gigerenzer, Gerd 2004 The Journal of Socio-Economics

The 'null ritual' — NHST as routinely practiced in psychology — is an incoherent hybrid of Fisher's null hypothesis testing and Neyman-Pearson decision theory that neither statistician endorsed. Surveys show ~90% of psychology students, teachers, and professors hold false beliefs about p-values (Haller & Krauss, 2002). Meehl's conjecture that null hypotheses in large non-experimental samples are virtually always false is empirically supported: 46% of random directional predictions confirmed as significant across 81,000+ MMPI-2 responses (Waller, 2004). Statistical power in psychology has averaged ~50% for medium effects since 1962 without improvement. Advocates replacing the null ritual with a toolbox including Bayesian methods, effect sizes, and exploratory analysis.

#null_hypothesis_significance_testing #p_value_misconceptions #statistical_power #effect_size #replication_crisis

A Proposal and Challenge for Proponents and Skeptics of Psi

Kennedy, J.E 2004 Journal of Parapsychology

Pharmaceutical clinical trials, with FDA-mandated power analyses, pre-specified primary outcomes, and independent protocol review, offer a rigorous model for psi research trapped in cycles of inconclusive meta-analysis. Examining published psi meta-analyses reveals that 70-80% of constituent studies are non-significant and z scores fail to increase with sample size — violating a fundamental assumption of statistical testing. A committee of parapsychologists, moderate skeptics, and a statistician is proposed to review pivotal study protocols prospectively. Kennedy challenges skeptics to specify adequate protocols in advance and proponents to demonstrate psi effects are reliable enough for prospective planning.

#power_analysis #pre_registration #pharmaceutical_research_model #meta_analysis_limitations #experimental_design

The Capricious, Actively Evasive, Unsustainable Nature of Psi: A Summary and Hypotheses

Kennedy, J.E 2003 The Journal of Parapsychology

After a century of psychical research, many investigators have concluded that psi may be capricious or actively evasive. Evidence includes negative reliability (significant direction reversals from pilot to confirmation), shifts from intended to unintended secondary effects, pervasive decline effects for subjects, experimenters, and entire research lines, and persistent failure of practical applications. Houtkooper found meta-analytic summary followed by 90% average reduction in effect size across seven series. Three hypotheses are evaluated: competing psi motivation from a polarized population, mechanistic constraints (Lucadou's pragmatic information model), and higher consciousness inducing mystery and wonder. Kennedy favors the last, arguing psi may be inherently unsustainable.

#decline_effects #psi_elusiveness #experimenter_effects #psi_missing #unsustainability

Weak Quantum Theory: Complementarity and Entanglement in Physics and Beyond

Atmanspacher, Harald; Römer, Hartmann; Walach, Harald 2002 Foundations of Physics

Proposes a minimal axiomatic generalization of quantum theory — "Weak Quantum Theory" (WQT) — in which observables are mappings on states forming a monoid rather than a C*-algebra. Six axioms preserve the key quantum features of complementarity (non-commuting observables) and entanglement (holistic correlations in composite systems) while discarding Planck's constant, Hilbert space, tensor products, and probability interpretation. Ordinary quantum theory is recoverable by progressively strengthening the axioms. Two applications are sketched: complementary dynamical descriptions in chaotic systems, and countertransference in psychotherapy as entangled mental states. Provides formal underpinning for applying entanglement concepts to mind-matter and psychophysical phenomena without claiming that quantum physics itself directly governs such processes.

#weak_quantum_theory #generalized_entanglement #complementarity #algebraic_quantum_theory #mind_matter_theoretical_framework

Why Is Psi So Elusive? A Review and Proposed Model

Kennedy, James E 2001 The Journal of Parapsychology

Eleven hypotheses for why psi phenomena are weak, unreliable, and rare are reviewed: methodological artifacts, rarity of ability, precarious psychological conditions, unnoticed psi, goal-oriented experimenter effects, fear of psi, evolutionary inhibition, ecological interconnectedness, spiritual growth, multiple-observer influence, and nonphysical beings. Meta-analyses consistently show z scores unrelated to sample size, contradicting standard statistical assumptions, and replication rates declined across paradigms despite decades of research. An integrative model proposes a bimodal distribution of anomalous experiences, with psi practitioners (~1%) forming a subset who reliably guide psi by intention and psi-conducive experimenters shaping outcomes via goal-oriented influence.

#elusiveness_of_psi #experimenter_effects #replication_decline #individual_differences #goal_oriented_psi

Methods for Investigating Goal-Oriented Psi

Kennedy, J.E 1995 Journal of Parapsychology

Experimental parapsychology faces a fundamental challenge if psi is goal-oriented and operates within a hierarchy of goals ranging from individual trials to entire research programs. Drawing on Schmidt (1974), where single-event PK trials (55.93%, z=5.55) and 100-event majority-vote trials (53.16%, z=2.89) produced approximately equal scoring rates despite communication theory predicting >90% for majority votes, the paper demonstrates that psi bypasses redundant opportunities to achieve goals directly. If psi operates efficiently at higher levels of the goal hierarchy, standard experimental methods may be unable to identify optimum conditions for psi.

#goal_oriented_psi #majority_vote #experimenter_effects #hierarchy_of_goals #signal_enhancement

Decision Augmentation Theory: Toward a Model of Anomalous Mental Phenomena

May, Edwin C; Utts, Jessica M; Spottiswoode, S. James P 1995 The Journal of Parapsychology

Introduces Decision Augmentation Theory (DAT), proposing that statistical anomalies in micro-psychokinesis experiments arise not from a mental force perturbing physical systems but from anomalous cognition (precognition) biasing human decisions toward favorable outcomes within an unperturbed world. Mathematical expressions are derived for normal and binomial distributions, yielding a key testable prediction: under DAT, the expected z² is independent of n (items per trial), whereas force-like models predict z² increases linearly with n. Statistical power curves show that ~1,368 runs at n=10⁴ suffice to separate models at 95% confidence for typical RNG effect sizes. The theory implies all anomalous mental phenomena may reduce to a single mechanism—information transfer from future to past.

#decision_augmentation_theory #anomalous_perturbation #anomalous_cognition #rng_experiments #phenomenological_model

Adverse Effects of Meditation: A Preliminary Investigation of Long-Term Meditators

Shapiro, Deane H. Jr 1992 International Journal of Psychosomatics

Surveyed 27 long-term Vipassana meditators (mean 4.27 years experience; 17 men, 10 women) at an intensive retreat in Barre, Massachusetts, using retrospective and prospective questionnaires at one month and six months post-retreat. Of the 27 subjects, 62.9% reported at least one adverse effect across the three time periods, and 7.4% suffered profound adverse effects leading them to stop meditating. Adverse effects were primarily intrapersonal (76.4%), including increased negativity, emotional pain, anxiety, and disorientation. Despite this, subjects reported significantly more positive than negative effects (p=.002 at time one). Frequency of adverse effects was not reduced by longer practice — the longest-term group (105 months average) reported the highest rate (75%). Replicates and extends Otis (1984) findings from Transcendental Meditation to the Vipassana tradition.

#vipassana #adverse_effects #meditation_safety #long_term_meditators #retreat_effects

Biological Utilisation of Quantum NonLocality

Josephson, Brian D; Pallikari-Viras, Fotini 1991 Foundations of Physics

Nobel laureate Brian Josephson and Fotini Pallikari-Viras argue that Bell's theorem demonstrates the existence of nonlocal connections between spatially separated objects, but that standard quantum mechanics predicts these connections vanish under statistical averaging. Drawing on Bohm's causal interpretation, they propose that this vanishing is an artifact of the specific probability distributions relevant to quantum measurement, and that biosystems — through evolutionary adaptation and developmental learning — may access qualitatively different, goal-focused distributions enabling practical use of nonlocality for telepathy and psychokinesis. The theory is entirely qualitative, offering no quantitative predictions.

#quantum_nonlocality #bell_theorem #bohm_causal_interpretation #theoretical_framework #mind_matter_interaction

Replication and Meta-Analysis in Parapsychology

Utts, Jessica 1991 Statistical Science

Invited review in Statistical Science (Vol. 6, No. 4, pp. 363-403) with seven formal commentaries by Bayarri & Berger, Dawson, Diaconis, Greenhouse, Hyman, Morris, and Mosteller. Demonstrates through worked examples that misunderstandings of statistical power cause scientists to misinterpret nonsignificant replications as failures. Synthesizes meta-analyses across four psi domains: ganzfeld telepathy (autoganzfeld: 122/355 hits = 34.4% vs. 25% chance, p=0.00005, h=0.20), forced-choice precognition (309 studies, z=11.41), RNG/micro-PK (832 studies, z=4.1), and dice-PK (148 studies, z=18.2). Ganzfeld effect size triple that of aspirin on heart attacks. Concludes an anomalous effect exists requiring explanation.

#meta_analysis #replication #ganzfeld_autoganzfeld #statistical_power #effect_size

A Joint Communiqué: The Psi Ganzfeld Controversy

Hyman, Ray; Honorton, Charles 1986 Journal of Parapsychology

A joint methodological communiqué in which skeptic Ray Hyman and parapsychologist Charles Honorton replace their planned debate continuation with a collaborative statement of shared positions. Reviewing the existing ganzfeld database (28 studies), both authors agree that an overall significant effect is present that cannot be explained by selective reporting or multiple analysis, while continuing to disagree on whether this constitutes psi evidence. They issue detailed prescriptive recommendations across six domains: sensory leakage control, target randomization with full documentation, judging and feedback procedures, pre-specified multiple analysis corrections, file-drawer registration, and statistical reporting standards. They advocate for planning experiments with future meta-analysis in mind and call for skeptics to participate as investigators. Originally published Journal of Parapsychology 50, 351–364 (1986).

#ganzfeld #methodology #replication_standards #meta_analysis #psi_debate

On the Quantum Mechanics of Consciousness, with Application to Anomalous Phenomena

Jahn, Robert G; Dunne, Brenda J 1986 Foundations of Physics

Proposes that consciousness be represented by a quantum mechanical wave function in a generalized space/time domain, with the Schrodinger equation defining eigenfunctions within a centered potential well associated with the physical body. Summarizes seven years of PEAR laboratory data — 683,700 REG trials (28 operators), 217,500 pseudo-REG trials, 22 million RMC trials, and 400+ precognitive remote perception trials — showing small but statistically significant anomalous effects (REG z = 2.95 for directional PK). Develops metaphoric analogues of atomic structure, covalent bonds, indistinguishability, exclusion, correspondence, uncertainty, and quantum statistics to model both psychokinesis and remote perception as resonance phenomena between consciousness and device wave functions.

#quantum_mechanics_metaphor #consciousness_wave_function #pear_laboratory #theoretical_framework #mind_matter_interaction

Skepticism and Negative Results in Borderline Areas of Science

Kennedy, J.E 1981

Examines whether biased errors by skeptics play a decisive role in producing their negative results in borderline science, particularly parapsychology. Four case studies are analyzed: Martin Gardner's factual misrepresentations of Coover's ESP experiments in 'Fads and Fallacies,' John Wheeler's fabricated accusation of data fraud against J.B. Rhine at the 1979 AAAS meeting (later retracted in Science), Warner Wilson's strategically underpowered ESP replication that reduced 6,210 trials to 450 after finding p<.005, and Zelen, Kurtz, and Abell's post-hoc subgroup analysis used to undermine the overall significant (p<.03) Mars effect. Concludes that negative results by skeptics require the same methodological scrutiny as positive findings by proponents.

#experimenter_bias #skeptical_methodology #sheep_goat_effect #mars_effect #philosophy_of_science

Intensive Insight Meditation: A Phenomenological Study

Kornfield, Jack 1979 The Journal of Transpersonal Psychology

A phenomenological survey of experiences reported by over 100 two-week and 63 three-month vipassana (insight meditation) retreat students, supplemented by a 21-person non-retreat control group. Questionnaires administered every 2-3 days and teacher interviews yielded 22 categories of unusual experiences. Over 80% of three-month students reported altered perceptions, spontaneous body movements (55%), dramatic mood swings (47%), and rapture/bliss states (95%). Sleep decreased 25% on average and food intake dropped by one-third. A strong positive correlation emerged between concentration levels and frequency of altered states. The control group showed dramatically fewer effects (2/21 sleep decreases, 2 unusual perceptions), indicating intensive practice rather than instruction or social context drives these phenomena. Kornfield concluded these experiences are normative developmental stages, not psychopathology.

#vipassana #phenomenological_study #altered_states #retreat_effects #contemplative_science

The "File Drawer Problem" and Tolerance for Null Results

Rosenthal, Robert 1979 Psychological Bulletin

Introduces the "file drawer problem" — the concern that journals publish the 5% of studies showing Type I errors while 95% of null results remain unpublished — and derives a quantitative solution: the fail-safe N (tolerance for future null results). Using the method of adding standard normal deviates across k independent studies, the formula X = (k/2.706)[k(Z-bar)^2 - 2.706] computes how many additional null-result studies would be needed to reduce a combined significance level to p = .05. Illustrated with interpersonal expectancy research: 94 studies require 3,263 null studies to overturn; 311 studies require 49,457. Proposes X >= 5k + 10 as a threshold for file-drawer resistance.

#publication_bias #fail_safe_n #meta_analytic_methodology #file_drawer_problem #combined_significance

Theoretical Risks and Tabular Asterisks: Sir Karl, Sir Ronald, and the Slow Progress of Soft Psychology

Meehl, Paul E 1978 Journal of Consulting and Clinical Psychology

Arguing that theories in soft psychology (clinical, social, personality) neither cumulate nor get clearly refuted but merely fade away, Meehl identifies 20 intrinsic difficulties of the subject matter and then turns to an extrinsic one: the near-universal reliance on null hypothesis significance testing. Drawing on Popper's falsificationism and Bayesian reasoning, he shows that since the null hypothesis is quasi-always false, refuting it depends on statistical power, not theoretical verisimilitude, making significance tests a feeble method of theory corroboration. As an alternative, he demonstrates 'consistency tests' within his MAXCOV-HITMAX taxometric framework, achieving 94% accuracy with zero false negatives across 600 Monte Carlo samples. He prescribes multiple non-redundant estimates of theoretical quantities over significance tests.

#significance_testing_critique #philosophy_of_science #falsificationism #taxometrics #replication_crisis