1
|
Lala S, Jha NK. SECRETS: Subject-efficient clinical randomized controlled trials using synthetic intervention. Contemp Clin Trials Commun 2024; 38:101265. [PMID: 38352896 PMCID: PMC10862504 DOI: 10.1016/j.conctc.2024.101265] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2023] [Revised: 12/15/2023] [Accepted: 01/28/2024] [Indexed: 02/16/2024] Open
Abstract
Background The parallel-group randomized controlled trial (RCT) is commonly used in Phase-3 clinical trials to establish treatment effectiveness but requires hundreds-to-thousands of subjects, making it difficult to implement, which leads to high Phase-3 trial failure rates. One approach to increasing power of a trial is to augment data collected from an RCT with external data from prospective studies or prior RCTs. However, this requires that external data be comparable to data from the study of interest, a condition that does not hold for new interventions or populations being studied. Another approach is to lower sample size requirements by using the cross-over design, which measures individual treatment effects (ITEs) to remove inter-subject variability; however, this design is only suitable for chronic conditions and interventions with effects that wash out rapidly. Method We propose a novel and practical framework called SECRETS (Subject-Efficient Clinical Randomized Controlled Trials using Synthetic Intervention) to increase power of any parallel-group RCT by simulating the cross-over design using only data collected from the study. SECRETS first estimates ITEs across all subjects recruited to the RCT by using a state-of-the-art counterfactual estimation algorithm called synthetic intervention (SI). Since SI induces dependencies among the ITEs, we introduce a novel hypothesis testing strategy to test for treatment effectiveness. Results We show that SECRETS can increase the power of an RCT while maintaining comparable significance levels; in particular, on three real-world clinical RCTs (Phase-3 trials), SECRETS increases power over the baseline method by 6 - 54 % (average: 21.5%, standard deviation: 15.8%), thereby reducing the number of subjects needed to obtain a typically desired statistical operating point of 80% power and 5% significance level by 25 - 76 % (10-3,957 fewer subjects per arm). Our analyses show that SECRETS increases power by consistently reducing the variance of the average treatment effect, thereby mimicking the effects of a cross-over design. Conclusion SECRETS increases subject efficiency of an RCT by simulating the cross-over design using only data collected from the RCT; therefore, it is a feasible solution for increasing the trial's power, especially under settings where satisfying sample size requirements is difficult.
Collapse
Affiliation(s)
- Sayeri Lala
- Department of Electrical and Computer Engineering, Princeton University, Princeton, 08544, NJ, USA
| | - Niraj K. Jha
- Department of Electrical and Computer Engineering, Princeton University, Princeton, 08544, NJ, USA
| |
Collapse
|
2
|
Sarafoglou A, Bartoš F, Stefan A, Haaf JM, Wagenmakers EJ. "This behavior strikes us as ideal": assessment and anticipations of Huisman (2022). Psychon Bull Rev 2024; 31:242-248. [PMID: 37542014 PMCID: PMC10866761 DOI: 10.3758/s13423-023-02299-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/23/2023] [Indexed: 08/06/2023]
Abstract
Huisman (Psychonomic Bulletin & Review, 1-10. 2022) argued that a valid measure of evidence should indicate more support in favor of a true alternative hypothesis when sample size is large than when it is small. Bayes factors may violate this pattern and hence Huisman concluded that Bayes factors are invalid as a measure of evidence. In this brief comment we call attention to the following: (1) Huisman's purported anomaly is in fact dictated by probability theory; (2) Huisman's anomaly has been discussed and explained in the statistical literature since 1939; the anomaly was also highlighted in the Psychonomic Bulletin & Review article by Rouder et al. (2009), who interpreted the anomaly as "ideal": an interpretation diametrically opposed to that of Huisman. We conclude that when intuition clashes with probability theory, chances are that it is intuition that needs schooling.
Collapse
Affiliation(s)
- Alexandra Sarafoglou
- Department of Psychology, University of Amsterdam, Nieuwe Achtergracht 129B, 1001, NK, Amsterdam, Netherlands.
| | - František Bartoš
- Department of Psychology, University of Amsterdam, Nieuwe Achtergracht 129B, 1001, NK, Amsterdam, Netherlands
| | - Angelika Stefan
- Department of Psychology, University of Amsterdam, Nieuwe Achtergracht 129B, 1001, NK, Amsterdam, Netherlands
- Department of Psychology, Universität der Bundeswehr München, Munich, Germany
| | - Julia M Haaf
- Department of Psychology, University of Amsterdam, Nieuwe Achtergracht 129B, 1001, NK, Amsterdam, Netherlands
| | - Eric-Jan Wagenmakers
- Department of Psychology, University of Amsterdam, Nieuwe Achtergracht 129B, 1001, NK, Amsterdam, Netherlands
| |
Collapse
|
3
|
Luo R. Hypothesis testing of Poisson rates in COVID-19 offspring distributions. Infect Dis Model 2023; 8:980-1001. [PMID: 37663920 PMCID: PMC10469988 DOI: 10.1016/j.idm.2023.07.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2022] [Revised: 07/14/2023] [Accepted: 07/18/2023] [Indexed: 09/05/2023] Open
Abstract
In the present study, we undertake the task of hypothesis testing in the context of Poisson-distributed data. The primary objective of our investigation is to ascertain whether two distinct sets of discrete data share the same Poisson rate. We delve into a comprehensive review and comparative analysis of various frequentist and Bayesian methodologies specifically designed to address this problem. Among these are the conditional test, the likelihood ratio test, and the Bayes factor. Additionally, we employ the posterior predictive p-value in our analysis, coupled with its corresponding calibration procedures. As the culmination of our investigation, we apply these diverse methodologies to test both simulated datasets and real-world data. The latter consists of the offspring distributions linked to COVID-19 cases in two disparate geographies - Hong Kong and Rwanda. This allows us to provide a practical demonstration of the methodologies' applications and their potential implications in the field of epidemiology.
Collapse
Affiliation(s)
- Rui Luo
- Department of Systems Engineering, City University of Hong Kong, Kowloon Town, Hong Kong Special Administrative Region
| |
Collapse
|
4
|
Ramli M, Budiantara IN, Ratnasari V. A method for parameter hypothesis testing in nonparametric regression with Fourier series approach. MethodsX 2023; 11:102468. [PMID: 37964783 PMCID: PMC10641682 DOI: 10.1016/j.mex.2023.102468] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2023] [Accepted: 10/29/2023] [Indexed: 11/16/2023] Open
Abstract
Nonparametric regression model with the Fourier series approach was first introduced by Bilodeau in 1994. In the later years, several researchers developed a nonparametric regression model with the Fourier series approach. However, these researches are limited to parameter estimation and there is no research related to parameter hypothesis testing. Parameter hypothesis testing is a statistical method used to test the significance of the parameters. In nonparametric regression model with the Fourier series approach, parameter hypothesis testing is used to determine whether the estimated parameters have significance influence on the model or not. Therefore, the purpose of this research is for parameter hypothesis testing in the nonparametric regression model with the Fourier series approach. The method that we use for hypothesis testing is the LRT method. The LRT method is a method that compares the likelihood functions under the parameter space of the null hypothesis and the hypothesis. By using the LRT method, we obtain the form of the statistical test and its distribution as well as the rejection region of the null hypothesis. To apply the method, we use ROA data from 47 go public banks that are listed on the Indonesia stock exchange in 2020. The highlights of this research are:•The Fourier series function is assumed as a non-smooth function.•The form of the statistical test is obtained using the LRT method and is distributed as F distribution.•The estimated parameters on modelling ROA data have a significant influence on the model.
Collapse
Affiliation(s)
- Mustain Ramli
- Department of Statistics, Faculty of Science and Data Analytics, Institut Teknologi Sepuluh Nopember, Kampus ITS-Sukolilo, Surabaya 60111, Indonesia
| | - I Nyoman Budiantara
- Department of Statistics, Faculty of Science and Data Analytics, Institut Teknologi Sepuluh Nopember, Kampus ITS-Sukolilo, Surabaya 60111, Indonesia
| | - Vita Ratnasari
- Department of Statistics, Faculty of Science and Data Analytics, Institut Teknologi Sepuluh Nopember, Kampus ITS-Sukolilo, Surabaya 60111, Indonesia
| |
Collapse
|
5
|
Chen YT, Gao LL. Testing for a difference in means of a single feature after clustering. ArXiv 2023:arXiv:2311.16375v1. [PMID: 38076519 PMCID: PMC10705581] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Figures] [Subscribe] [Scholar Register] [Indexed: 12/23/2023]
Abstract
For many applications, it is critical to interpret and validate groups of observations obtained via clustering. A common validation approach involves testing differences in feature means between observations in two estimated clusters. In this setting, classical hypothesis tests lead to an inflated Type I error rate. To overcome this problem, we propose a new test for the difference in means in a single feature between a pair of clusters obtained using hierarchical or k-means clustering. The test based on the proposed p-value controls the selective Type I error rate in finite samples and can be efficiently computed. We further illustrate the validity and power of our proposal in simulation and demonstrate its use on single-cell RNA-sequencing data.
Collapse
Affiliation(s)
- Yiqun T Chen
- Department of Biomedical Data Science, Stanford University
| | - Lucy L Gao
- Department of Statistics, University of British Columbia, November 29, 2023
| |
Collapse
|
6
|
Brönmark C, Hellström G, Baktoft H, Hansson LA, McCallum ES, Nilsson PA, Skov C, Brodin T, Hulthén K. Ponds as experimental arenas for studying animal movement: current research and future prospects. Mov Ecol 2023; 11:68. [PMID: 37880741 PMCID: PMC10601242 DOI: 10.1186/s40462-023-00419-9] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/14/2023] [Accepted: 09/02/2023] [Indexed: 10/27/2023]
Abstract
Animal movement is a multifaceted process that occurs for multiple reasons with powerful consequences for food web and ecosystem dynamics. New paradigms and technical innovations have recently pervaded the field, providing increasingly powerful means to deliver fine-scale movement data, attracting renewed interest. Specifically in the aquatic environment, tracking with acoustic telemetry now provides integral spatiotemporal information to follow individual movements in the wild. Yet, this technology also holds great promise for experimental studies, enhancing our ability to truly establish cause-and-effect relationships. Here, we argue that ponds with well-defined borders (i.e. "islands in a sea of land") are particularly well suited for this purpose. To support our argument, we also discuss recent experiences from studies conducted in an innovative experimental infrastructure, composed of replicated ponds equipped with modern aquatic telemetry systems that allow for unparalleled insights into the movement patterns of individual animals.
Collapse
Affiliation(s)
- Christer Brönmark
- Department of Biology-Aquatic Ecology, Lund University, Ecology building, Sölvegatan 37 223 62, Lund, Sweden.
| | - Gustav Hellström
- Department of Wildlife, Fish and Environmental Studies, Swedish University of Agricultural Sciences (SLU), Umeå, 90183, Sweden
| | - Henrik Baktoft
- National Institute of Aquatic Resources, Technical University of Denmark (DTU), Silkeborg, Denmark
| | - Lars-Anders Hansson
- Department of Biology-Aquatic Ecology, Lund University, Ecology building, Sölvegatan 37 223 62, Lund, Sweden
| | - Erin S McCallum
- Department of Wildlife, Fish and Environmental Studies, Swedish University of Agricultural Sciences (SLU), Umeå, 90183, Sweden
| | - P Anders Nilsson
- Department of Biology-Aquatic Ecology, Lund University, Ecology building, Sölvegatan 37 223 62, Lund, Sweden
| | - Christian Skov
- National Institute of Aquatic Resources, Technical University of Denmark (DTU), Silkeborg, Denmark
| | - Tomas Brodin
- Department of Wildlife, Fish and Environmental Studies, Swedish University of Agricultural Sciences (SLU), Umeå, 90183, Sweden
| | - Kaj Hulthén
- Department of Biology-Aquatic Ecology, Lund University, Ecology building, Sölvegatan 37 223 62, Lund, Sweden.
| |
Collapse
|
7
|
Trinh P, Clausen DS, Willis AD. happi: a hierarchical approach to pangenomics inference. Genome Biol 2023; 24:214. [PMID: 37773075 PMCID: PMC10540326 DOI: 10.1186/s13059-023-03040-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2022] [Accepted: 08/16/2023] [Indexed: 09/30/2023] Open
Abstract
Recovering metagenome-assembled genomes (MAGs) from shotgun sequencing data is an increasingly common task in microbiome studies, as MAGs provide deeper insight into the functional potential of both culturable and non-culturable microorganisms. However, metagenome-assembled genomes vary in quality and may contain omissions and contamination. These errors present challenges for detecting genes and comparing gene enrichment across sample types. To address this, we propose happi, an approach to testing hypotheses about gene enrichment that accounts for genome quality. We illustrate the advantages of happi over existing approaches using published Saccharibacteria MAGs, Streptococcus thermophilus MAGs, and via simulation.
Collapse
Affiliation(s)
- Pauline Trinh
- Department of Environmental & Occupational Health Sciences, University of Washington, Seattle, WA, USA
| | - David S Clausen
- Department of Biostatistics, University of Washington, Seattle, WA, USA
| | - Amy D Willis
- Department of Biostatistics, University of Washington, Seattle, WA, USA.
| |
Collapse
|
8
|
Robertson EP, Walsh DP, Martin J, Work TM, Kellogg CA, Evans JS, Barker V, Hawthorn A, Aeby G, Paul VJ, Walker BK, Kiryu Y, Woodley CM, Meyer JL, Rosales SM, Studivan M, Moore JF, Brandt ME, Bruckner A. Rapid prototyping for quantifying belief weights of competing hypotheses about emergent diseases. J Environ Manage 2023; 337:117668. [PMID: 36958278 DOI: 10.1016/j.jenvman.2023.117668] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/01/2022] [Revised: 02/10/2023] [Accepted: 03/02/2023] [Indexed: 06/18/2023]
Abstract
Emerging diseases can have devastating consequences for wildlife and require a rapid response. A critical first step towards developing appropriate management is identifying the etiology of the disease, which can be difficult to determine, particularly early in emergence. Gathering and synthesizing existing information about potential disease causes, by leveraging expert knowledge or relevant existing studies, provides a principled approach to quickly inform decision-making and management efforts. Additionally, updating the current state of knowledge as more information becomes available over time can reduce scientific uncertainty and lead to substantial improvement in the decision-making process and the application of management actions that incorporate and adapt to newly acquired scientific understanding. Here we present a rapid prototyping method for quantifying belief weights for competing hypotheses about the etiology of disease using a combination of formal expert elicitation and Bayesian hierarchical modeling. We illustrate the application of this approach for investigating the etiology of stony coral tissue loss disease (SCTLD) and discuss the opportunities and challenges of this approach for addressing emergent diseases. Lastly, we detail how our work may apply to other pressing management or conservation problems that require quick responses. We found the rapid prototyping methods to be an efficient and rapid means to narrow down the number of potential hypotheses, synthesize current understanding, and help prioritize future studies and experiments. This approach is rapid by providing a snapshot assessment of the current state of knowledge. It can also be updated periodically (e.g., annually) to assess changes in belief weights over time as scientific understanding increases. Synthesis and applications: The rapid prototyping approaches demonstrated here can be used to combine knowledge from multiple experts and/or studies to help with fast decision-making needed for urgent conservation issues including emerging diseases and other management problems that require rapid responses. These approaches can also be used to adjust belief weights over time as studies and expert knowledge accumulate and can be a helpful tool for adapting management decisions.
Collapse
Affiliation(s)
- Ellen P Robertson
- Contract Quantitative Ecologist, US Geological Survey, Wetland and Aquatic Research Center, Gainesville, FL, USA.
| | - Daniel P Walsh
- U.S. Geological Survey, Montana Cooperative Wildlife Research Unit, University of Montana, Missoula, MT, USA.
| | - Julien Martin
- U.S. Geological Survey, Eastern Ecological Science Center, Laurel, MD, USA.
| | - Thierry M Work
- U.S. Geological Survey, National Wildlife Health Center, Honolulu Field Station, Honolulu, HI, USA
| | - Christina A Kellogg
- U.S. Geological Survey, St. Petersburg Coastal and Marine Science Center, St. Petersburg, FL, USA
| | - James S Evans
- U.S. Geological Survey, St. Petersburg Coastal and Marine Science Center, St. Petersburg, FL, USA
| | | | - Aine Hawthorn
- U.S. Geological Survey National Wildlife Health Center, Western Fisheries Research Center, Seattle, WA, USA
| | - Greta Aeby
- Smithsonian Marine Station, Fort Pierce, FL, USA
| | | | - Brian K Walker
- Nova Southeastern University, Halmos College of Arts and Sciences, Dania Beach, FL, USA
| | - Yasunari Kiryu
- Fish and Wildlife Research Institute, Florida Fish and Wildlife Conservation Commission, St. Petersburg, FL, USA
| | - Cheryl M Woodley
- Hollings Marine Laboratory, Center for Coastal Environmental Health and Biomolecular Research, National Oceanic and Atmospheric Administration's National Ocean Service, Charleston, SC, USA
| | - Julie L Meyer
- Department of Soil, Water, and Ecosystem Sciences, University of Florida, Gainesville, FL, USA
| | - Stephanie M Rosales
- Cooperative Institute for Marine and Atmospheric Studies, University of Miami, Miami, FL, USA; Atlantic Oceanographic and Meteorological Laboratory, National Oceanic and Atmospheric Administration, Miami, FL, USA
| | - Michael Studivan
- Cooperative Institute for Marine and Atmospheric Studies, University of Miami, Miami, FL, USA; Atlantic Oceanographic and Meteorological Laboratory, National Oceanic and Atmospheric Administration, Miami, FL, USA
| | - Jennifer F Moore
- Moore Ecological Analysis and Management, LLC, Gainesville, FL, USA
| | - Marilyn E Brandt
- Center for Marine and Environmental Studies, University of the Virgin Islands, St. Thomas, USVI, USA
| | - Andrew Bruckner
- Florida Keys National Marine Sanctuary, NOAA, Key Largo, FL, USA
| |
Collapse
|
9
|
Chen YT, Witten DM. Selective inference for k-means clustering. J Mach Learn Res 2023; 24:152. [PMID: 38264325 PMCID: PMC10805457] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Figures] [Subscribe] [Scholar Register] [Indexed: 01/25/2024]
Abstract
We consider the problem of testing for a difference in means between clusters of observations identified via k -means clustering. In this setting, classical hypothesis tests lead to an inflated Type I error rate. In recent work, Gao et al. (2022) considered a related problem in the context of hierarchical clustering. Unfortunately, their solution is highly-tailored to the context of hierarchical clustering, and thus cannot be applied in the setting of k -means clustering. In this paper, we propose a p-value that conditions on all of the intermediate clustering assignments in the k -means algorithm. We show that the p-value controls the selective Type I error for a test of the difference in means between a pair of clusters obtained using k -means clustering in finite samples, and can be efficiently computed. We apply our proposal on hand-written digits data and on single-cell RNA-sequencing data.
Collapse
Affiliation(s)
- Yiqun T Chen
- Data Science Institute and Department of Biomedical Data Science, Stanford University, Stanford, CA 94305, USA
| | - Daniela M Witten
- Departments of Statistics and Biostatistics, University of Washington, Seattle, WA 98195-4322, USA
| |
Collapse
|
10
|
Chen YT, Jewell SW, Witten DM. Quantifying uncertainty in spikes estimated from calcium imaging data. Biostatistics 2023; 24:481-501. [PMID: 34654923 PMCID: PMC10449000 DOI: 10.1093/biostatistics/kxab034] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2021] [Revised: 07/28/2021] [Accepted: 09/04/2021] [Indexed: 11/12/2022] Open
Abstract
In recent years, a number of methods have been proposed to estimate the times at which a neuron spikes on the basis of calcium imaging data. However, quantifying the uncertainty associated with these estimated spikes remains an open problem. We consider a simple and well-studied model for calcium imaging data, which states that calcium decays exponentially in the absence of a spike, and instantaneously increases when a spike occurs. We wish to test the null hypothesis that the neuron did not spike-i.e., that there was no increase in calcium-at a particular timepoint at which a spike was estimated. In this setting, classical hypothesis tests lead to inflated Type I error, because the spike was estimated on the same data used for testing. To overcome this problem, we propose a selective inference approach. We describe an efficient algorithm to compute finite-sample $p$-values that control selective Type I error, and confidence intervals with correct selective coverage, for spikes estimated using a recent proposal from the literature. We apply our proposal in simulation and on calcium imaging data from the $\texttt{spikefinder}$ challenge.
Collapse
Affiliation(s)
- Yiqun T Chen
- Department of Biostatistics, University of Washington,
Seattle, WA 98195, USA
| | - Sean W Jewell
- Department of Statistics, University of Washington, Seattle,
WA 98195, USA
| | - Daniela M Witten
- Departments of Statistics & Biostatistics, University of
Washington, Seattle, WA 98195, USA
| |
Collapse
|
11
|
Fierro R. Cumulative damage for multi-type epidemics and an application to infectious diseases. J Math Biol 2023; 86:47. [PMID: 36797526 PMCID: PMC9934514 DOI: 10.1007/s00285-023-01880-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2022] [Revised: 12/04/2022] [Accepted: 01/23/2023] [Indexed: 02/18/2023]
Abstract
A continuous time multivariate stochastic model is proposed for assessing the damage of a multi-type epidemic cause to a population as it unfolds. The instants when cases occur and the magnitude of their injure are random. Thus, we define a cumulative damage based on counting processes and a multivariate mark process. For a large population we approximate the behavior of this damage process by its asymptotic distribution. Also, we analyze the distribution of the stopping times when the numbers of cases caused by the epidemic attain levels beyond certain thresholds. We focus on introducing some tools for statistical inference on the parameters related with the epidemic. In this regard, we present a general hypothesis test for homogeneity in epidemics and apply it to data of Covid-19 in Chile.
Collapse
Affiliation(s)
- Raúl Fierro
- Instituto de Matemáticas, Pontificia Universidad Católica de Valparaíso, Casilla 4059, Valparaíso, Chile.
| |
Collapse
|
12
|
Abstract
Recent insights into problems with common statistical practice in psychology have motivated scientists to consider alternatives to the traditional frequentist approach that compares p-values to a significance criterion. While these alternatives have worthwhile attributes, Francis (Behavior Research Methods, 40, 1524-1538, 2017) showed that many proposed test statistics for the situation of a two-sample t-test are based on precisely the same information in a given data set; and for a given sample size, one can convert from any statistic to the others. Here, we show that the same relationship holds for the equivalent of a one-sample t-test. We derive the relationships and provide an on-line app that performs the computations. A key conclusion of this analysis is that many types of tests are based on the same information, so the choice of which approach to use should reflect the intent of the scientist and the appropriateness of the corresponding inferential framework for that intent.
Collapse
Affiliation(s)
- Gregory Francis
- Department of Psychological Sciences, Purdue University, 703 Third Street, West Lafayette, IN, 47907-2004, USA.
| | - Victoria Jakicic
- Department of Psychological Sciences, Purdue University, 703 Third Street, West Lafayette, IN, 47907-2004, USA
| |
Collapse
|
13
|
Stefan AM, Schönbrodt FD, Evans NJ, Wagenmakers EJ. Efficiency in sequential testing: Comparing the sequential probability ratio test and the sequential Bayes factor test. Behav Res Methods 2022; 54:3100-3117. [PMID: 35233752 PMCID: PMC9729330 DOI: 10.3758/s13428-021-01754-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/16/2021] [Indexed: 12/16/2022]
Abstract
In a sequential hypothesis test, the analyst checks at multiple steps during data collection whether sufficient evidence has accrued to make a decision about the tested hypotheses. As soon as sufficient information has been obtained, data collection is terminated. Here, we compare two sequential hypothesis testing procedures that have recently been proposed for use in psychological research: Sequential Probability Ratio Test (SPRT; Psychological Methods, 25(2), 206-226, 2020) and the Sequential Bayes Factor Test (SBFT; Psychological Methods, 22(2), 322-339, 2017). We show that although the two methods have different philosophical roots, they share many similarities and can even be mathematically regarded as two instances of an overarching hypothesis testing framework. We demonstrate that the two methods use the same mechanisms for evidence monitoring and error control, and that differences in efficiency between the methods depend on the exact specification of the statistical models involved, as well as on the population truth. Our simulations indicate that when deciding on a sequential design within a unified sequential testing framework, researchers need to balance the needs of test efficiency, robustness against model misspecification, and appropriate uncertainty quantification. We provide guidance for navigating these design decisions based on individual preferences and simulation-based design analyses.
Collapse
Affiliation(s)
- Angelika M Stefan
- Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands.
| | - Felix D Schönbrodt
- Department of Psychology, Ludwig-Maximilians-Universität München, München, Germany
| | - Nathan J Evans
- School of Psychology, University of Queensland, St Lucia, Australia
| | | |
Collapse
|
14
|
Cui Y, Peng L. Assessing dynamic covariate effects with survival data. Lifetime Data Anal 2022; 28:675-699. [PMID: 35962886 PMCID: PMC9901566 DOI: 10.1007/s10985-022-09571-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/15/2021] [Accepted: 06/06/2022] [Indexed: 06/15/2023]
Abstract
Dynamic (or varying) covariate effects often manifest meaningful physiological mechanisms underlying chronic diseases. However, a static view of covariate effects is typically adopted by standard approaches to evaluating disease prognostic factors, which can result in depreciation of some important disease markers. To address this issue, in this work, we take the perspective of globally concerned quantile regression, and propose a flexible testing framework suited to assess either constant or dynamic covariate effects. We study the powerful Kolmogorov-Smirnov (K-S) and Cramér-Von Mises (C-V) type test statistics and develop a simple resampling procedure to tackle their complicated limit distributions. We provide rigorous theoretical results, including the limit null distributions and consistency under a general class of alternative hypotheses of the proposed tests, as well as the justifications for the presented resampling procedure. Extensive simulation studies and a real data example demonstrate the utility of the new testing procedures and their advantages over existing approaches in assessing dynamic covariate effects.
Collapse
Affiliation(s)
- Ying Cui
- Department of Biostatistics and Bioinformatics, Rollins School of Public Health, Emory University, Atlanta, GA, USA
| | - Limin Peng
- Department of Biostatistics and Bioinformatics, Rollins School of Public Health, Emory University, Atlanta, GA, USA.
| |
Collapse
|
15
|
Melikechi O, Young AL, Tang T, Bowman T, Dunson D, Johndrow J. Limits of epidemic prediction using SIR models. J Math Biol 2022; 85:36. [PMID: 36125562 PMCID: PMC9487859 DOI: 10.1007/s00285-022-01804-5] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2021] [Revised: 08/12/2022] [Accepted: 08/30/2022] [Indexed: 11/27/2022]
Abstract
The Susceptible-Infectious-Recovered (SIR) equations and their extensions comprise a commonly utilized set of models for understanding and predicting the course of an epidemic. In practice, it is of substantial interest to estimate the model parameters based on noisy observations early in the outbreak, well before the epidemic reaches its peak. This allows prediction of the subsequent course of the epidemic and design of appropriate interventions. However, accurately inferring SIR model parameters in such scenarios is problematic. This article provides novel, theoretical insight on this issue of practical identifiability of the SIR model. Our theory provides new understanding of the inferential limits of routinely used epidemic models and provides a valuable addition to current simulate-and-check methods. We illustrate some practical implications through application to a real-world epidemic data set.
Collapse
Affiliation(s)
- Omar Melikechi
- Department of Mathematics, Duke University, Durham, NC, USA.
| | | | - Tao Tang
- Department of Mathematics, Duke University, Durham, NC, USA
| | - Trevor Bowman
- Department of Mathematics, Duke University, Durham, NC, USA
| | - David Dunson
- Department of Mathematics, Duke University, Durham, NC, USA.,Department of Statistics, Duke University, Durham, NC, USA
| | - James Johndrow
- Department of Statistics, University of Pennsylvania, Philadelphia, PA, USA
| |
Collapse
|
16
|
McNulty R. A logical analysis of null hypothesis significance testing using popular terminology. BMC Med Res Methodol 2022; 22:244. [PMID: 36123631 DOI: 10.1186/s12874-022-01696-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2022] [Accepted: 07/20/2022] [Indexed: 11/16/2022] Open
Abstract
Background Null Hypothesis Significance Testing (NHST) has been well criticised over the years yet remains a pillar of statistical inference. Although NHST is well described in terms of statistical models, most textbooks for non-statisticians present the null and alternative hypotheses (H0 and HA, respectively) in terms of differences between groups such as (μ1 = μ2) and (μ1 ≠ μ2) and HA is often stated to be the research hypothesis. Here we use propositional calculus to analyse the internal logic of NHST when couched in this popular terminology. The testable H0 is determined by analysing the scope and limits of the P-value and the test statistic’s probability distribution curve. Results We propose a minimum axiom set NHST in which it is taken as axiomatic that H0 is rejected if P-value< α. Using the common scenario of the comparison of the means of two sample groups as an example, the testable H0 is {(μ1 = μ2) and [(\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\overline{x}$$\end{document}x¯1 ≠ \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\overline{x}$$\end{document}x¯2) due to chance alone]}. The H0 and HA pair should be exhaustive to avoid false dichotomies. This entails that HA is ¬{(μ1 = μ2) and [(\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\overline{x}$$\end{document}x¯1 ≠ \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\overline{x}$$\end{document}x¯2) due to chance alone]}, rather than the research hypothesis (HT). To see the relationship between HA and HT, HA can be rewritten as the disjunction HA: ({(μ1 = μ2) ∧ [(\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\overline{x}$$\end{document}x¯1 ≠ \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\overline{x}$$\end{document}x¯2) not due to chance alone]} ∨ {(μ1 ≠ μ2) ∧ [\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$(\overline{x}$$\end{document}(x¯1 ≠ \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\overline{x}$$\end{document}x¯2) not due to (μ1 ≠ μ2) alone]} ∨ {(μ1 ≠ μ2) ∧ [(\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\overline{\boldsymbol{x}}$$\end{document}x¯1≠\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\overline{\boldsymbol{x}}$$\end{document}x¯2) due to (μ1 ≠ μ2) alone]}). This reveals that HT (the last disjunct in bold) is just one possibility within HA. It is only by adding premises to NHST that HT or other conclusions can be reached. Conclusions Using this popular terminology for NHST, analysis shows that the definitions of H0 and HA differ from those found in textbooks. In this framework, achieving a statistically significant result only justifies the broad conclusion that the results are not due to chance alone, not that the research hypothesis is true. More transparency is needed concerning the premises added to NHST to rig particular conclusions such as HT. There are also ramifications for the interpretation of Type I and II errors, as well as power, which do not specifically refer to HT as claimed by texts.
Collapse
|
17
|
Abstract
The graph fused lasso-which includes as a special case the one-dimensional fused lasso-is widely used to reconstruct signals that are piecewise constant on a graph, meaning that nodes connected by an edge tend to have identical values. We consider testing for a difference in the means of two connected components estimated using the graph fused lasso. A naive procedure such as a z-test for a difference in means will not control the selective Type I error, since the hypothesis that we are testing is itself a function of the data. In this work, we propose a new test for this task that controls the selective Type I error, and conditions on less information than existing approaches, leading to substantially higher power. We illustrate our approach in simulation and on datasets of drug overdose death rates and teenage birth rates in the contiguous United States. Our approach yields more discoveries on both datasets. Supplementary materials for this article are available online.
Collapse
Affiliation(s)
- Yiqun Chen
- Department of Biostatistics, University of Washington, Seattle, WA
| | - Sean Jewell
- Department of Statistics, University of Washington, Seattle, WA
| | - Daniela Witten
- Department of Biostatistics, University of Washington, Seattle, WA
- Department of Statistics, University of Washington, Seattle, WA
| |
Collapse
|
18
|
Verploegh ISC, Lazar NA, Bartels RHMA, Volovici V. Evaluation of the Use of P Values in Neurosurgical Literature: from Statistical Significance to Clinical Irrelevance. World Neurosurg 2022; 161:280-283.e3. [PMID: 35505545 DOI: 10.1016/j.wneu.2022.02.018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2021] [Revised: 02/03/2022] [Accepted: 02/03/2022] [Indexed: 11/25/2022]
Abstract
The application and interpretation of P values have caused debate for several decades, and this debate has become particularly relevant in the past few years. The P value represents the probability of seeing results as extreme or more extreme than those observed in a data analysis, were the null hypothesis and other underlying assumptions to be true. While P values are useful in pointing out where an effect may be present, they have often been misused in an attempt to oversell "statistically significant" findings. As P values rely on the spread and number of measurements, a smaller P value does not necessarily imply a larger effect size, which is better assessed via an effect estimate and confidence interval interpreted in the context of the study. The clinical relevance of a computed P value is context dependent. We investigated the current use of P values in a small sample of recent neurosurgical literature. Only a minority of manuscripts that reported statistical significance described confounder adjustment, or effect sizes. A common, incorrect assumption often observed was that statistical significance equals clinical relevance. To enable correct interpretation of clinical significance, it is crucial that authors describe the clinical implications of their findings.
Collapse
Affiliation(s)
- Iris S C Verploegh
- Department of Neurosurgery, Erasmus Medical Center, Rotterdam, the Netherlands.
| | - Nicole A Lazar
- Department of Statistics, Pennsylvania State University, University Park, Pennsylvania, USA
| | | | - Victor Volovici
- Department of Neurosurgery, Erasmus Medical Center, Rotterdam, the Netherlands
| |
Collapse
|
19
|
Stefan AM, Katsimpokis D, Gronau QF, Wagenmakers EJ. Expert agreement in prior elicitation and its effects on Bayesian inference. Psychon Bull Rev 2022. [PMID: 35378671 DOI: 10.3758/s13423-022-02074-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 02/14/2022] [Indexed: 11/08/2022]
Abstract
Bayesian inference requires the specification of prior distributions that quantify the pre-data uncertainty about parameter values. One way to specify prior distributions is through prior elicitation, an interview method guiding field experts through the process of expressing their knowledge in the form of a probability distribution. However, prior distributions elicited from experts can be subject to idiosyncrasies of experts and elicitation procedures, raising the spectre of subjectivity and prejudice. Here, we investigate the effect of interpersonal variation in elicited prior distributions on the Bayes factor hypothesis test. We elicited prior distributions from six academic experts with a background in different fields of psychology and applied the elicited prior distributions as well as commonly used default priors in a re-analysis of 1710 studies in psychology. The degree to which the Bayes factors vary as a function of the different prior distributions is quantified by three measures of concordance of evidence: We assess whether the prior distributions change the Bayes factor direction, whether they cause a switch in the category of evidence strength, and how much influence they have on the value of the Bayes factor. Our results show that although the Bayes factor is sensitive to changes in the prior distribution, these changes do not necessarily affect the qualitative conclusions of a hypothesis test. We hope that these results help researchers gauge the influence of interpersonal variation in elicited prior distributions in future psychological studies. Additionally, our sensitivity analyses can be used as a template for Bayesian robustness analyses that involve prior elicitation from multiple experts.
Collapse
|
20
|
Wang C, Sabo RT, Mukhopadhyay ND, Perera RA. Early termination in single-parameter model phase II clinical trial designs using decreasingly informative priors. Int J Clin Trials 2022; 9:107-117. [PMID: 36846554 PMCID: PMC9957559 DOI: 10.18203/2349-3259.ijct20221110] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/18/2023]
Abstract
Background To exchange the type of subjective Bayesian prior selection for assumptions more directly related to statistical decision making in clinician studies and trials, the decreasingly informative prior (DIP) is considered. We expand standard Bayesian early termination methods in one-parameter statistical models for Phase II clinical trials to include decreasingly informative priors (DIP). These priors are designed to reduce the chance of erroneously adapting trials too early by parameterize skepticism in an amount always equal to the unobserved sample size. Method We show how to parameterize these priors based on effective prior sample size and provide examples for common single-parameter models, include Bernoulli, Poisson, and Gaussian distributions. We use a simulation study to search through possible values of total sample sizes and termination thresholds to find the smallest total sample size (N) under admissible designs, which we define as having at least 80% power and no greater than 5% type I error rate. Results For Bernoulli, Poisson, and Gaussian distributions, the DIP approach requires fewer patients when admissible designs are achieved. In situations where type I error or power are not admissible, the DIP approach yields similar power and better-controlled type I error with comparable or fewer patients than other Bayesian priors by Thall and Simon. Conclusions The DIP helps control type I error rates with comparable or fewer patients, especially for those instances when increased type I error rates arise from erroneous termination early in a trial.
Collapse
Affiliation(s)
- Chen Wang
- Department of Biostatistics, Virginia Commonwealth University, Richmond VA, U. S. A
| | - Roy T. Sabo
- Department of Biostatistics, Virginia Commonwealth University, Richmond VA, U. S. A
| | | | - Robert A. Perera
- Department of Biostatistics, Virginia Commonwealth University, Richmond VA, U. S. A
| |
Collapse
|
21
|
Abstract
Change point analysis aims to detect structural changes in a data sequence. It has always been an active research area since it was introduced in the 1950s. In modern statistical applications, however, high-throughput data with increasing dimensions are ubiquitous in fields ranging from economics, finance to genetics and engineering. For those problems, the earlier works are typically no longer applicable. As a result, the problem of testing a change point for high dimensional data sequences has been an important yet challenging task. In this paper, we first focus on models for at most one change point, and review recent state-of-art techniques for change point testing of high dimensional mean vectors and compare their theoretical properties. Based on that, we provide a survey of some extensions to general high dimensional parameters beyond mean vectors as well as strategies for testing multiple change points in high dimensions. Finally, we discuss some open problems for possible future research directions.
Collapse
Affiliation(s)
- Bin Liu
- School of Management, Fudan University, Shanghai, 200433, China
| | - Xinsheng Zhang
- School of Management, Fudan University, Shanghai, 200433, China
| | - Yufeng Liu
- Department of Statistics and Operations Research, Department of Genetics, and Department of Biostatistics, Carolina Center for Genome Sciences, Lineberger Comprehensive Cancer Center, University of North Carolina at Chapel Hill, U.S.A.,Corresponding author. . (Yufeng Liu)
| |
Collapse
|
22
|
Abstract
BACKGROUND Network meta-analysis (NMA) is a statistical method used to combine results from several clinical trials and simultaneously compare multiple treatments using direct and indirect evidence. Statistical heterogeneity is a characteristic describing the variability in the intervention effects being evaluated in the different studies in network meta-analysis. One approach to dealing with statistical heterogeneity is to perform a random effects network meta-analysis that incorporates a between-study variance into the statistical model. A common assumption in the random effects model for network meta-analysis is the homogeneity of between-study variance across all interventions. However, there are applications of NMA where the single between-study assumption is potentially incorrect and instead the model should incorporate more than one between-study variances. METHODS In this paper, we develop an approach to testing the homogeneity of between-study variance assumption based on a likelihood ratio test. A simulation study was conducted to assess the type I error and power of the proposed test. This method is then applied to a network meta-analysis of antibiotic treatments for Bovine respiratory disease (BRD). RESULTS The type I error rate was well controlled in the Monte Carlo simulation. We found statistical evidence (p value = 0.052) against the homogeneous between-study variance assumption in the network meta-analysis BRD. The point estimate and confidence interval of relative effect sizes are strongly influenced by this assumption. CONCLUSIONS Since homogeneous between-study variance assumption is a strong assumption, it is crucial to test the validity of this assumption before conducting a network meta-analysis. Here we propose and validate a method for testing this single between-study variance assumption which is widely used for many NMA.
Collapse
Affiliation(s)
- Dapeng Hu
- Department of Statistics, College of Liberal Arts and Sciences, Iowa State University, Ames, IA, USA
| | - Chong Wang
- Department of Statistics, College of Liberal Arts and Sciences, Iowa State University, Ames, IA, USA. .,Department of Veterinary Diagnostic and Production Animal Medicine, College of Veterinary Medicine, Iowa State University, Ames, IA, USA.
| | - Annette M O'Connor
- Department of Veterinary Diagnostic and Production Animal Medicine, College of Veterinary Medicine, Iowa State University, Ames, IA, USA.,Professor of Epidemiology, Chairperson of the Department of Large Animal Clinical Sciences, College of Veterinary Medicine, Michigan State University, East Lansing, MI, USA
| |
Collapse
|
23
|
Dick AS, Lopez DA, Watts AL, Heeringa S, Reuter C, Bartsch H, Fan CC, Kennedy DN, Palmer C, Marshall A, Haist F, Hawes S, Nichols TE, Barch DM, Jernigan TL, Garavan H, Grant S, Pariyadath V, Hoffman E, Neale M, Stuart EA, Paulus MP, Sher KJ, Thompson WK. Meaningful associations in the adolescent brain cognitive development study. Neuroimage 2021; 239:118262. [PMID: 34147629 PMCID: PMC8803401 DOI: 10.1016/j.neuroimage.2021.118262] [Citation(s) in RCA: 74] [Impact Index Per Article: 24.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2021] [Revised: 05/07/2021] [Accepted: 06/10/2021] [Indexed: 02/08/2023] Open
Abstract
The Adolescent Brain Cognitive Development (ABCD) Study is the largest single-cohort prospective longitudinal study of neurodevelopment and children's health in the United States. A cohort of n = 11,880 children aged 9-10 years (and their parents/guardians) were recruited across 22 sites and are being followed with in-person visits on an annual basis for at least 10 years. The study approximates the US population on several key sociodemographic variables, including sex, race, ethnicity, household income, and parental education. Data collected include assessments of health, mental health, substance use, culture and environment and neurocognition, as well as geocoded exposures, structural and functional magnetic resonance imaging (MRI), and whole-genome genotyping. Here, we describe the ABCD Study aims and design, as well as issues surrounding estimation of meaningful associations using its data, including population inferences, hypothesis testing, power and precision, control of covariates, interpretation of associations, and recommended best practices for reproducible research, analytical procedures and reporting of results.
Collapse
Affiliation(s)
- Anthony Steven Dick
- Department of Psychology and Center for Children and Families, Florida International University, Miami, FL, United States
| | - Daniel A Lopez
- Division of Epidemiology, Department of Public Health Sciences, University of Rochester Medical Center, Rochester, NY 14642, United States
| | - Ashley L Watts
- Department of Psychology, University of Missouri, MO, United States
| | - Steven Heeringa
- Institute for Social Research, University of Michigan, Ann Arbor, MI 48109, United States
| | - Chase Reuter
- Herbert Wertheim School of Public Health and Human Longevity Science, University of California, San Diego, La Jolla, CA 92093, United States
| | - Hauke Bartsch
- Mohn Medical Imaging and Visualization Center, Department of Radiology, Haukeland University Hospital, Bergen, Norway
| | - Chun Chieh Fan
- Population Neuroscience and Genetics Lab, University of California, San Diego, La Jolla, CA 92093, United States
| | - David N Kennedy
- Department of Psychiatry, University of Massachusetts Medical School, MA United States, 01604
| | - Clare Palmer
- Center for Human Development, University of California, San Diego, La Jolla, CA 92093, United States
| | - Andrew Marshall
- Children's Hospital Los Angeles, and the Department of Pediatrics, University of Southern California, Los Angeles, CA, United States
| | - Frank Haist
- Department of Radiology, University of California, San Diego, La Jolla, CA 92093, United States
| | - Samuel Hawes
- Department of Psychology and Center for Children and Families, Florida International University, Miami, FL, United States
| | - Thomas E Nichols
- Oxford Big Data Institute, Li Ka Shing Centre for Health Information and Discovery, Nuffield Department of Population Health, University of Oxford, Oxford, United Kingdom
| | - Deanna M Barch
- Departments of Psychological & Brain Sciences, Psychiatry and Radiology, Washington University, St. Louis, MO 63130, United States
| | - Terry L Jernigan
- Department of Psychiatry, University of Massachusetts Medical School, MA United States, 01604
| | - Hugh Garavan
- Department of Psychiatry, University of Vermont, Burlington, VT, 05405, United States
| | - Steven Grant
- Behavioral and Cognitive Neuroscience Branch, Division of Neuroscience and Behavior, National Institute on Drug Abuse, National Institutes of Health, Department of Health and Human Services, Bethesda, MD, United States
| | - Vani Pariyadath
- Behavioral and Cognitive Neuroscience Branch, Division of Neuroscience and Behavior, National Institute on Drug Abuse, National Institutes of Health, Department of Health and Human Services, Bethesda, MD, United States
| | - Elizabeth Hoffman
- National Institute on Drug Abuse, National Institutes of Health, Department of Heatlh and Human Services, Bethesda, MD, United States
| | - Michael Neale
- Department of Psychiatry, Virginia Commonwealth University, Richmond, VA 23298, United States
| | - Elizabeth A Stuart
- Department of Biostatistics, Bloomberg School of Public Health, Johns Hopkins University, Baltimore, MD, United States
| | - Martin P Paulus
- Laureate Institute for Brain Research, Tulsa, OK, United States
| | - Kenneth J Sher
- Department of Psychology, University of Missouri, MO, United States
| | - Wesley K Thompson
- Herbert Wertheim School of Public Health and Human Longevity Science, University of California, San Diego, La Jolla, CA 92093, United States; Population Neuroscience and Genetics Lab, University of California, San Diego, La Jolla, CA 92093, United States.
| |
Collapse
|
24
|
Morris AMA. English engineer John Smeaton's experimental method(s): Optimisation, hypothesis testing and exploratory experimentation. Stud Hist Philos Sci 2021; 89:283-294. [PMID: 34547653 DOI: 10.1016/j.shpsa.2021.07.004] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/08/2020] [Revised: 07/05/2021] [Accepted: 07/15/2021] [Indexed: 06/13/2023]
Abstract
In this paper I provide a detailed account of eighteenth-century engineer John Smeaton's experimental methods, with the aim of bringing our understanding of his work into line with recent research in the history and philosophy of science. Starting from his use of the technique of parameter variation, I identify three distinct methodological aims in the research he carried out on waterwheels, windmills and hydraulic mortars. These aims are: optimisation, hypothesis testing and maxim generation. The main claim of this paper is that Smeaton did more than merely improve engineering methods by systematising earlier artisanal approaches, which is the classic view of Smeaton's method developed by historians of technology in the 1990s. I argue instead that his approach bridged the divide between science and technology, by integrating both hypothesis testing and exploratory experimentation. This is borne out, in particular, by the way that Smeaton emphasised the exploratory side of the work he published in the Philosophical Transactions, in contrast to his account of the construction of the Eddystone lighthouse, which was aimed at a broader, non-specialist public. I contribute to recent research on exploratory experimentation by showing - in line with other work on this topic - that exploratory experimentation is not incompatible with hypothesis testing. This new perspective on Smeaton's method will hopefully lead to further research and new insights into the relation between science and technology at the start of the Industrial Revolution.
Collapse
Affiliation(s)
- Andrew M A Morris
- Centre for Logic and Philosophy of Science, Vrije Universiteit Brussel, Pleinlaan 2, Room 5B425, B-1050, Brussels, Belgium.
| |
Collapse
|
25
|
Wang C, Hu J, Blaser MJ, Li H. Microbial trend analysis for common dynamic trend, group comparison, and classification in longitudinal microbiome study. BMC Genomics 2021; 22:667. [PMID: 34525957 PMCID: PMC8442444 DOI: 10.1186/s12864-021-07948-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2020] [Accepted: 08/25/2021] [Indexed: 12/26/2022] Open
Abstract
BACKGROUND The human microbiome is inherently dynamic and its dynamic nature plays a critical role in maintaining health and driving disease. With an increasing number of longitudinal microbiome studies, scientists are eager to learn the comprehensive characterization of microbial dynamics and their implications to the health and disease-related phenotypes. However, due to the challenging structure of longitudinal microbiome data, few analytic methods are available to characterize the microbial dynamics over time. RESULTS We propose a microbial trend analysis (MTA) framework for the high-dimensional and phylogenetically-based longitudinal microbiome data. In particular, MTA can perform three tasks: 1) capture the common microbial dynamic trends for a group of subjects at the community level and identify the dominant taxa; 2) examine whether or not the microbial overall dynamic trends are significantly different between groups; 3) classify an individual subject based on its longitudinal microbial profiling. Our extensive simulations demonstrate that the proposed MTA framework is robust and powerful in hypothesis testing, taxon identification, and subject classification. Our real data analyses further illustrate the utility of MTA through a longitudinal study in mice. CONCLUSIONS The proposed MTA framework is an attractive and effective tool in investigating dynamic microbial pattern from longitudinal microbiome studies.
Collapse
Affiliation(s)
- Chan Wang
- Division of Biostatistics, Department of Population Health, New York University School of Medicine, New York, 10016 NY USA
| | - Jiyuan Hu
- Division of Biostatistics, Department of Population Health, New York University School of Medicine, New York, 10016 NY USA
| | - Martin J. Blaser
- Center for Advanced Biotechnology and Medicine, Rutgers University, Piscataway, 08854-8021 NJ USA
| | - Huilin Li
- Division of Biostatistics, Department of Population Health, New York University School of Medicine, New York, 10016 NY USA
| |
Collapse
|
26
|
Anderson RB, Crawford JC, Bailey MH. Biasing the input: A yoked-scientist demonstration of the distorting effects of optional stopping on Bayesian inference. Behav Res Methods 2021. [PMID: 34494220 DOI: 10.3758/s13428-021-01618-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 05/05/2021] [Indexed: 11/08/2022]
Abstract
Prior work by Michael R. Dougherty and colleagues (Yu et al., 2014) shows that when a scientist monitors the p value during data collection and uses a critical p as the signal to stop collecting data, the resulting p is distorted due to Type I error-rate inflation. They argued similarly that the use of a critical Bayes factor (BF(crit)) for stopping distorts the obtained Bayes factor (BF), a position that has met with controversy. The present paper clarified that when BF(crit) is used as a stopping criterion, the sample becomes biased in that data consistent with large effects have a greater chance to be included than do other data, thus biasing the input to Bayesian inference. We report simulations of yoked pairs of scientists in which Scientist A uses BF(crit) to optionally stop, while Scientist B, sampling from the same population, stops when A stops. Thus, optional stopping is compared not to a hypothetical in which no stopping occurs, but to a situation in which B stops for reasons unrelated to the characteristics of B's sample. The results indicated that optional stopping biased the input for Bayesian inference. We also simulated the use of effect-size stabilization as a stopping criterion and found no bias in that case.
Collapse
|
27
|
Khan U, Khan AM, Alkatheery N, Khan U. Pandemic and its effect on professional environment on the Kingdom of Saudi Arabia. Environ Sci Pollut Res Int 2021; 28:41162-41168. [PMID: 33779902 PMCID: PMC8006104 DOI: 10.1007/s11356-021-13501-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/12/2020] [Accepted: 03/15/2021] [Indexed: 06/12/2023]
Abstract
The pandemic has affected the world from many different perspectives, including environmental change. This research study aims to investigate the pandemic and its associated effect on the professional environment by measuring some of the parameters that are likely to disclose the impact of the pandemic. A structural questionnaire elicits design to capture the effect of COVID-19, where 284 respondents participated and present their views on a different statement based on the Likert scale. The factor analysis reveals five factors, which were further tested by hypothesis testing and binary logistic regression-and found factors 2, 3, and 5 to be significant in both tests.
Collapse
Affiliation(s)
- Uzma Khan
- Department of Finance, College of Business Administration, Prince Sattam Bin Abdulaziz University, Al-Kharj, Saudi Arabia
| | | | - Nouf Alkatheery
- College of Business Administration, Prince Sattam bin Abdulaziz University, Al-Kharj, Saudi Arabia
| | - Urooja Khan
- TGT, Aligarh Muslim University, Aligarh, India
| |
Collapse
|
28
|
Chen S, Lin X. Analysis in case-control sequencing association studies with different sequencing depths. Biostatistics 2021; 21:577-593. [PMID: 30590456 PMCID: PMC7308042 DOI: 10.1093/biostatistics/kxy073] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/19/2018] [Revised: 10/17/2018] [Accepted: 10/21/2018] [Indexed: 01/09/2023] Open
Abstract
With the advent of next-generation sequencing, investigators have access to higher quality sequencing data. However, to sequence all samples in a study using next generation sequencing can still be prohibitively expensive. One potential remedy could be to combine next generation sequencing data from cases with publicly available sequencing data for controls, but there could be a systematic difference in quality of sequenced data, such as sequencing depths, between sequenced study cases and publicly available controls. We propose a regression calibration (RC)-based method and a maximum-likelihood method for conducting an association study with such a combined sample by accounting for differential sequencing errors between cases and controls. The methods allow for adjusting for covariates, such as population stratification as confounders. Both methods control type I error and have comparable power to analysis conducted using the true genotype with sufficiently high but different sequencing depths. We show that the RC method allows for analysis using naive variance estimate (closely approximates true variance in practice) and standard software under certain circumstances. We evaluate the performance of the proposed methods using simulation studies and apply our methods to a combined data set of exome sequenced acute lung injury cases and healthy controls from the 1000 Genomes project.
Collapse
Affiliation(s)
- Sixing Chen
- Department of Biostatistics, Harvard TH Chan School of Public Health, 655 Huntington Avenue, Building 2, 4th Floor, Boston, MA 02115, USA
| | - Xihong Lin
- Department of Biostatistics, Harvard TH Chan School of Public Health, 655 Huntington Avenue, Building 2, 4th Floor, Boston, MA 02115, USA.,Department of Statistics, Harvard University, One Oxford Street, Suite 400, Cambridge, MA 02138-2901, USA
| |
Collapse
|
29
|
Abstract
Recently, optional stopping has been a subject of debate in the Bayesian psychology community. Rouder (Psychonomic Bulletin & Review 21(2), 301-308, 2014) argues that optional stopping is no problem for Bayesians, and even recommends the use of optional stopping in practice, as do (Wagenmakers, Wetzels, Borsboom, van der Maas & Kievit, Perspectives on Psychological Science 7, 627-633, 2012). This article addresses the question of whether optional stopping is problematic for Bayesian methods, and specifies under which circumstances and in which sense it is and is not. By slightly varying and extending Rouder's (Psychonomic Bulletin & Review 21(2), 301-308, 2014) experiments, we illustrate that, as soon as the parameters of interest are equipped with default or pragmatic priors-which means, in most practical applications of Bayes factor hypothesis testing-resilience to optional stopping can break down. We distinguish between three types of default priors, each having their own specific issues with optional stopping, ranging from no-problem-at-all (type 0 priors) to quite severe (type II priors).
Collapse
Affiliation(s)
- Rianne de Heide
- Leiden University, Leiden, Amsterdam, The Netherlands
- The Netherlands Centre for Mathematics & Computer Science (CWI), Amsterdam, The Netherlands
| | - Peter D Grünwald
- Leiden University, Leiden, Amsterdam, The Netherlands.
- The Netherlands Centre for Mathematics & Computer Science (CWI), Amsterdam, The Netherlands.
| |
Collapse
|
30
|
Liu Z, Shen J, Barfield R, Schwartz J, Baccarelli AA, Lin X. Large-Scale Hypothesis Testing for Causal Mediation Effects with Applications in Genome-wide Epigenetic Studies. J Am Stat Assoc 2021; 117:67-81. [PMID: 35989709 PMCID: PMC9385159 DOI: 10.1080/01621459.2021.1914634] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2020] [Revised: 04/04/2021] [Accepted: 04/04/2021] [Indexed: 01/03/2023]
Abstract
In genome-wide epigenetic studies, it is of great scientific interest to assess whether the effect of an exposure on a clinical outcome is mediated through DNA methylations. However, statistical inference for causal mediation effects is challenged by the fact that one needs to test a large number of composite null hypotheses across the whole epigenome. Two popular tests, the Wald-type Sobel's test and the joint significant test using the traditional null distribution are underpowered and thus can miss important scientific discoveries. In this paper, we show that the null distribution of Sobel's test is not the standard normal distribution and the null distribution of the joint significant test is not uniform under the composite null of no mediation effect, especially in finite samples and under the singular point null case that the exposure has no effect on the mediator and the mediator has no effect on the outcome. Our results explain why these two tests are underpowered, and more importantly motivate us to develop a more powerful Divide-Aggregate Composite-null Test (DACT) for the composite null hypothesis of no mediation effect by leveraging epigenome-wide data. We adopted Efron's empirical null framework for assessing statistical significance of the DACT test. We showed analytically that the proposed DACT method had improved power, and could well control type I error rate. Our extensive simulation studies showed that, in finite samples, the DACT method properly controlled the type I error rate and outperformed Sobel's test and the joint significance test for detecting mediation effects. We applied the DACT method to the US Department of Veterans Affairs Normative Aging Study, an ongoing prospective cohort study which included men who were aged 21 to 80 years at entry. We identified multiple DNA methylation CpG sites that might mediate the effect of smoking on lung function with effect sizes ranging from -0.18 to -0.79 and false discovery rate controlled at level 0.05, including the CpG sites in the genes AHRR and F2RL3. Our sensitivity analysis found small residual correlations (less than 0.01) of the error terms between the outcome and mediator regressions, suggesting that our results are robust to unmeasured confounding factors.
Collapse
Affiliation(s)
- Zhonghua Liu
- Department of Statistics and Actuarial Science, University of Hong Kong
| | - Jincheng Shen
- Department of Population Health Sciences, University of Utah School of Medicine
| | - Richard Barfield
- Department of Biostatistics and Bioinformatics, Duke University School of Medicine
| | - Joel Schwartz
- Environmental Epidemiology, Harvard T.H. Chan School of Public Health
| | - Andrea A. Baccarelli
- Environmental Health Sciences, Mailman School of Public Health, Columbia University
| | - Xihong Lin
- Biostatistics at Harvard T.H. Chan School of Public Health and Statistics at Faculty of Arts and Sciences, Harvard University
| |
Collapse
|
31
|
Abstract
When data are not normally distributed, researchers are often uncertain whether it is legitimate to use tests that assume Gaussian errors, or whether one has to either model a more specific error structure or use randomization techniques. Here we use Monte Carlo simulations to explore the pros and cons of fitting Gaussian models to non-normal data in terms of risk of type I error, power and utility for parameter estimation. We find that Gaussian models are robust to non-normality over a wide range of conditions, meaning that p values remain fairly reliable except for data with influential outliers judged at strict alpha levels. Gaussian models also performed well in terms of power across all simulated scenarios. Parameter estimates were mostly unbiased and precise except if sample sizes were small or the distribution of the predictor was highly skewed. Transformation of data before analysis is often advisable and visual inspection for outliers and heteroscedasticity is important for assessment. In strong contrast, some non-Gaussian models and randomization techniques bear a range of risks that are often insufficiently known. High rates of false-positive conclusions can arise for instance when overdispersion in count data is not controlled appropriately or when randomization procedures ignore existing non-independencies in the data. Hence, newly developed statistical methods not only bring new opportunities, but they can also pose new threats to reliability. We argue that violating the normality assumption bears risks that are limited and manageable, while several more sophisticated approaches are relatively error prone and particularly difficult to check during peer review. Scientists and reviewers who are not fully aware of the risks might benefit from preferentially trusting Gaussian mixed models in which random effects account for non-independencies in the data.
Collapse
|
32
|
Robinson MA, Vanrenterghem J, Pataky TC. Sample size estimation for biomechanical waveforms: Current practice, recommendations and a comparison to discrete power analysis. J Biomech 2021; 122:110451. [PMID: 33933866 DOI: 10.1016/j.jbiomech.2021.110451] [Citation(s) in RCA: 20] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2020] [Revised: 04/06/2021] [Accepted: 04/09/2021] [Indexed: 12/16/2022]
Abstract
Testing a prediction is fundamental to scientific experiments. Where biomechanical experiments involve analysis of 1-Dimensional (waveform) data, sample size estimation should consider both 1D variance and hypothesised 1D effects. This study exemplifies 1D sample size estimation using typical biomechanical signals and contrasts this with 0D (discrete) power analysis. For context, biomechanics papers from 2018 and 2019 were reviewed to characterise current practice. Sample size estimation occurred in approximately 4% of 653 papers and reporting practice was mixed. To estimate sample sizes, common biomechanical signals were sourced from the literature and 1D effects were generated artificially using the open-source power1d software. Smooth Gaussian noise was added to the modelled 1D effect to numerically estimate the sample size required. Sample sizes estimated using 1D power procedures varied according to the characteristics of the dataset, requiring only small-to-moderate sample sizes of approximately 5-40 to achieve target powers of 0.8 for reported 1D effects, but were always larger than 0D sample sizes (from N + 1 to >N + 20). The importance of a priori sample size estimation is highlighted and recommendations are provided to improve the consistency of reporting. This study should enable researchers to construct 1D biomechanical effects to address adequately powered, hypothesis-driven, predictive research questions.
Collapse
Affiliation(s)
- Mark A Robinson
- School of Sport and Exercise Sciences, Liverpool John Moores University, UK.
| | - Jos Vanrenterghem
- Musculoskeletal Rehabilitation Research Group, Faculty of Movement and Rehabilitation Sciences, Leuven KU, Belgium
| | - Todd C Pataky
- Department of Human Health Sciences, Kyoto University, Japan
| |
Collapse
|
33
|
Li G, Walter SD, Thabane L. Shifting the focus away from binary thinking of statistical significance and towards education for key stakeholders: revisiting the debate on whether it's time to de-emphasize or get rid of statistical significance. J Clin Epidemiol 2021; 137:104-112. [PMID: 33839240 DOI: 10.1016/j.jclinepi.2021.03.033] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2020] [Revised: 03/03/2021] [Accepted: 03/10/2021] [Indexed: 01/01/2023]
Abstract
There has been a long-standing controversy among scientists regarding the appropriate use of P-values and statistical significance in clinical research. This debate has resurfaced through recent calls to modify the threshold of P-value required to declare significance, or to retire statistical significance entirely. In this article, we revisit the issue by discussing: i) the connection between statistical thinking and evidence-based practice; ii) some history of statistical significance and P-values; iii) some practical challenges with statistical significance or P-value thresholds in clinical research; iv) the on-going debate on what to do with statistical significance; v) suggestions to shift the focus away from binary thinking of statistical significance and towards education for key stakeholders on research essentials including statistical thinking, critical thinking, good reporting, basic clinical research concepts and methods, and more. We then conclude with remarks and illustrations of the potential deleterious public health consequences of poor methods including selective choice of analysis approach and misguided reliance on binary use of P-values to report and interpret scientific findings.
Collapse
Affiliation(s)
- Guowei Li
- Center for Clinical Epidemiology and Methodology (CCEM), Guangdong Second Provincial General Hospital, Guangzhou City, Guangdong Province, China 510317; Department of Health research methods, Evidence, and Impact (HEI), McMaster University, Hamilton, Ontario, Canada
| | - Stephen D Walter
- Department of Health research methods, Evidence, and Impact (HEI), McMaster University, Hamilton, Ontario, Canada
| | - Lehana Thabane
- Department of Health research methods, Evidence, and Impact (HEI), McMaster University, Hamilton, Ontario, Canada; Population Health Research Institute, Hamilton Health Sciences and McMaster University, Hamilton, Ontario, Canada; Father Sean O'Sullivan Research Centre, St. Joseph's Healthcare Hamilton, Hamilton, Ontario, Canada.
| |
Collapse
|
34
|
Abstract
The effect of treatment on binary disease outcome can differ across subgroups characterized by other covariates. Testing for the existence of subgroups that are associated with heterogeneous treatment effects can provide valuable insight regarding the optimal treatment recommendation in practice. Our research in this paper is motivated by the question of whether host genetics could modify a vaccine's effect on HIV acquisition risk. To answer this question, we used data from an HIV vaccine trial with a two-phase sampling design and developed a general threshold-based model framework to test for the existence of subgroups associated with the heterogeneity in disease risks, allowing for subgroups based on multivariate covariates. We developed a testing procedure based on maximum of likelihood-ratio statistics over change planes and demonstrated its advantage over alternative methods. We further developed the testing procedure to account for bias sampling of expensive (i.e. resource-intensive to measure) covariates through the incorporation of inverse probability weighting techniques. We used the proposed method to analyze the motivating HIV vaccine trial data. Our proposed testing procedure also has broad applications in epidemiological studies for assessing heterogeneity in disease risk with respect to univariate or multivariate predictors.
Collapse
Affiliation(s)
- Ying Huang
- Biostatistics, Bioinformatics, & Epidemiology Program, Fred Hutchinson Cancer Research Center, Seattle, WA, 98109
| | - Juhee Cho
- Biostatistics, Bioinformatics, & Epidemiology Program, Fred Hutchinson Cancer Research Center, Seattle, WA, 98109
| | - Youyi Fong
- Biostatistics, Bioinformatics, & Epidemiology Program, Fred Hutchinson Cancer Research Center, Seattle, WA, 98109
| |
Collapse
|
35
|
Taylor JET, Taylor GW. Artificial cognition: How experimental psychology can help generate explainable artificial intelligence. Psychon Bull Rev 2021; 28:454-75. [PMID: 33159244 DOI: 10.3758/s13423-020-01825-5] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 10/02/2020] [Indexed: 11/08/2022]
Abstract
Artificial intelligence powered by deep neural networks has reached a level of complexity where it can be difficult or impossible to express how a model makes its decisions. This black-box problem is especially concerning when the model makes decisions with consequences for human well-being. In response, an emerging field called explainable artificial intelligence (XAI) aims to increase the interpretability, fairness, and transparency of machine learning. In this paper, we describe how cognitive psychologists can make contributions to XAI. The human mind is also a black box, and cognitive psychologists have over 150 years of experience modeling it through experimentation. We ought to translate the methods and rigor of cognitive psychology to the study of artificial black boxes in the service of explainability. We provide a review of XAI for psychologists, arguing that current methods possess a blind spot that can be complemented by the experimental cognitive tradition. We also provide a framework for research in XAI, highlight exemplary cases of experimentation within XAI inspired by psychological science, and provide a tutorial on experimenting with machines. We end by noting the advantages of an experimental approach and invite other psychologists to conduct research in this exciting new field.
Collapse
|
36
|
Abstract
A powerful way to evaluate scientific explanations (hypotheses) is to test the predictions that they make. In this way, predictions serve as an important bridge between abstract hypotheses and concrete experiments. Experimental biologists, however, generally receive little guidance on how to generate quality predictions. Here, we identify two important components of good predictions - criticality and persuasiveness - which relate to the ability of a prediction (and the experiment it implies) to disprove a hypothesis or to convince a skeptic that the hypothesis has merit. Using a detailed example, we demonstrate how striving for predictions that are both critical and persuasive can speed scientific progress by leading us to more powerful experiments. Finally, we provide a quality control checklist to assist students and researchers as they navigate the hypothetico-deductive method from puzzling observations to experimental tests.
Collapse
Affiliation(s)
- Douglas S Fudge
- Schmid College of Science and Technology, Chapman University, 1 University Dr., Orange, CA 92866, USA
| | - Andy J Turko
- Department of Biology, McMaster University, 1280 Main St W, Hamilton, ON L8S 4L8, Canada
| |
Collapse
|
37
|
Abstract
While the applied psychology community relies on statistics to assist drawing conclusions from quantitative data, the methods being used mostly today do not reflect several of the advances in statistics that have been realized over the past decades. We show in this paper how a number of issues with how statistical analyses are presently executed and reported in the literature can be addressed by applying more modern methods. Unfortunately, such new methods are not always supported by widely available statistical packages, such as SPSS, which is why we also introduce a new software platform, called ILLMO (for Interactive Log-Likelihood MOdeling), which offers an intuitive interface to such modern statistical methods. In order to limit the complexity of the material being covered in this paper, we focus the discussion on a fairly simple, but nevertheless very frequent and important statistical task, i.e., comparing two experimental conditions.
Collapse
|
38
|
Garrido Wainer JM, Espinosa JF, Hirmas N, Trujillo N. Free-viewing as experimental system to test the Temporal Correlation Hypothesis: A case of theory-generative experimental practice. Stud Hist Philos Biol Biomed Sci 2020; 83:101307. [PMID: 32467019 DOI: 10.1016/j.shpsc.2020.101307] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/02/2020] [Revised: 05/13/2020] [Accepted: 05/14/2020] [Indexed: 06/11/2023]
Abstract
Theory-free characterizations of experimental systems miss normative and conceptual components that sometimes are crucial to understanding their historical development. In the following paper, we show that these components may be part of the intrinsic capacities of experimental systems themselves. We study a case of non-exploratory and theory-oriented research in experimental neuroscience that concerns the construction of free-viewing as an experimental system to test one particular pre-existing hypothesis, the Temporal Correlation Hypothesis (TCH), at a laboratory in Santiago de Chile, during 2002-2008. We show that the system does not take well-formulated pre-existing predictions or hypotheses to test them directly, but re-creates them and re-signifies them in terms that are not implied by the theoretical background from which they originally derived. Therefore, we conclude that there is a sui generis way in which experimental systems produce proper theoretical knowledge.
Collapse
Affiliation(s)
| | - Juan Felipe Espinosa
- Universidad Andres Bello, Escuela de Ingeniería Comercial, Facultad de Economía y Negocios, Quillota #980, Viña del Mar, Chile
| | - Natalia Hirmas
- Faculty of Education, Pontificia Universidad Católica de Chile, Vicuña Mackenna 4860, 7810000, Macul, Santiago, Chile
| | - Nicolás Trujillo
- Philosophy Institute, Universidad Diego Portales / Leiden University, Ejército Libertador 260, 8370056, Santiago, Chile
| |
Collapse
|
39
|
Abstract
Background In medical research and practice, the p-value is arguably the most often used statistic and yet it is widely misconstrued as the probability of the type I error, which comes with serious consequences. This misunderstanding can greatly affect the reproducibility in research, treatment selection in medical practice, and model specification in empirical analyses. By using plain language and concrete examples, this paper is intended to elucidate the p-value confusion from its root, to explicate the difference between significance and hypothesis testing, to illuminate the consequences of the confusion, and to present a viable alternative to the conventional p-value. Main text The confusion with p-values has plagued the research community and medical practitioners for decades. However, efforts to clarify it have been largely futile, in part, because intuitive yet mathematically rigorous educational materials are scarce. Additionally, the lack of a practical alternative to the p-value for guarding against randomness also plays a role. The p-value confusion is rooted in the misconception of significance and hypothesis testing. Most, including many statisticians, are unaware that p-values and significance testing formed by Fisher are incomparable to the hypothesis testing paradigm created by Neyman and Pearson. And most otherwise great statistics textbooks tend to cobble the two paradigms together and make no effort to elucidate the subtle but fundamental differences between them. The p-value is a practical tool gauging the “strength of evidence” against the null hypothesis. It informs investigators that a p-value of 0.001, for example, is stronger than 0.05. However, p-values produced in significance testing are not the probabilities of type I errors as commonly misconceived. For a p-value of 0.05, the chance a treatment does not work is not 5%; rather, it is at least 28.9%. Conclusions A long-overdue effort to understand p-values correctly is much needed. However, in medical research and practice, just banning significance testing and accepting uncertainty are not enough. Researchers, clinicians, and patients alike need to know the probability a treatment will or will not work. Thus, the calibrated p-values (the probability that a treatment does not work) should be reported in research papers.
Collapse
Affiliation(s)
- Jian Gao
- Department of Veterans Affairs, Office of Productivity, Efficiency and Staffing (OPES, RAPID), Albany, USA.
| |
Collapse
|
40
|
Sreekumar S, Cohen A, Gündüz D. Privacy-Aware Distributed Hypothesis Testing. Entropy (Basel) 2020; 22:e22060665. [PMID: 33286437 PMCID: PMC7517198 DOI: 10.3390/e22060665] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/01/2020] [Revised: 06/11/2020] [Accepted: 06/12/2020] [Indexed: 11/16/2022]
Abstract
A distributed binary hypothesis testing (HT) problem involving two parties, a remote observer and a detector, is studied. The remote observer has access to a discrete memoryless source, and communicates its observations to the detector via a rate-limited noiseless channel. The detector observes another discrete memoryless source, and performs a binary hypothesis test on the joint distribution of its own observations with those of the observer. While the goal of the observer is to maximize the type II error exponent of the test for a given type I error probability constraint, it also wants to keep a private part of its observations as oblivious to the detector as possible. Considering both equivocation and average distortion under a causal disclosure assumption as possible measures of privacy, the trade-off between the communication rate from the observer to the detector, the type II error exponent, and privacy is studied. For the general HT problem, we establish single-letter inner bounds on both the rate-error exponent-equivocation and rate-error exponent-distortion trade-offs. Subsequently, single-letter characterizations for both trade-offs are obtained (i) for testing against conditional independence of the observer's observations from those of the detector, given some additional side information at the detector; and (ii) when the communication rate constraint over the channel is zero. Finally, we show by providing a counter-example where the strong converse which holds for distributed HT without a privacy constraint does not hold when a privacy constraint is imposed. This implies that in general, the rate-error exponent-equivocation and rate-error exponent-distortion trade-offs are not independent of the type I error probability constraint.
Collapse
Affiliation(s)
- Sreejith Sreekumar
- Department of Electrical and Computer Engineering , Cornell University, Ithaca, NY 14850, USA
- Correspondence:
| | - Asaf Cohen
- The School of Electrical and Computer Engineering, Ben-Gurion University of the Negev, Beer-Sheva 8410501, Israel;
| | - Deniz Gündüz
- Department of Electrical and Electronic Engineering, Imperial College London, London SW72AZ, UK;
| |
Collapse
|
41
|
Rich LR, Patrick JA, Hamner MA, Ransom BR, Brown AM. A method for reducing animal use whilst maintaining statistical power in electrophysiological recordings from rodent nerves. Heliyon 2020; 6:e04143. [PMID: 32529085 DOI: 10.1016/j.heliyon.2020.e04143] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2019] [Revised: 11/21/2019] [Accepted: 06/02/2020] [Indexed: 12/29/2022] Open
Abstract
The stimulus evoked compound action potential, recorded from ex vivo nerve trunks such as the rodent optic and sciatic nerve, is a popular model system used to study aspects of nervous system metabolism. This includes (1) the role of glycogen in supporting axon conduction, (2) the injury mechanisms resulting from metabolic insults, and (3) to test putative benefits of clinically relevant neuroprotective strategies. We demonstrate the benefit of simultaneously recording from pairs of nerves in the same superfusion chamber compared with conventional recordings from single nerves. Experiments carried out on mouse optic and sciatic nerves demonstrate that our new recording configuration decreased the relative standard deviation from samples when compared with recordings from an equivalent number of individually recorded nerves. The new method reduces the number of animals required to produce equivalent Power compared with the existing method, where single nerves are used. Adopting this method leads to increased experimental efficiency and productivity. We demonstrate that reduced animal use and increased Power can be achieved by recording from pairs of rodent nerve trunks simultaneously.
Collapse
|
42
|
Russo L, Russo S. Search engines, cognitive biases and the man-computer interaction: a theoretical framework for empirical researches about cognitive biases in online search on health-related topics. Med Health Care Philos 2020; 23:237-246. [PMID: 32056071 DOI: 10.1007/s11019-020-09940-9] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
The widespread use of online search engines to answer the general public's needs for information has raised concerns about possible biases and the emerging of a 'filter bubble' in which users are isolated from attitude-discordant messages. Research is split between approaches that largely focus on the intrinsic limitations of search engines and approaches that investigate user search behavior. This work evaluates the findings and limitations of both approaches and advances a theoretical framework for empirical investigations of cognitive biases in online search activities about health-related topics. We aim to investigate the interaction between the user and the search engine as a whole. Online search activity about health-related topics is considered as a hypothesis-testing process. Two questions emerge: whether the retrieved information provided by the search engines are fit to fulfill their role as evidence, and whether the use of this information by users is cognitively and epistemologically valid and unbiased.
Collapse
Affiliation(s)
| | - Selena Russo
- Department of Surgery and Medicine, University of Milan-Bicocca, Milan, Italy.
- Behavioural Sciences Unit, Kids Cancer Centre, Level 1 South, Sydney Children's Hospital, High Street, Randwick, NSW, 2031, Australia.
| |
Collapse
|
43
|
Abstract
Correlation and association analyses are one of the most widely used statistical methods in research fields, including microbiome and integrative multiomics studies. Correlation and association have two implications: dependence and co-occurrence. Microbiome data are structured as phylogenetic tree and have several unique characteristics, including high dimensionality, compositionality, sparsity with excess zeros, and heterogeneity. These unique characteristics cause several statistical issues when analyzing microbiome data and integrating multiomics data, such as large p and small n, dependency, overdispersion, and zero-inflation. In microbiome research, on the one hand, classic correlation and association methods are still applied in real studies and used for the development of new methods; on the other hand, new methods have been developed to target statistical issues arising from unique characteristics of microbiome data. Here, we first provide a comprehensive view of classic and newly developed univariate correlation and association-based methods. We discuss the appropriateness and limitations of using classic methods and demonstrate how the newly developed methods mitigate the issues of microbiome data. Second, we emphasize that concepts of correlation and association analyses have been shifted by introducing network analysis, microbe-metabolite interactions, functional analysis, etc. Third, we introduce multivariate correlation and association-based methods, which are organized by the categories of exploratory, interpretive, and discriminatory analyses and classification methods. Fourth, we focus on the hypothesis testing of univariate and multivariate regression-based association methods, including alpha and beta diversities-based, count-based, and relative abundance (or compositional)-based association analyses. We demonstrate the characteristics and limitations of each approaches. Fifth, we introduce two specific microbiome-based methods: phylogenetic tree-based association analysis and testing for survival outcomes. Sixth, we provide an overall view of longitudinal methods in analysis of microbiome and omics data, which cover standard, static, regression-based time series methods, principal trend analysis, and newly developed univariate overdispersed and zero-inflated as well as multivariate distance/kernel-based longitudinal models. Finally, we comment on current association analysis and future direction of association analysis in microbiome and multiomics studies.
Collapse
Affiliation(s)
- Yinglin Xia
- Department of Medicine, University of Illinois at Chicago, Chicago, IL, United States.
| |
Collapse
|
44
|
Hu M, Crainiceanu C, Schindler MK, Dewey B, Reich DS, Shinohara RT, Eloyan A. Matrix decomposition for modeling lesion development processes in multiple sclerosis. Biostatistics 2020; 23:83-100. [PMID: 32318692 DOI: 10.1093/biostatistics/kxaa016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/01/2019] [Revised: 03/11/2019] [Accepted: 03/12/2020] [Indexed: 11/14/2022] Open
Abstract
Our main goal is to study and quantify the evolution of multiple sclerosis lesions observed longitudinally over many years in multi-sequence structural magnetic resonance imaging (sMRI). To achieve that, we propose a class of functional models for capturing the temporal dynamics and spatial distribution of the voxel-specific intensity trajectories in all sMRI sequences. To accommodate the hierarchical data structure (observations nested within voxels, which are nested within lesions, which, in turn, are nested within study participants), we use structured functional principal component analysis. We propose and evaluate the finite sample properties of hypothesis tests of therapeutic intervention effects on lesion evolution while accounting for the multilevel structure of the data. Using this novel testing strategy, we found statistically significant differences in lesion evolution between treatment groups.
Collapse
Affiliation(s)
- Menghan Hu
- Department of Biostatistics, Brown University, Providence, RI 02903, USA
| | - Ciprian Crainiceanu
- Department of Biostatistics, Johns Hopkins Bloomberg School of Public Health, Baltimore, MD 21205, USA
| | - Matthew K Schindler
- Translational Neuroradiology Section, Division of Neuroimmunology and Neurovirology, National Institute of Neurological Disorders and Stroke, National Institutes of Health, Bethesda, MD 20892, USA
| | - Blake Dewey
- Translational Neuroradiology Section, Division of Neuroimmunology and Neurovirology, National Institute of Neurological Disorders and Stroke, National Institutes of Health, Bethesda, MD 20892, USA and Department of Electrical and Computer Engineering, Johns Hopkins Whiting School of Engineering, Baltimore, MD 21218, USA
| | - Daniel S Reich
- Translational Neuroradiology Section, National Institute of Neurological Disorders and Stroke, National Institutes of Health, Bethesda, MD 20892, USA
| | - Russell T Shinohara
- Department of Biostatistics, Epidemiology, and Informatics, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA 19104, USA and Department of Radiology, Center for Biomedical Image Computing and Analytics, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA 19104, USA
| | - Ani Eloyan
- Department of Biostatistics, Brown University, Providence, RI 02903, USA
| |
Collapse
|
45
|
Baduashvili A, Evans AT, Cutler T. How to understand and teach P values: a diagnostic test framework. J Clin Epidemiol 2020; 122:49-55. [PMID: 32169596 DOI: 10.1016/j.jclinepi.2020.03.003] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2019] [Revised: 02/09/2020] [Accepted: 03/05/2020] [Indexed: 11/25/2022]
Abstract
OBJECTIVES The aim of the tutorial is to help educators address misconceptions about P values and provide a tool that can be used to teach a more contemporary interpretation. STUDY DESIGN AND SETTING A scripted tutorial using problem-based learning and a diagnostic test analogy to deconstruct the misunderstandings about P values and develop a more Bayesian approach to study interpretation. RESULTS A diagnostic test analogy is an effective teaching tool. Learners' understanding of Bayes' theorem in diagnostic testing can be used as a bridge to the realization that the prestudy probability of a true difference is crucial for study interpretation. The analogy has several caveats and shortcomings. The limitations of this analogy and the conceptual difficulties with the Bayesian study analyses are addressed. CONCLUSION P values do not provide the information many assume they do-they are not equivalent to a probability of a chance finding. This tutorial helps move learners from these incorrect notions to new insights.
Collapse
Affiliation(s)
- Amiran Baduashvili
- Section of Hospital Medicine, Division of General Internal Medicine, Weill Cornell Medical College, 525 East 68th Street, Box 331, New York, NY 10065, USA; Division of Hospital Medicine, University of Colorado School of Medicine, Aurora, CO, USA.
| | - Arthur T Evans
- Section of Hospital Medicine, Division of General Internal Medicine, Weill Cornell Medical College, 525 East 68th Street, Box 331, New York, NY 10065, USA
| | - Todd Cutler
- Section of Hospital Medicine, Division of General Internal Medicine, Weill Cornell Medical College, 525 East 68th Street, Box 331, New York, NY 10065, USA
| |
Collapse
|
46
|
Abstract
Preclinical studies using animals to study the potential of a therapeutic drug or strategy are important steps before translation to clinical trials. However, evidence has shown that poor quality in the design and conduct of these studies has not only impeded clinical translation but also led to significant waste of valuable research resources. It is clear that experimental biases are related to the poor quality seen with preclinical studies. In this chapter, we will focus on hypothesis testing type of preclinical studies and explain general concepts and principles in relation to the design of in vivo experiments, provide definitions of experimental biases and how to avoid them, and discuss major sources contributing to experimental biases and how to mitigate these sources. We will also explore the differences between confirmatory and exploratory studies, and discuss available guidelines on preclinical studies and how to use them. This chapter, together with relevant information in other chapters in the handbook, provides a powerful tool to enhance scientific rigour for preclinical studies without restricting creativity.
Collapse
|
47
|
Adams RH, Castoe TA. Probabilistic Species Tree Distances: Implementing the Multispecies Coalescent to Compare Species Trees Within the Same Model-Based Framework Used to Estimate Them. Syst Biol 2020; 69:194-207. [PMID: 31086978 DOI: 10.1093/sysbio/syz031] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2018] [Accepted: 05/02/2019] [Indexed: 11/14/2022] Open
Abstract
Despite the ubiquitous use of statistical models for phylogenomic and population genomic inferences, this model-based rigor is rarely applied to post hoc comparison of trees. In a recent study, Garba et al. derived new methods for measuring the distance between two gene trees computed as the difference in their site pattern probability distributions. Unlike traditional metrics that compare trees solely in terms of geometry, these measures consider gene trees and associated parameters as probabilistic models that can be compared using standard information theoretic approaches. Consequently, probabilistic measures of phylogenetic tree distance can be far more informative than simply comparisons of topology and/or branch lengths alone. However, in their current form, these distance measures are not suitable for the comparison of species tree models in the presence of gene tree heterogeneity. Here, we demonstrate an approach for how the theory of Garba et al. (2018), which is based on gene tree distances, can be extended naturally to the comparison of species tree models. Multispecies coalescent (MSC) models parameterize the discrete probability distribution of gene trees conditioned upon a species tree with a particular topology and set of divergence times (in coalescent units), and thus provide a framework for measuring distances between species tree models in terms of their corresponding gene tree topology probabilities. We describe the computation of probabilistic species tree distances in the context of standard MSC models, which assume complete genetic isolation postspeciation, as well as recent theoretical extensions to the MSC in the form of network-based MSC models that relax this assumption and permit hybridization among taxa. We demonstrate these metrics using simulations and empirical species tree estimates and discuss both the benefits and limitations of these approaches. We make our species tree distance approach available as an R package called pSTDistanceR, for open use by the community.
Collapse
Affiliation(s)
- Richard H Adams
- Department of Biology, University of Texas at Arlington, 501 S. Nedderman Dr., Arlington, TX 76019, USA
| | - Todd A Castoe
- Department of Biology, University of Texas at Arlington, 501 S. Nedderman Dr., Arlington, TX 76019, USA
| |
Collapse
|
48
|
Abstract
Phenotypic characterisation provides important information about novel crops that helps their developers to make technical and commercial decisions. Phenotypic characterisation comprises two activities. Product characterisation checks that the novel crop has the qualities of a viable product-the intended traits have been introduced and work as expected, and no unintended changes have been made that will adversely affect the performance of the final product. Risk assessment evaluates whether the intended and unintended changes are likely to harm human health or the environment. Product characterisation follows the principles of problem formulation, namely that the characteristics required in the final product are defined and criteria to decide whether the novel crop will have these properties are set. The hypothesis that the novel crop meets the criteria are tested during product development. If the hypothesis is corroborated, development continues, and if the hypothesis is falsified, the product is redesigned or its development is halted. Risk assessment should follow the same principles. Criteria that indicate the crop poses unacceptable risk should be set, and the hypothesis that the crop does not possess those properties should be tested. However, risk assessment, particularly when considering unintended changes introduced by new plant breeding methods such as gene editing, often ignores these principles. Instead, phenotypic characterisation seeks to catalogue all unintended changes by profiling methods and then proceeds to work out whether any of the changes are important. This paper argues that profiling is an inefficient and ineffective method of phenotypic characterisation for risk assessment. It discusses reasons why profiling is favoured and corrects some misconceptions about problem formulation.
Collapse
Affiliation(s)
- Alan Raybould
- Syngenta Crop Protection AG, Rosentalstrasse 67, 4002, Basel, Switzerland.
| |
Collapse
|
49
|
Yang Q, An X, Pan W. Computing and graphing probability values of pearson distributions: a SAS/IML macro. Source Code Biol Med 2020; 14:6. [PMID: 31889995 PMCID: PMC6923921 DOI: 10.1186/s13029-019-0076-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/09/2017] [Accepted: 11/22/2019] [Indexed: 11/10/2022]
Abstract
Background Any empirical data can be approximated to one of Pearson distributions using the first four moments of the data (Elderton WP, Johnson NL. Systems of Frequency Curves. 1969; Pearson K. Philos Trans R Soc Lond Ser A. 186:343–414 1895; Solomon H, Stephens MA. J Am Stat Assoc. 73(361):153–60 1978). Thus, Pearson distributions made statistical analysis possible for data with unknown distributions. There are both extant, old-fashioned in-print tables (Pearson ES, Hartley HO. Biometrika Tables for Statisticians, vol. II. 1972) and contemporary computer programs (Amos DE, Daniel SL. Tables of percentage points of standardized pearson distributions. 1971; Bouver H, Bargmann RE. Tables of the standardized percentage points of the pearson system of curves in terms of β1 and β2. 1974; Bowman KO, Shenton LR. Biometrika. 66(1):147–51 1979; Davis CS, Stephens MA. Appl Stat. 32(3):322–7 1983; Pan W. J Stat Softw. 31(Code Snippet 2):1–6 2009) available for obtaining percentage points of Pearson distributions corresponding to certain pre-specified percentages (or probability values; e.g., 1.0%, 2.5%, 5.0%, etc.), but they are little useful in statistical analysis because we have to rely on unwieldy second difference interpolation to calculate a probability value of a Pearson distribution corresponding to a given percentage point, such as an observed test statistic in hypothesis testing. Results The present study develops a SAS/IML macro program to identify the appropriate type of Pearson distribution based on either input of dataset or the values of four moments and then compute and graph probability values of Pearson distributions for any given percentage points. Conclusions The SAS macro program returns accurate approximations to Pearson distributions and can efficiently facilitate researchers to conduct statistical analysis on data with unknown distributions.
Collapse
Affiliation(s)
| | - Xinming An
- 2University of North Carolina at Chapel Hill, 27599, Chapel Hill, USA
| | - Wei Pan
- 1Duke University, Durham, 27710 USA
| |
Collapse
|
50
|
Abstract
Expression quantitative trait loci (eQTL) analysis identifies genetic variants that regulate the expression level of a gene. The genetic regulation may persist or vary in different tissues. When data are available on multiple tissues, it is often desired to borrow information across tissues and conduct an integrative analysis. Here we describe a multi-tissue eQTL analysis procedure, which improves the identification of different types of eQTL and facilitates the assessment of tissue specificity.
Collapse
Affiliation(s)
- Gen Li
- Department of Biostatistics, Mailman School of Public Health, Columbia University, New York, NY, USA.
| |
Collapse
|