Since the so-called “replication crisis”, the reproducibility of findings as a major quality criterion in natural science is the target of methodological discussions in many research areas (Munafò et al., 2017; Open Science Colloboration., 2015). To address this topic, the scientific community requires strategies to increase transparency. When preparing their data, researchers inevitably take more or less arbitrary decisions by picking a processing pipeline out of a large pool of (potentially) equally justifiable data processing and analysis choices (Brand, Dixon, & McBee, 2019). Usually there is more than one suitable path without a clear empirical evidence for the superiority of appropriateness of one specific path (Dragicevic, Jansen, Sarma, Kay, & Chevalier, 2019). These researchers’ degrees of freedom provoke a garden of forking paths (Gelman & Loken, 2014) which has been criticized in the current literature. Further, the lack of consensus about data processing and analysis pipelines used even within an area of research leads to difficulties in comparing findings across different studies. Consequently, the result of a single statistical model applied to a single data set generated via a specific preprocessing pipeline - as typically reported in research papers - is limited, since it doesn’t inform on the robustness of the result. To address the problem of choosing one data processing path while multiple suitable paths may exist, multiverse analysis has been proposed (Steegen, Tuerlinckx, Gelman, & Vanpaemel, 2016). Thereby researchers generate a ‘data multiverse’ from all or many possible decision paths in processing a single data set. The underlying idea of a data multiverse is that at every point the researchers have to take a decision, the multiverse grows by one more possible data set or by a whole branch of single data sets. This naturally leads to a multiverse of results based on this data multiverse even when applying an identical statistical model to all data sets (i.e., ‘data multiverse’). A similar approach is to set up a ‘model multiverse’ by applying different statistical models to a single data set (Donnelly, Brooks, & Homer, 2019; McBee, Brand, & Wallace E. Dixon, 2019; Patel, Burford, & Ioannidis, 2015). As a consequence, a data multiverse analysis illustrates the existence of possible outcomes based on equally justifiable choices within different preprocessing or transformation pipelines of a single data set (Stern, Arslan, Gerlach, & Penke, 2019). In case findings show homogeneity of results across the different forking paths, the robustness of an effect independent of the used preprocessing pipeline (for a data multiverse) or statistical pipeline (for a model multiverse), can be assumed. If inhomogeneity, however, is observed, this may inform us on potential boundary conditions (for instance inclusion of specific covariates). In this sense, we want to address a topical problem concerning skin conductance response (SCR) quantification approaches which arises from recent methodological discussions about comparability of methods and procedures. In human fear conditioning research the most commonly used psychophysiological index to infer fear learning are SCR. In practice, different methods to quantify SCR exist: First of all, the trough-to-peak (TTP) approach consists of quantifying the SCR amplitude as the difference of the peak score of the response and the preceding trough within specific time-windows and criteria (Boucsein, 1992). It can be conducted manually or semi-manually (i.e., computer-assisted) through trained raters and is supported in a fully computerized (though manually modifiable) way in the software Autonomate, which implements the TTP criteria (Boucsein, 1992) in computer-based algorithms (Green, Kragel, Fecteau, & LaBar, 2014). Second, computational approaches for SCR quantification exist with software programs such as Ledalab or PsPM providing computerized model-based SCR quantification (Bach, 2014; Staib, Castegnetti, & Bach, 2015). A third SCR quantification approach is what we here refer to as baseline correction approach (Pineles, Orr, & Orr, 2009): The baseline of each SCR is defined as the mean in a pre-CS period (conditioned stimulus). This mean is subtracted from the subsequent peak response in a post-CS period. Of note, the pre- and post-CS periods vary in the literature. In sum, there is no standard quantification approach for SCR, but all approaches mentioned are currently used interchangeably. As part of previous work, we conducted a systemic literature research (described in detail in: Lonsdorf et al., 2019) covering all publications (including e-pubs ahead of print) in PubMed from the 22nd September 2018 to the 22nd March 2019 to identify records which measure SCRs in human fear conditioning research. From 854 records identified through database searching, 50 records remained after the exclusion of review articles/ meta-analysis, rodent work, irrelevant topics, duplicates and no SCR measurement (for further information on screening and exclusion of records see the supplemental flow chart according to PRISMA guidelines (Liberati et al., 2009)). In these articles, TTP and baseline correction for SCR quantification were most common in the field with few studies employing computational approaches (an overview can be found in the additional bar chart). Yet, the “baseline correction category” is itself a particularly heterogeneous category as we identified 16 different ways to perform SCR quantification through baseline correction through our systematic literature search. These differed in the duration and length of the time windows used (yet, only partly explained by different experimental timings of the employed paradigm). Because TTP SCR quantification approaches are conducted semi-manually by independent raters they may be affected by the raters’ decision even with extensive rater-training. Here, we aim to investigate whether differences in the operationalization of SCR quantification through baseline correction and semi-manual TTP scoring as used in the field (described in detail in: Lonsdorf et al., 2019) result in divergent or convergent outcomes. Hence, we employ these different definitions in a multiverse analysis by applying all of them to a single preexisting data set to generate a data multiverse. Beside the frequentist approach we implement a Bayesian inference approach and compute Bayes Factors. Bayes Factors quantify the evidence for the absence of an effect or the absence of evidence for or against an effect. As our primary aim, we want to test for CS+/CS- discrimination within fear acquisition training (effect of main interest) as a prime example in this data multiverse based on different SCR quantification approaches. Our secondary aim is to use trait anxiety, as assessed by the STAI-Questionnaire (Spielberger, Gorsuch, & Lushene, 1983), as a case example for individual differences in fear conditioning to test robustness of a possible association between CS+/CS- discrimination in fear acquisition training and trait anxiety regarding different baseline correction approaches and TTP scoring in a data multiverse approach. This is important as individual difference effects are much more subtle and hence potentially more susceptible to data processing differences (Eysenck, 1963; Lonsdorf & Merz, 2017; Naveteur & Freixa i Baque, 1987). Findings regarding the association between trait anxiety as a criterion variable and differential SCRs in fear conditioning research show inconsistent results and were mainly assessed in small samples fear conditioning experiments (Indovina, Robbins, Núñez-Elizalde, Dunn, & Bishop, 2011; Lissek et al., 2005; Mineka & Oehlberg, 2008; Plehn & Peterson, 2002; Torrents-Rodas et al., 2013; larger sample: Sjouwerman, Scharfenort, & Lonsdorf, preprint). Hence we want to investigate the agreement between the baseline correction variants and TTP scoring from the multiverse approach using the Bland-Altman approach for method comparison (Bland & Altman, 1986, 1995). References Bach, D. R. (2014). A head-to-head comparison of SCRalyze and Ledalab, two model-based methods for skin conductance analysis. Biological Psychology, 103(1), 63–68. https://doi.org/10.1016/j.biopsycho.2014.08.006 Berth, M., & Singmann, H. (2020). Overview of MPT Multiverse: An Example Application. Retrieved from https://cran.r-project.org/web/packages/MPTmultiverse/vignettes/introduction-bayen_kuhlmann_2011.html Bland, J. M., & Altman, D. G. (1986). Statistical Methods for Assessing Agreement Between Two Methods of Clinical Measurement. The Lancet, 327(8476), 307–310. https://doi.org/10.1016/S0140-6736(86)90837-8 Bland, J. M., & Altman, D. G. (1995). Comparing methods of measurement: why plotting difference against standard method is misleading. The Lancet, 346(8982), 1085–1087. https://doi.org/10.1016/S0140-6736(95)91748-9 Bland, J. M., & Altman, D. G. (1999). Measuring agreement in method comparison studies. Statistical Methods in Medical Research, 8(2), 135–160. https://doi.org/10.1177/096228029900800204 Boucsein, W. (1992). Electrodermal activity. Retrieved from https://books.google.com/books?hl=de&lr=&id=6N6rnOEZEEoC&oi=fnd&pg=PR3&ots=B2pDkyGLMG&sig=vr7FnIhAtMiWYJzNLiyhRB6H340 Brand, R., Dixon, W., & McBee, M. (2019). Challenging the Link Between Early Childhood Television Exposure and Later Attention Problems: a Multiverse Analysis. ETSU Faculty Works. Retrieved from https://dc.etsu.edu/etsu-works/4902 Donnelly, S., Brooks, P. J., & Homer, B. D. (2019). Is there a bilingual advantage on interference-control tasks? A multiverse meta-analysis of global reaction time and interference cost. Psychonomic Bulletin and Review, 26(4), 1122–1147. https://doi.org/10.3758/s13423-019-01567-z Dragicevic, P., Jansen, Y., Sarma, A., Kay, M., & Chevalier, F. (2019). Increasing the transparency of research papers with explorable multiverse analyses. Conference on Human Factors in Computing Systems - Proceedings. https://doi.org/10.1145/3290605.3300295 Eysenck, H. J. (1963). Biological basis of personality. Transaction Publishers. Retrieved from https://link.springer.com/content/pdf/10.1038/1991031a0.pdf Gelman, A., & Loken, E. (2014). The garden of forking paths: Why multiple comparisons can be a problem, even when there is no “fishing expedition” or “p-hacking” and the research hypothesis was posited ahead of time. Psychological Bulletin, 140(5), 1272–1280. https://doi.org/dx.doi.org/10.1037/a0037714 Green, S. R., Kragel, P. A., Fecteau, M. E., & LaBar, K. S. (2014). Development and validation of an unsupervised scoring system (Autonomate) for skin conductance response analysis. International Journal of Psychophysiology, 91(3), 186–193. https://doi.org/10.1016/j.ijpsycho.2013.10.015 Indovina, I., Robbins, T. W., Núñez-Elizalde, A. O., Dunn, B. D., & Bishop, S. J. (2011). Fear-Conditioning Mechanisms Associated with Trait Vulnerability to Anxiety in Humans. Neuron, 69(3), 563–571. https://doi.org/10.1016/j.neuron.2010.12.034 Liberati, A., Altman, D. G., Tetzlaff, J., Mulrow, C., Gøtzsche, P. C., Ioannidis, J. P. A., … Moher, D. (2009). The PRISMA Statement for Reporting Systematic Reviews and Meta-Analyses of Studies That Evaluate Health Care Interventions: Explanation and Elaboration. PLoS Medicine, 6(7), e1000100. https://doi.org/10.1371/journal.pmed.1000100 Lissek, S., Powers, A. S., McClure, E. B., Phelps, E. A., Woldehawariat, G., Grillon, C., & Pine, D. S. (2005). Classical fear conditioning in the anxiety disorders: A meta-analysis. Behaviour Research and Therapy, 43(11), 1391–1424. https://doi.org/10.1016/j.brat.2004.10.007 Lonsdorf, T. B., Klingelhöfer-Jens, M., Andreatta, M., Beckers, T., Chalkia, A., Gerlicher, A., … Merz, C. J. (2019). Navigating the garden of forking paths for data exclusions in fear conditioning research. ELife, 8, 1–36. https://doi.org/10.7554/eLife.52465 Lonsdorf, T. B., & Merz, C. J. (2017). More than just noise: Inter-individual differences in fear acquisition, extinction and return of fear in humans - Biological, experiential, temperamental factors, and methodological pitfalls. Neuroscience and Biobehavioral Reviews, 80(July), 703–728. https://doi.org/10.1016/j.neubiorev.2017.07.007 Lykken, D. T. (1972). Range Correction Applied to Heart Rate and to GSR Data. Psychophysiology, 9(3), 373–379. https://doi.org/10.1111/j.1469-8986.1972.tb03222.x Lykken, David T., & Venables, P. H. (1971). Direct measurement of skin conductance: A proposal for standardization. Psychophysiology, 8(5), 656–672. https://doi.org/10.1111/j.1469-8986.1971.tb00501.x McBee, M., Brand, R., & Wallace E. Dixon, J. (2019). Challenging the Link Between Early Childhood Television Exposure and Later Attention Problems: A Multiverse Approach. https://doi.org/10.31234/OSF.IO/5HD4R Mineka, S., & Oehlberg, K. (2008). The relevance of recent developments in classical conditioning to understanding the etiology and maintenance of anxiety disorders. Acta Psychologica, 127(3), 567–580. https://doi.org/10.1016/j.actpsy.2007.11.007 Morey, R. D., & Rouder, J. N. (2018). BayesFactor 0.9.12-4.2. Comprehensive R Archive Network. Retrieved from http://cran.r-project.org/web/packages/ BayesFactor/index.html Munafò, M. R., Nosek, B. A., Bishop, D. V. M., Button, K. S., Chambers, C. D., Percie Du Sert, N., … Ioannidis, J. P. A. (2017). A manifesto for reproducible science. Nature Human Behaviour, 1(1), 1–9. https://doi.org/10.1038/s41562-016-0021 Naveteur, J., & Freixa i Baque, E. (1987). Individual differences in electrodermal activity as a function of subjects’ anxiety. Personality and Individual Differences, 8(5), 615–626. https://doi.org/10.1016/0191-8869(87)90059-6 Open Science Colloboration. (2015). Estimating the reproducibility of psychological science. Science, 349(6251). https://doi.org/10.1126/science.aac4716 Patel, C. J., Burford, B., & Ioannidis, J. P. A. (2015). Assessment of vibration of effects due to model specification can demonstrate the instability of observational associations. Journal of Clinical Epidemiology, 68(9), 1046–1058. https://doi.org/10.1016/j.jclinepi.2015.05.029 Pineles, S. L., Orr, M. R., & Orr, S. P. (2009). An alternative scoring method for skin conductance responding in a differential fear conditioning paradigm with a long-duration conditioned stimulus. Psychophysiology, 146(5), 984–995. https://doi.org/10.1038/jid.2014.371 Plehn, K., & Peterson, R. A. (2002). Anxiety sensitivity as a predictor of the development of panic symptoms, panic attacks, and panic disorder: A prospective study. Journal of Anxiety Disorders, 16(4), 455–474. https://doi.org/10.1016/S0887-6185(02)00129-9 Sjouwerman, R., Scharfenort, R., & Lonsdorf, T. B. (n.d.). Individual differences in predicting fear: Specificity to trait anxiety and mediation via amygdala activation (in revision). https://doi.org/https://www.biorxiv.org/content/10.1101/233528v2 Spielberger, C., Gorsuch, R., & Lushene, R. (1983). Manual for the State-Trait Anxiety Inventory. In Consulting Psychologists Press. Palo Alto, CA. Staib, M., Castegnetti, G., & Bach, D. R. (2015). Optimising a model-based approach to inferring fear learning from skin conductance responses. Journal of Neuroscience Methods, 255, 131–138. https://doi.org/10.1016/j.jneumeth.2015.08.009 Steegen, S., Tuerlinckx, F., Gelman, A., & Vanpaemel, W. (2016). Increasing Transparency Through a Multiverse Analysis. Perspectives on Psychological Science, 11(5), 702–712. https://doi.org/10.1177/1745691616658637 Stern, J., Arslan, R. C., Gerlach, T. M., & Penke, L. (2019). No robust evidence for cycle shifts in preferences for men’s bodies in a multiverse analysis: A response to Gangestad, Dinh, Grebe, Del Giudice, and Emery Thompson (2019). Evolution and Human Behavior, 40(6), 517–525. https://doi.org/10.1016/j.evolhumbehav.2019.08.005 Torrents-Rodas, D., Fullana, M. A., Bonillo, A., Caseras, X., Andión, O., & Torrubia, R. (2013). No effect of trait anxiety on differential fear conditioning or fear generalization. Biological Psychology, 92(2), 185–190. https://doi.org/10.1016/j.biopsycho.2012.10.006 Wagenmakers, E.-J., Love, J., Marsman, M., Jamil, T., Ly, A., Verhagen, J., … Morey, R. D. (2018). Bayesian inference for psychology. Part II: Example applications with JASP. Psychonomic Bulletin & Review, 25(1), 58–76. https ://doi.org/10.3758/s13423-017-1323-7