The proposed integrative experiment design approach consists of three steps: (1) Explicitly define the design space of the experiments in terms of features of the decision situation and the population sample, (2) systematically sample from that design space, and (3) build theories by quantifying the outcome heterogeneity over that space. This approach will guarantee commensurability between different experiments and findings and foster cumulative knowledge. The authors’ concept of “research cartography” is brilliant – the idea is to itemize, standardize, categorize, and quantify the information that we typically and only partially reveal in the Methodology and Discussion sections of our research papers. The image of a Wikidata-style database containing all experimental (and in fact, any other) social and behavioral knowledge is incredibly appealing! The Cooperation Databank, for instance, offers a glimpse of how such a database could look like (Spadaro et al., Reference Spadaro, Tiddi, Columbus, Jin, ten Teije and Balliet2022). Developing research cartography will help identify research gaps, established findings, and controversial problems. The approach will also aid the reuse and reanalysis of existing data to answer new research questions (Almaatouq, Rahimian, Burton, & Alhajri, Reference Almaatouq, Rahimian, Burton and Alhajri2022; Rand, Greene, & Nowak, Reference Rand, Greene and Nowak2012; Tsvetkova, Wagner, & Mao, Reference Tsvetkova, Wagner and Mao2018). In short, whether retrospective or prospective, a comprehensive and systematic research cartography will help consolidate knowledge and stimulate new research.
The integrative experiment design approach, however, presses further – steps (2) and (3) propose to consolidate research and theory-testing efforts by sampling and generalizing over many points in the experimental design space simultaneously, rather than “one-at-a-time.” Yet, these steps are not necessary for commensurability and more importantly, carry negative implications for diversity, innovation, and productive debate in academic research. There are several issues I would like to raise here.
First, the proposed paradigm threatens to entrench and exacerbate existing inequalities within and between scholarly communities. Participating in global research consortia may be open to many but who leads these consortia will likely befall on those with status, prestige, and funding. It is hard to overlook the fact that the authors speak from a position of privilege – they work at prestigious US universities, with access to hefty research funds and numerous PhD students and postdoctoral researchers. The large-scale research they propose is simply not accessible to many experimental researchers.
Second, the proposed paradigm aims to optimize efficiency in research but this is a misguided ideal. Academic research is not just about results but also about exploration and discovery, critique and debate, learning and training. Consolidating research activities in hierarchically structured labs or consortia with established protocols and routines may reduce labor costs but stifle entrepreneurship, critical thinking, and iconoclastic innovation. Based on some of the authors’ empirical examples, the complexities of group synergies imply that different problems would be best addressed by teams of different size and composition (Almaatouq, Alsobay, Yin, & Watts, Reference Almaatouq, Alsobay, Yin and Watts2021; Mao, Mason, Suri, & Watts, Reference Mao, Mason, Suri and Watts2016; Straub, Tsvetkova, & Yasseri, Reference Straub, Tsvetkova and Yasseri2023). This calls for independence, plurality, redundancy, and diversification of research effort, not consolidation.
Third, in the social and behavioral sciences, raising the question is often more important than finding out the answer. Much like the observer effect related to measuring physical systems, studying a social system changes it. Posing a specific social research problem can shape political debate, policy decisions, organization strategies, and collective behavior. Consolidating funding and research efforts forebodes a monopoly over setting research agendas and directions, the muffling of marginalized voices, the sidelining of localized problems, and the suppression of new perspectives and paradigms. Large-scale integrative experiments may be good for providing definitive evidence to integrate and reconcile existing theories but restricted when it comes to launching new research agendas.
Related to the latter issue, the proposed result-driven active-learning sampling strategy for experiments threatens to shift the focus to effects that are sizeable but not necessarily meaningful or important. Specifically, certain combinations of context and population features may be impossible or unlikely and hence, practically irrelevant. In short, the integrative experiment design approach does not alleviate and may even exacerbate the thorniest problem of experimental research – external validity. Explaining all variation is not always the best strategy for good or efficient science: The power of good general theories is not that they are universally true but that they apply to statistically likely/common situations and hence, they are useful.
I acknowledge that the authors present integrative experiment design as an additional, and not the only true, approach to experimentation in the social and behavioral sciences. I assumed an exaggeratedly antagonistic stance here to caution against consolidation. There are alternatives, such as adversarial collaboration (Killingsworth, Kahneman, & Mellers, Reference Killingsworth, Kahneman and Mellers2023; Mellers, Hertwig, & Kahneman, Reference Mellers, Hertwig and Kahneman2001), that can help reconcile contradictory findings without compromising debate, plurality, and diversity.
The proposed integrative experiment design approach consists of three steps: (1) Explicitly define the design space of the experiments in terms of features of the decision situation and the population sample, (2) systematically sample from that design space, and (3) build theories by quantifying the outcome heterogeneity over that space. This approach will guarantee commensurability between different experiments and findings and foster cumulative knowledge. The authors’ concept of “research cartography” is brilliant – the idea is to itemize, standardize, categorize, and quantify the information that we typically and only partially reveal in the Methodology and Discussion sections of our research papers. The image of a Wikidata-style database containing all experimental (and in fact, any other) social and behavioral knowledge is incredibly appealing! The Cooperation Databank, for instance, offers a glimpse of how such a database could look like (Spadaro et al., Reference Spadaro, Tiddi, Columbus, Jin, ten Teije and Balliet2022). Developing research cartography will help identify research gaps, established findings, and controversial problems. The approach will also aid the reuse and reanalysis of existing data to answer new research questions (Almaatouq, Rahimian, Burton, & Alhajri, Reference Almaatouq, Rahimian, Burton and Alhajri2022; Rand, Greene, & Nowak, Reference Rand, Greene and Nowak2012; Tsvetkova, Wagner, & Mao, Reference Tsvetkova, Wagner and Mao2018). In short, whether retrospective or prospective, a comprehensive and systematic research cartography will help consolidate knowledge and stimulate new research.
The integrative experiment design approach, however, presses further – steps (2) and (3) propose to consolidate research and theory-testing efforts by sampling and generalizing over many points in the experimental design space simultaneously, rather than “one-at-a-time.” Yet, these steps are not necessary for commensurability and more importantly, carry negative implications for diversity, innovation, and productive debate in academic research. There are several issues I would like to raise here.
First, the proposed paradigm threatens to entrench and exacerbate existing inequalities within and between scholarly communities. Participating in global research consortia may be open to many but who leads these consortia will likely befall on those with status, prestige, and funding. It is hard to overlook the fact that the authors speak from a position of privilege – they work at prestigious US universities, with access to hefty research funds and numerous PhD students and postdoctoral researchers. The large-scale research they propose is simply not accessible to many experimental researchers.
Second, the proposed paradigm aims to optimize efficiency in research but this is a misguided ideal. Academic research is not just about results but also about exploration and discovery, critique and debate, learning and training. Consolidating research activities in hierarchically structured labs or consortia with established protocols and routines may reduce labor costs but stifle entrepreneurship, critical thinking, and iconoclastic innovation. Based on some of the authors’ empirical examples, the complexities of group synergies imply that different problems would be best addressed by teams of different size and composition (Almaatouq, Alsobay, Yin, & Watts, Reference Almaatouq, Alsobay, Yin and Watts2021; Mao, Mason, Suri, & Watts, Reference Mao, Mason, Suri and Watts2016; Straub, Tsvetkova, & Yasseri, Reference Straub, Tsvetkova and Yasseri2023). This calls for independence, plurality, redundancy, and diversification of research effort, not consolidation.
Third, in the social and behavioral sciences, raising the question is often more important than finding out the answer. Much like the observer effect related to measuring physical systems, studying a social system changes it. Posing a specific social research problem can shape political debate, policy decisions, organization strategies, and collective behavior. Consolidating funding and research efforts forebodes a monopoly over setting research agendas and directions, the muffling of marginalized voices, the sidelining of localized problems, and the suppression of new perspectives and paradigms. Large-scale integrative experiments may be good for providing definitive evidence to integrate and reconcile existing theories but restricted when it comes to launching new research agendas.
Related to the latter issue, the proposed result-driven active-learning sampling strategy for experiments threatens to shift the focus to effects that are sizeable but not necessarily meaningful or important. Specifically, certain combinations of context and population features may be impossible or unlikely and hence, practically irrelevant. In short, the integrative experiment design approach does not alleviate and may even exacerbate the thorniest problem of experimental research – external validity. Explaining all variation is not always the best strategy for good or efficient science: The power of good general theories is not that they are universally true but that they apply to statistically likely/common situations and hence, they are useful.
I acknowledge that the authors present integrative experiment design as an additional, and not the only true, approach to experimentation in the social and behavioral sciences. I assumed an exaggeratedly antagonistic stance here to caution against consolidation. There are alternatives, such as adversarial collaboration (Killingsworth, Kahneman, & Mellers, Reference Killingsworth, Kahneman and Mellers2023; Mellers, Hertwig, & Kahneman, Reference Mellers, Hertwig and Kahneman2001), that can help reconcile contradictory findings without compromising debate, plurality, and diversity.
Financial support
This research received no specific grant from any funding agency, commercial, or not-for-profit sectors.
Competing interest
None.