We use cookies to distinguish you from other users and to provide you with a better experience on our websites. Close this message to accept cookies or find out how to manage your cookie settings.
To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Survey research is a method commonly used to understand what members of a population think, feel, and do. This chapter uses the total survey error perspective and the fitness for use perspective to explore how biasing and variable errors occur in surveys. Coverage error and sample frames, nonprobability samples and web panels, sampling error, nonresponse rates and nonresponse bias, and sources of measurement error are discussed. Different pretesting methods and modes of data collection commonly used in surveys are described. The chapter concludes that survey research is a tool that social psychologists may use to improve the generalizability of studies, to evaluate how different populations react to different experimental conditions, and to understand patterns in outcomes that may vary over time, place, or people.
This chapter provides an overview of methods for data collection in Conversation Analysis and practical advice on collecting interactional data. We touch on several recurrent issues that researchers encounter in the process. These issues include accessing data; the use of existing data (including user-uploaded, like YouTube); navigating gatekeepers in accessing a setting; building trust with members of a setting; building ethnographic understanding of activities under examination; obtaining ethical approvals; protecting privacy of participants; methods and materials for informed consent (including with populations with diminished capacities); devising a recording schedule; deciding when/how often to record; selecting the right quantity and type of recording equipment; considerations of spatial and audio environments; the use of alternative technologies for recording; recording mediated interactions; procedures and check-lists for before recording; positioning and framing the camera; when to press record and when to press stop; navigating the presence of the researcher-recorder on site; and gathering supplementary documentation from the setting.
While the preceding chapters of the Handbook have focused on practical skills in CA research methods, this chapter looks towards the path ahead. A diverse group of conversation analysts were asked to outline possible projects, point readers toward un- or under-described interactional phenomena, and discuss persistent issues in the field. The contributions address future advances in data collection, specific interactional practices, the complex interplay between language and the body, and cross-cultural and crosslinguistic comparisons, among other issues. The chapter concludes with a concise reiteration of the bedrock principle that underpins all CA research methods.
Conversation-analytic (CA) research projects have begun to involve the collection of interaction data in laboratory settings, as opposed to field settings, not for the purpose of experimentation, but in order to systematically analyze interactional phenomena that are elusive, not in the sense of being rare (i.e., ‘seldom occurring’), but in the sense of not being reliably or validly detected by analysts in the field using relatively standard recording equipment. This chapter (1) describes two, CA, methodological mandates – ‘maintaining mundane realism’ and ‘capturing the entirety of settings’ features’ – and their tensions; (2) provides four examples of elusive phenomena that expose these tensions, including gaze orientation, blinking, phonetic features during overlapping talk, and inhaling; and (3) discusses analytic ramifications of elusive phenomena, and provides a resultant series of data collection recommendations for both field and lab settings.
Once the theory is specified and an operationalization has been chosen for the nodes and links, the next step is to acquire the data. This chapter goes deep into issues that arise when designing surveys to collect data. Although this is not the only method of data collection, it is one that illuminates issues that pertain to all others. This chapter covers the practical question of how to use surveys to elicit network information. The advice leans heavily on a well-formulated theory.
This chapter provides an overview of selected studies assessing technology-aided programs to promote independent leisure and communication or combinations of independent leisure, communication, and daily activities in people with mild to moderate intellectual disability often associated with sensory and/or motor impairments. The studies included in the overview offer an opportunity to describe the development of those programs, the technology solutions used to support them, and their outcomes in terms of participants’ independent performance. Following the presentation of the programs and their outcomes, the discussion focuses on three main issues: (a) effectiveness of the programs and methodological considerations, (b) accessibility and affordability of the programs, and (c) implications of the programs for professionals working in daily contexts. With regard to the last issue, an effort was made to examine ethical and moral questions that may accompany the possible decisions of professionals to adopt those programs in daily contexts.
In the previous chapters we have considered the ‘nuts and bolts’ of epidemiology. In this and the next few chapters we look at how epidemiology is used in practice to improve public health. We start with ‘surveillance’ because without timely information on emerging and changing health problems, public health action can be paralysed or, at best, inefficient. In this chapter we discuss the design and use of surveillance systems that enable health officials to detect new risks and diseases such as mpox promptly, track known diseases and health problems, and generate data needed for effective health planning and resource allocation.
A user-friendly introductory guide to the empirical study of social networks. Jennifer M. Larson presents the fundamentals of social networks in an intuition-forward way which guides theory-driven research design. Substantial attention is devoted to a framework for developing a network theory that will steer data collection to be maximally informative and minimally frustrating. Other features include: Coverage of a range of practical topics including selecting operationalizations, cutting survey costs, and cleaning data; A tutorial for getting started in analyzing networks in R; Technical sections full of examples, points to hone intuition, and practice problems with solutions. Designing Empirical Social Networks Research will be a valuable tool for advanced undergraduates, Ph.D. students in the social sciences, especially political science, and researchers across the social sciences who are new to the study of networks.
Taking a simplified approach to statistics, this textbook teaches students the skills required to conduct and understand quantitative research. It provides basic mathematical instruction without compromising on analytical rigor, covering the essentials of research design; descriptive statistics; data visualization; and statistical tests including t-tests, chi-squares, ANOVAs, Wilcoxon tests, OLS regression, and logistic regression. Step-by-step instructions with screenshots are used to help students master the use of the freely accessible software R Commander. Ancillary resources include a solutions manual and figure files for instructors, and datasets and further guidance on using STATA and SPSS for students. Packed with examples and drawing on real-world data, this is an invaluable textbook for both undergraduate and graduate students in public administration and political science.
Research studies involving human subjects require collection of and reporting on demographic data related to race and ethnicity. However, existing practices lack standardized guidelines, leading to misrepresentation and biased inferences and conclusions for underrepresented populations in research studies. For instance, sometimes there is a misconception that self-reported racial or ethnic identity may be treated as a biological variable with underlying genetic implications, overlooking its role as a social construct reflecting lived experiences of specific populations. In this manuscript, we use the We All Count data equity framework, which organizes data projects across seven stages: Funding, Motivation, Project Design, Data Collection, Analysis, Reporting, and Communication. Focusing on data collection and analysis, we use examples – both real and hypothetical – to review common practice and provide critiques and alternative recommendations. Through these examples and recommendations, we hope to provide the reader with some ideas and a starting point as they consider embedding a lens of justice, equity, diversity, and inclusivity from research conception to dissemination of findings.
Scholars often use monetary incentives to boost participation rates in online surveys. This technique follows existing literature from western countries, which suggests egoistic incentives effectively boost survey participation. Positing that incentives’ effectiveness vary by country context, we tested this proposition through an experiment in Australia, India, and the USA. We compared three types of monetary lotteries to narrative and altruistic appeals. We find that egoistic rewards are most effective in the USA and to some extent, in Australia. In India, respondents are just as responsive to altruistic incentives as to egoistic incentives. Results from an adapted dictator game corroborate these patterns. Our results caution scholars against exporting survey participation incentives to areas where they have not been tested.
How do international crises unfold? We conceptualize international relations as a strategic chess game between adversaries and develop a systematic way to measure pieces, moves, and gambits accurately and consistently over a hundred years of history. We introduce a new ontology and dataset of international events called ICBe based on a very high-quality corpus of narratives from the International Crisis Behavior (ICB) Project. We demonstrate that ICBe has higher coverage, recall, and precision than existing state of the art datasets and conduct two detailed case studies of the Cuban Missile Crisis (1962) and the Crimea-Donbas Crisis (2014). We further introduce two new event visualizations (event iconography and crisis maps), an automated benchmark for measuring event recall using natural language processing (synthetic narratives), and an ontology reconstruction task for objectively measuring event precision. We make the data, supplementary appendix, replication material, and visualizations of every historical episode available at a companion website crisisevents.org.
This study serves as an exemplar to demonstrate the scalability of a research approach using survival analysis applied to general practice electronic health record data from multiple sites. Collection of these data, the subsequent analysis, and the preparation of practice-specific reports were performed using a bespoke distributed data collection and analysis software tool.
Background:
Statins are a very commonly prescribed medication, yet there is a paucity of evidence for their benefits in older patients. We examine the relationship between statin prescriptions for general practice patients over 75 and all-cause mortality.
Methods:
We carried out a retrospective cohort study using survival analysis applied to data extracted from the electronic health records of five Australian general practices.
Findings:
The data from 8025 patients were analysed. The median duration of follow-up was 6.48 years. Overall, 52 015 patient-years of data were examined, and the outcome of death from any cause was measured in 1657 patients (21%), with the remainder being censored. Adjusted all-cause mortality was similar for participants not prescribed statins versus those who were (HR 1.05, 95% CI 0.92–1.20, P = 0.46), except for patients with diabetes for whom all-cause mortality was increased (HR = 1.29, 95% CI: 1.00–1.68, P = 0.05). In contrast, adjusted all-cause mortality was significantly lower for patients deprescribed statins compared to those who were prescribed statins (HR 0.81, 95% CI 0.70–0.93, P < 0.001), including among females (HR = 0.75, 95% CI: 0.61–0.91, P < 0.001) and participants treated for secondary prevention (HR = 0.72, 95% CI: 0.60–0.86, P < 0.001). This study demonstrated the scalability of a research approach using survival analysis applied to general practice electronic health record data from multiple sites. We found no evidence of increased mortality due to statin-deprescribing decisions in primary care.
This chapter describes lab verification and clinical validation of tests for the detection of SARS-CoV-2. As new SARS-CoV-2 tests were being developed early in the pandemic, extensive lab verification studies to “test the tests” were conducted at ACME POCT at Emory University. Initial testing was performed in a Biosafety Level 3 facility to determine if the assays could detect propagated SARS-CoV-2 in ideal conditions and evaluate the specificity of these tests. We then describe the establishment of a Biorepository to bank SARS-CoV-2 variant samples and use these samples to determine whether tests could detect new variants with equal sensitivity as the original wild-type virus. This chapter also describes the clinical validation of tests using samples collected from individuals at testing centers. The clinical validation core requires careful planning for staffing and personnel training, semi-permanent and mobile clinical sites, defining inclusion and exclusion parameters, and data collection and reporting. Our experience demonstrated the importance of developing strong relationships with academic and private partners to facilitate clinical site setup, marketing, and purchasing.
Policymakers and scholars have long proposed that willingness to participate in armed conflict is influenced by citizens' income-earning opportunities. Testing this opportunity cost mechanism has led to mixed results. One reason for this might be the fact that current proxies can also serve as indicators to test grievance-based theories. In this study, we construct a more suitable measure. We use crop calendars and crop location data to build an index of agricultural idle time for first administration units on the African continent from 1990 to 2017. We test the explanatory power of this measure by examining its relationship with armed conflict. Our results show that agricultural idle time increases the probability of observing armed conflict by more than 20 percent.
Archaeologists frequently use written guidelines such as site manuals, recording forms, and digital prompts during excavations to create usable data within and across projects. Most written guidelines emphasize creating either standardized datasets or narrative summaries; however, previous research has demonstrated that the resulting datasets are often difficult to (re)use. Our study analyzed observations and interviews conducted with four archaeological excavation teams, as well as interviews with archaeological data reusers, to evaluate how archaeologists use and implement written guidelines. These excavation team and reuser experiences suggest that archaeologists need more specific best practices to create and implement written guidelines that improve the quality and usability of archaeological data. We present recommendations to improve written guidelines that focus on a project's methods, end-of-season documentation, and naming practices. We also present a Written Guidelines Checklist to help project directors improve their written guidelines before, during, and after fieldwork as part of a collaborative process. Ideally, these best practices for written guidelines will make it easier for team members and future reusers to incorporate their own and others’ archaeological data into their research.
Chapter 3 shows why the contracts model doesn’t work: consent is absent in the information economy. Privacy harm can’t be seen as a risk that people accept in exchange for a service. Inferences, relational data, and de-identified data aren’t captured by consent provisions. Consent is unattainable in the information economy more broadly because the dynamic between corporations and users is plagued with uneven knowledge, inequality, and a lack of choices. Data harms are collective and unknowable, making individual choices to reduce them impossible. Worse, privacy has a moral hazard problem: corporations have incentives to behave against our best interests, creating profitable harms after obtaining agreements. Privacy’s moral hazard leads to informational exploitation. One manifestation of valid consent in the information economy are consent refusals. We can consider them by thinking of people’s data as part of them, as their bodies are.
The role of lay health workers in data collection for clinical and translational research studies is not well described. We explored lay health workers as data collectors in clinical and translational research studies. We also present several methods for examining their work, i.e., qualitative interviews, fidelity checklists, and rates of unusable/missing data.
Methods:
We conducted 2 randomized, controlled trials that employed lay health research personnel (LHR) who were employed by community-based organizations. In one study, n = 3 Latina LHRs worked with n = 107 Latino diabetic participants. In another study, n = 6 LHR worked with n = 188 Cambodian American refugees with depression. We investigated proficiency in biological, behavioral, and psychosocial home-based data collection conducted by LHR. We also conducted in-depth interviews with lay LHR to explore their experience in this research role. Finally, we described the training, supervision, and collaboration for LHR to be successful in their research role.
Results:
Independent observers reported a very high degree of fidelity to technical data collection protocols (>95%) and low rates of missing/unusable data (1.5%–11%). Qualitative results show that trust, training, communication, and supervision are key and that LHR report feeling empowered by their role. LHR training included various content areas over several weeks with special attention to LHR and participant safety. Training and supervision from both the academic researchers and the staff at the community-based organizations were necessary and had to be well-coordinated.
Conclusions:
Carefully selected, trained, and supervised LHRs can collect sophisticated data for community-based clinical and translational research.
Choosing an appropriate electronic data capture system (EDC) is a critical decision for all randomized controlled trials (RCT). In this paper, we document our process for developing and implementing an EDC for a multisite RCT evaluating the efficacy and implementation of an enhanced primary care model for individuals with opioid use disorder who are returning to the community from incarceration.
Methods:
Informed by the Knowledge-to-Action conceptual framework and user-centered design principles, we used Claris Filemaker software to design and implement CRICIT, a novel EDC that could meet the varied needs of the many stakeholders involved in our study.
Results:
CRICIT was deployed in May 2021 and has been continuously iterated and adapted since. CRICIT’s features include extensive participant tracking capabilities, site-specific adaptability, integrated randomization protocols, and the ability to generate both site-specific and study-wide summary reports.
Conclusions:
CRICIT is highly customizable, adaptable, and secure. Its implementation has enhanced the quality of the study’s data, increased fidelity to a complicated research protocol, and reduced research staff’s administrative burden. CRICIT and similar systems have the potential to streamline research activities and contribute to the efficient collection and utilization of clinical research data.
This chapter describes the process of creating and annotating a corpus. This process involves, for instance, collecting data (speech and writing), transcribing recorded speech, and adding annotation, markup indicating in a conversation, for instance, when one person’s speech overlaps another speaker. While written texts are relatively easy to collect – most writing is readily available in digital formats – speech, especially spontaneous conversations, has to be transcribed, though voice recognition software has made progress in automating the transcription of certain kinds of speech, such as monologues. Other stages of building a corpus are also discussed, ranging from the administrative (keeping records of texts collected) to transcribing recordings of speech. The chapter concludes with a description of various kinds of textual markup and linguistic annotation that can be added to texts. Topics discussed include how to create a “header” for a particular text. Headers contain various kinds of information. For written texts, the header would include, for instance, the title of the text; the author(s); if published, where it was published. Other textual markup is internal to the text, and in a spoken text would include such information as speaker IDs, and the beginnings and ends of overlapping speech.