Skip to main content Accessibility help
×
Hostname: page-component-78c5997874-xbtfd Total loading time: 0 Render date: 2024-11-13T01:31:14.631Z Has data issue: false hasContentIssue false

13 - Publishing Your Research

from Part III - Your Research/Academic Career

Published online by Cambridge University Press:  21 July 2022

Mitchell J. Prinstein
Affiliation:
University of North Carolina, Chapel Hill

Summary

We focus on preparing a manuscript for publication, emphasizing description, explanation, and contextualization.Description is the most straightforward, providing study details.Explanation is more demanding, as it refers to presenting the rationale of several facets of the study.Contextualization moves a step furtherfrom description and addresses how the study fits in the context of other studies and the knowledge base more generally.Authors are often frustrated that their study and manuscript was not understood or appreciated.No doubt this is often true, but the responsibility lies with us as authors.There is much an author can do to make a compelling case for the study.One summary question to guide an author in preparing a manuscript might be, “Why is this study important, needed, or especially interesting?”The answer serves as the basis of the Introduction and Discussion.We discuss each section of a manuscript, what to include and explain, and how the gestalt can present a storyline that connects the sections.The chapter also discusses guidelines for conducting, evaluating, and reporting research and the journal submission and manuscript review process.

Type
Chapter
Information
The Portable Mentor
Expert Guide to a Successful Career in Psychology
, pp. 235 - 265
Publisher: Cambridge University Press
Print publication year: 2022
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NC
This content is Open Access and distributed under the terms of the Creative Commons Attribution licence CC-BY-NC 4.0 https://creativecommons.org/cclicenses/

A key characteristic of science is the accumulation of knowledge. This accumulation depends not only on the completion of research but also on preparation of reports that disseminate the results. Consequently, publication of research is an essential part of science. Publication can serve other goals as well. Preparing a manuscript for publication helps the investigator to consider the current study in a broader context and chart a course for a series of studies. In addition, many professional and career goals are served by publishing one’s research. Publication of one’s research signals a level of competence and mastery that includes developing an idea, designing, executing and completing the study, analyzing the results, preparing a written report, submitting it for publication, and traversing the peer-review process. This chapter focuses on publishing one’s research. The topics include preparing a manuscript, selecting a publication outlet, submitting the manuscript for review, and revising the manuscript as needed for publication.

There are many outlets to communicate the results of one’s research. Prominent among these are presentations at professional meetings, chapters in edited books, full-length books, and professional journals. Journal publication, the focus of this chapter, holds special status because it is the primary outlet for original research. In terms of one’s career, journal publication also plays a special role primarily because articles accepted for publication usually have undergone peer review. Acceptance and publication attest to the views of one’s peers that there is merit in the work. For any given article, only a few peers (one editor, two to three reviewers) may actually see the manuscript. Multiple publications add to this and after a few publications one can assume there is a building consensus about one’s work, i.e., others view the contributions as important and worthy of publication.

1. Preparing a Manuscript for Publication

1.1 Writing the Article

A central goal of scientific writing is to convey what was actually done so that the methods and procedures can be replicated. Concrete, specific, operational, objective, and precise are some of the characteristics that describe the writing style. The effort to describe research in concrete and specific ways is critically important. However, the task of the author goes well beyond description.

Preparation of the report for publication involves three interrelated tasks that I refer to as description, explanation, and contextualization. Failure to appreciate or to accomplish these tasks serves as a main source of frustration for authors, as their papers traverse the process of manuscript review toward journal publication. Description is the most straightforward task and includes providing details of the study. Even though this is an obvious requirement of the report, basic details often are omitted in published articles (e.g., sex, socioeconomic status, and race of the participants; means and standard deviations) (e.g., Reference Case and SmithCase & Smith, 2000; Reference Gerber, Arceneaux, Boudreau, Dowling, Hillygus, Palfrey, Biggers and HendryGerber et al., 2014; Reference Tate, Perdices, Rosenkoetter, McDonald, Togher, Shadish, Horner, Kratochwill, Barlow, Kazdin, Sampson, Shamseer and SampsonTate et al., 2016). Omission of basic details can hamper scientific progress. If a later study fails to replicate the findings, it could be because the sample is very different along some dimension or characteristic. Yet, we cannot surmise that without knowing at least basic details of the sample in both studies. If a study does repeat the findings, that is important, but is the new finding an extension to a new type of sample? Again, we need basic information in the studies to allow such comparisons.

Explanation is more demanding insofar as this refers to presenting the rationale of several facets of the study. The justification, decision-making process, and the connections between the decisions and the goals of the study move well beyond description. Here the reader of the manuscript has access to the author decision points. There are numerous decision points pertaining to such matters as selecting the sample, choosing among many options of how to test the idea, selecting the measures, and including various control and comparison groups. The author is obliged to explain why the specific options elected are well suited to the hypotheses or the goals of the study. There is a persuasion feature that operates here. The author of the manuscript is persuaded that the decisions are reasonable ways to address the overriding research question. Now the author must convey that to persuade the reader. In other words, explanation conveys why the procedures, measures, and so on were selected, but that explanation ought to be cogent, persuasive, and above all explicit. We do not want the reader to think, “This is an important research question, but why study it that way?” And in many cases, the related prior question of the same ilk emerges, why do we even need this study or why is study important? For the many decision points beginning with selection of the research question, these are very reasonable questions that the author ought to anticipate and pre-empt.

Finally, contextualization moves one step further away from description and addresses how the study fits in the context of other studies and in the knowledge base more generally. This latter facet of the article preparation reflects such lofty notions as scholarship and perspective, because the author places the descriptive and explanatory material into a broader context. Essentially, the author is making the case for the study based on the knowledge base. Relatively vacuous claims (e.g., this is the first study of this or the first study to include this or that control condition or measure) are rarely a strong basis for the study and often means or is interpreted as meaning that the author could not come up with something better. Without context, any “first” is not very important by itself. Indeed, it is easy to be first for a topic that is not very important and has been purposely neglected. We need a more compelling rationale.

For example, if this study is done on why people commit suicide we need the context of why this specific study ought to be done and where in the puzzle of understanding this piece fits. Perhaps prior research omitted some critical control procedure; perhaps there is a special group that has a novel characteristic that reduces (or increases) the likelihood of suicide that would inform the field in unique ways; or perhaps some new twist on a theory or intervention will have clear implications for reducing suicide attempts. These and other such comments convey three points that are wise to address: (1) there is a gap in knowledge, (2) that gap is important, and (3) that gap will be filled in whole or in part by this study.

1.2 General Comments

The three components I identified vary in difficulty. When individuals write their first project for publication, they focus heavily on the descriptive part to make sure all the material and sections are included. And this part is fundamental. Explanation and contextualization are much more difficult. Explanation requires having considered options and conveying to the reader why the one selection was a good choice. Yet one’s first study is often with or from an advisor who has made these decisions and the bases of these decisions might be buried in one of the advisor’s other articles but otherwise is tacit. As authors we need to be prepared for other scientists looking at our paper and doing their job by asking, “Why on earth did we [use: that population, measure, control condition, means of data evaluation, and so on]”. These are not only legitimate questions but are central to science.

Contextualization is even more difficult. Contextualization benefits from experience, scholarship, time, and knowledge of as many related areas of work as one can bring to bear. How is the study connected to the literature or topic, how does it relate to theory, to other disciplines, to a critical problem we ought to care about or that is now facing society? The puzzle analogy might help. A given study is one puzzle piece and merely showing that piece to someone is not inherently interesting. It may be inherently boring. Yet, the piece becomes more interesting as all the other pieces are shown (e.g., the outside box with a full photo of the puzzle) and even more interesting, fascinating actually, if one can paint a verbal picture of the whole puzzle and show how one or two pieces are needed and this study is that part! Explanation gives the rationales for decisions; contextualization determines whether the study is compelling or not. Authors often complain that the reviewers did not understand, “get it,” appreciate the importance of their study. The authors are usually completely right, but guess whose responsibility that is?

The extent to which description, explanation, and contextualization are accomplished increases the likelihood that the report will be viewed as a publishable article and facilitates integration of the report into the knowledge base. Guidelines are provided later in the chapter to convey these tasks more concretely in the preparation and evaluation of research reports. The guidelines focus on the logic of the study, the interrelations of the different sections, the rationale for specific procedures and analyses, the strengths and limitations, and where the study fits in the knowledge base. Consider main sections of the manuscript that are prepared for journal publication and how these components can be addressed.Footnote 1

2. Sections of an Article

2.1 Title

The title of an article includes the key variables, focus, and population with an economy of words. The special features of the study are included to convey the focus immediately to potential readers. It is critical here to be direct, clear, and concise (e.g., “Memory loss and gains associated with aging” or “Predictors of drug use and abuse among adolescents”). These examples are especially concise. Ordinarily an author is encouraged to fit the title within 10–12 words. The words ought to be selected carefully. Titles occasionally are used to index articles in large databases. Words that are not needed or that say little (e.g., “preliminary findings,” “implications,” “new findings”) might be more judiciously replaced by substantive or content words (e.g., among preschool children, the elderly; consequences for sleep and stress) that permit the article to be indexed more broadly than it otherwise would have been.

Occasionally, comments about the method are included in the title or more commonly in the subtitle. Terms like “a pilot study” or “preliminary report” may have many different meanings, such as the fact that this is an initial or interim report of a larger research program. These words could also be gently preparing readers for some methodological surprises and even tell us not to expect too much from the design. These qualifying terms might be accurate, but they implicitly apologize or ask for mercy as well. Better to give a strong title and in the write up give the explanation (decision making) and contexts to convey why this study was done, where it fits in the scheme of this literature, and why this was important. No apologies needed; just let the reader know your thinking on the matter. Although I am reluctant to boast, my dissertation won a prize for the best qualifying terms in a title. (In the subtitle of my dissertation, I conveyed this as: “A pre-preliminary, tentative, exploratory pilot study©.”)

In some cases, terms are added to the study such as, “A Controlled Investigation,” which moves our expectation in the other direction, namely, that the present study is somehow well conducted and controlled, and perhaps by implication stands in contrast to other studies in the field (or in the author’s repertoire). Usually words noting that the investigation is controlled are not needed unless this is truly a novel feature of research on the topic. Some words when added can be important because they are novel. An example would be the subheading, “A replication.” That is important because replications are of interest and not too often published. They have taken on even increased importance given the concerns in science that many studies produce findings that are not replicable. Another word to add as a subtitle might be: A review or A meta-analysis. These are important to convey that the article is not an individual investigation but an evaluation of a broad literature.

Occasionally authors are wont to use titles with simple questions, “Is depression really a detriment to health?” or “Is childhood bullying among boys a predictor of domestic violence in adulthood?” In general, it is advisable to avoid “yes, no” questions in the title. Science and findings are often nuanced and findings are likely to be both yes and no but under very different circumstances or for some subgroups of people but not for others. As an example, consider a hypothetical yes–no question for the title of a study as, “Is cigarette smoking bad for one’s health?” For anyone on the planet, the answer might be a resounding yes. Yet, the yes–no nature of the question makes this a poor choice of title because the answer is likely to depend on either how smoking is defined (e.g., how much smoking – a cigarette a year, a pack after each meal) and how health is defined (e.g., mental, physical, what diseases, disorders). Very familiar is how horrible smoking is for one’s physical health in so many domains (e.g., heart disease, cancer, chronic respiratory disease), but the question in the title can be answered both yes and no. Less familiar is the fact that cigarette smoking and exposure to cigarette smoke (among nonsmokers) reduce the risk for Parkinson’s disease and there are reasonable explanations for that based on brain chemistry and neurotransmitters (Reference Ma, Liu, Neumann and GaoMa et al., 2017; Reference Miller and DasMiller & Das, 2007). Clearly, the hypothetical title is plainly simplistic and not very helpful or informative because we can show many circumstances in which yes and no are correct answers. I am not arguing in favor of cigarette smoking (although I used to be a chain smoker until I switched to cigarettes). I am advising against titles of empirical articles that have a yes–no question given that most answers involve essays. Few phenomena allow the simplistic thinking the question can reflect.

2.2 Abstract

The Abstract is likely to be read by many more people than is the full article. The Abstract will be entered into various databases and be accessible through Internet and online searches. Many journals list the tables of contents for their issues and provide free access on the Web to abstracts of the articles but charge for the full article. Consequently, the Abstract is the only information that most readers will have about the study. For reviewers of the manuscript and readers of the journal article, the Abstract conveys what the author studied and found. Ambiguity, illogic, and fuzziness here are ominous. Thus, the Title and Abstract are sometimes the only impression or first impression one may have about the study.

Obviously, the purpose of the Abstract is to provide a relatively brief but comprehensive statement of goals, methods, findings, and conclusions of the study. Critical methodological descriptors pertain to the participants and their characteristics, experimental and control groups or conditions, design, and major findings. Often space is quite limited; indeed a word limit (e.g., 150–250 words maximum) may be placed on the abstract. It is useful to deploy the words to make substantive statements about the characteristics of the study and the findings, rather than to provide general and minimally informative comments. For example, vacuous statements (“Implications of the results were discussed” or “Future directions for research were suggested”) ought to be replaced with more specific comments of what one or two implications and research directions are (e.g., “The findings suggest that the family and peers might be mobilized to prevent drug abuse among adolescents and that cultural influences play a major role.”). Also, the more specific comments can convey the study’s relevance and interest value beyond what is suggested by the manuscript title or the opening comments of the Abstract. As a reader, I am not going to read very eagerly an article with the vacuous “implications” or “future directions” sentences, but if I am interested in the specific topics mentioned as implications (brain activity, the immune system, family, peers, culture), this article is a must for me to read. As authors, we often lament the word restrictions placed on us in the Abstract, but the first task is to make sure we are using the existing allotment with maximum information.

2.3 Introduction

The Introduction is designed to convey the overall rationale and objectives. The task of the author is to convey in a crisp and concise fashion why this study is needed and the current questions or deficiencies the study is designed to address. The section should not review the literature in a study-by-study fashion, but rather convey issues and evaluative comments that set the stage for the study. Placing the study in the context of what is and is not known (contextualization) and the essential next step in research in the field requires mastery of the pertinent literatures, apart from reasonable communication skills. Ironically, mastery of the literature is needed so the author knows precisely what to omit from the Introduction. A vast amount of material one has mastered and that is very interesting will need to be omitted because it does not set the stage or convey the precise context for this study.

Saying that the study is important (without systematically establishing the context) or noting that no one else has studied this phenomenon (measure or sample) usually are feeble attempts to short-circuit the contextualization of the study. In a manuscript I reviewed, the author mentioned four times in the Introduction (and three more in the Discussion) that this was the first time this study has been done. This was not amusing. Someone had not advised or helped the author very much and a very poor case was made for the study. Among the tasks of the Introduction is to lead the reader to the conclusion that the study is important and worthwhile. Telling the reader that the study is important is an argument from authority and that is not how science works. This might even strongly suggest that the author has not done his or her contextualization homework.

It may be relevant to consider limitations of previous work and how those limitations can be overcome. These statements build the critical transition from an existing literature to the present study and the rationale for design improvements or additions in relation to those studies. It is important to emphasize that “fixing limitations” of prior work is not necessarily a strong basis for publishing a study. The author must convey that the limitations of prior work are central to a key building block in theory or the knowledge base. Convey that because of that specific limitation, we really do not know what we thought we did or that there is a new ambiguity that is important but hidden in prior studies considering what was studied and by what means. Alternatively, the study may build along new dimensions to extend the theory and constructs to a broader range of domains of performance, samples, and settings. The rationale for the specific study must be very clearly established. Theory and previous research usually are the proper springboard to convey the importance of the current study.

In general, the Introduction will move from the very general to the specific. The very general refers to the opening of the Introduction that conveys the area, general topic, and significance of a problem. For example, in studies of diagnosis, assessment, treatment, or prevention of clinical dysfunction, the Introduction invariably includes a paragraph to orient the reader about the seriousness, prevalence or incidence, and economic and social costs of the disorder. Reviewers of the manuscript are likely to be specialists in the area of the study and hence know the context very well. Yet, many potential readers would profit from a statement that conveys the significance, interest, and value of the main focus of the study.

After the initial material, the Introduction moves to the issues that underlie this specific study. Here the context that frames the specific hypotheses of the study are provided and reflect theory and research that are the impetus for the investigation. There is an introduction syllogism, as it were, a logic that will lead the reader from previous theory and research to the present study with a direct path. Extended paragraphs that are background without close connections to the hypotheses of the study serve as a common weakness of manuscripts rejected for publication.

The Introduction does not usually permit us to convey all the information we wish to present. In fact, the limit is usually 2–5 manuscript pages. A reasonable use of this space is in brief paragraphs or implicit sections that describe the nature of the problem, the current status of the literature, the extension to theory and research this study is designed to provide, and how the methods to be used are warranted. The penultimate or final paragraph of the Introduction usually includes a statement of the purpose of the study and the specific hypotheses and predictions. By the time the reader reaches this paragraph or set of paragraphs, it should be very clear that these hypotheses make sense, are important, and address a critical issue or need in the knowledge base. In short, the Introduction must establish that the study addresses a central issue. To the extent that the author conveys a grasp of the issues in the area and can identify the lacunae that the study is designed to fill greatly improves the quality of the report and the chances of acceptance for journal publication. By the time the readers arrive at the purpose of the study or hypotheses paragraph, they should be nodding enthusiastically and saying to themselves, “This study is really needed, it should have been done years ago, I am so glad this is being done now.” As authors we often believe a description of the study is all that is needed. The identical study (description of what was done) can be viewed as a weak and just another study or strong, compelling, and sorely needed. All this can be decided by how the Introduction is cast.

2.4 Method

This section of the paper encompasses several points related to who was studied, why, and how. The section not only describes critical procedures, but also provides the rationale for methodological decisions. Subject selection, recruitment, screening, and other features ought to be covered in detail. Initially, the subjects or clients are described. Why was this sample included and how is this appropriate to the substantive area and question of interest? In some cases, the sample is obviously relevant because participants have the characteristic of interest (e.g., parents accused of child abuse, siblings of children with autism spectrum disorder) or are in a setting of interest (e.g., daycare center, wilderness camp). In other cases, samples are included merely because they are available. Such samples, referred to as samples of convenience, may include college students or a clinic population recruited for some other purpose than to test the hypotheses of this study. The rationale for the sample should be provided to convey why this sample provides a good – or if not good, a reasonable – test of the hypotheses and whether any special features may be relevant to the conclusions.

The design is likely to include two or more groups that are treated in a specific fashion. The precise purpose of each group and the procedures to which they are exposed should be clarified. Control groups should not merely be labeled as such with the idea that the name is informative. The author should convey precisely what the group(s) is designed to control. The author is advised to identify the critical methodological concerns and to convey how these are controlled in the design. Plausible threats to experimental validity that are uncontrolled deserve explicit comment to arrest the reasonable concerns of the reviewers (see Reference KazdinKazdin, 2017).

Several measures are usually included in the study. Why the constructs were selected for study should have been clarified in the Introduction. Then the specific measures and why they were selected to operationalize the constructs should be presented in the Method section. Information about the psychometric characteristics of the measures is often highlighted. This information relates directly to the credibility of the results. Apart from individual assessment devices, the rationale for including or omitting areas that might be regarded as crucial (e.g., multiple measures, informants, settings) deserves comment.

Occasionally, ambiguous statements may enter into descriptions of measures. For example, measures may be referred to as “reliable” or “valid” in previous research, as part of the rationale for their use. There are, of course, many different types of reliability and validity. It is important to identify those characteristics of the measure found in prior research that are relevant to the present research. For example, high internal consistency (reliability) in a prior study may not be a strong argument for use of the measure in a longitudinal design where the author cares more about test–retest reliability. Even previous data on test–retest reliability (e.g., over 2 weeks) may not provide a sound basis for repeated testing over annual intervals. The author ought to present information to convey the suitability of the measures for the study.

It often appears that reliability and validity of assessment are not routinely taught, at least if one looks at Method sections of articles in clinical psychology. These are important concepts because they can determine what is measured by a given instrument and how well. One sees more routinely that authors report Cronbach’s alpha for a measure and then move on. Alpha is one measure of reliability (internal consistency) and can be very useful to know. However, this has little to do with validity of the measure and by itself is not a justification for using a specific measure without much more explanation. Perhaps add a couple of sentences in this section to comment specifically on reliability and validity and what types have been supported in prior research. This is not merely to convince a reader but also ourselves on the wisdom of electing this measure. It is unreasonable to expect the measures to have the ideal reliability and validity data that the investigator would like to make a flawless case for use of these measures. Yet, make the case from what psychometric data there are. If data are not available, include some analyses in the study to suggest the measure(s) behave in ways that suggest pertinent forms of reliability or validity (Reference KazdinKazdin, 2017).

2.5 Results

It is important to convey why specific statistical tests were selected and how these serve the goals of the study. A useful exercise is for the investigator to read that paragraph about hypotheses and predictions from the Introduction and then immediately start reading the Results section, i.e., for the moment completely bypass the Methods. The results ought to speak directly to and flow from that narrative statement in the Introduction.

Analyses often are reported in a rote fashion in which, for example, the main effects are presented and then interactions for each measure. The author presents the analyses in very much the same way as the software output. Similarly, if several dependent measures are available, a set of analyses is automatically run (e.g., omnibus tests of multivariate analyses of variance followed by univariate analyses of variance for individual measures). The tests may not relate to the hypotheses, predictions, or expectations outlined at the beginning of the paper. It is important that the statistical tests be seen and presented as tools to answer questions or enlighten features of those questions and to convey this to the reader. The reader should not be able to legitimately ask, “Why was that statistical test done?” Knowledge of statistics is critical for selecting the analyses to address the hypotheses and conditions met by the data. Yet, as important in the presentation is to convey precisely why a given statistical test or procedure is suitable to test the hypotheses and then again what the results of that test reveal in relation to those hypotheses.

It is often useful to begin the Results by presenting basic descriptors of the data (e.g., means, standard deviations for each group or condition), so the reader has access to the numbers themselves. The main body of the Results is to test the hypotheses or to evaluate the predictions. Organization of the Results (subheadings) or brief statements of hypotheses before the analyses are often helpful to prompt the author to clarify how the statistical test relates to the substantive questions and to draw connections for the reader.

Several additional or ancillary analyses may be presented to elaborate the primary hypotheses. For example, one might be able to reduce the plausibility that certain biases may have accounted for group differences based on supplementary or ancillary data analyses. Ancillary analyses may be more exploratory and diffuse than tests of primary hypotheses. Manifold variables can be selected for these analyses (e.g., sex, race, height differences) that are not necessarily conceptually interesting in relation to the goals of the study. The author may wish to present data, data analyses, and findings that were unexpected, were not of initial interest, and were not the focus of the study. The rationale for these excursions and the limitations of interpretation are worth noting. From the standpoint of the reviewer and reader, the results should make clear what the main hypotheses were, how the analyses provide appropriate and pointed tests, and what conclusions can be reached as a result. As in other portions of the manuscript, how the author has reached a decision (what analysis) and why are very important.

2.6 Discussion

The Introduction began with a statement of the need for this study and issues or lacunae in theory or research the study was designed to address. The Discussion continues the storyline by noting what we know now and how the findings address or fulfill the points noted previously. With the present findings, what puzzle piece has been added to the knowledge base, what new questions or ambiguities were raised, what other substantive areas might be relevant for this line of research, and what new studies are needed? From the standpoint of contextualization, the new studies referred to here are not merely those that overcome methodological limitations of the present study, but rather focus on the substantive next steps for research. Also, this is not the place for vacuous suggestions such as, “This study needs to be replicated with people who are …”. If one is suggesting an extension of the study to different subjects, settings, or other dimensions, specify exactly why this specific extension would be of special interest.

The Discussion usually includes paragraphs to provide an overview of the major findings, integration or relation of these findings to theory and prior research, limitations and ambiguities and their implications for interpretation, and future directions. These are implicit rather than formally delineated sections and the author ought to consider the balance of attention to each topic. Usually, the Discussion is completed within 3–5 manuscript pages.

Description and interpretation of the findings can raise a tension between what the author wishes to say about the findings and their meaning versus what can be said in light of how the study was designed and evaluated. It is in the Discussion that one can see the interplay of the Introduction, Methods, and Results sections. For example, the author might draw conclusions that are not quite appropriate given the method and findings. The Discussion may convey flaws, problems or questionable methodological decisions within the design that were not previously evident. That is, the reader of the manuscript can now state that if these are the statements the author wishes to make, the present study (design, measures, or sample) is not well suited. The slight mismatch of interpretative statements in the Discussion and Methods is a common, albeit tacit basis for not considering a study as well conceived and executed. A slightly different study may be required to support the specific statements the author makes in the Discussion. It is important to be precise about what can and cannot be asserted in light of the design and findings.

It is usually to the author’s credit to examine potential limitations or sources of ambiguity of the study. A candid, non-defensive appraisal of the study is very helpful. Here too, contextualization may be helpful because limitations of a study also are related to the body of prior research, what other studies have and have not accomplished, and whether a finding is robust across different methods of investigation. Although it is to the author’s credit to acknowledge the limitations of the study, there are limits on the extent to which reviewers grant a pardon for true confessions. At some point, the flaw is sufficient to preclude publication, whether or not the author acknowledges it. For example, the authors of the study might note, “A significant limitation of the present study is the absence of a suitable control group. We are aware that this might limit the strength of the conclusions.” Awareness here does not strengthen the demonstration itself. A huge limitation in the study is sufficiently damaging to preclude drawing valid inferences.

In noting the limitations of the study, there is a useful structure for the presentation. First note the limitation. Then discuss to the extent reasonable that this limitation is not likely to influence the conclusion (if this is the case). If the role of the limitation cannot be diminished or dismissed by sound reasoning or related data, note that addressing this issue is a logical if not important next step for research. All studies have limitations by their very nature so reasoning about their likely and unlikely impact on the findings is invariably relevant.

At other points, acknowledging potential limitations conveys critical understanding of the issues and guides future work. For example, in explaining the findings, the author may note that although the dependent measures are valid, there are many specific facets of the construct of interest that are not covered. Thus, the results may not extend to different facets of the construct as measured in different ways. Here too it is useful to be specific and to note precisely why other constructs and their measure might show different results. In short, be specific as to why a limitation or point might really make a difference. This latter use of acknowledgment augments the contribution of the study and suggests concrete lines of research.

3. Questions to Guide Manuscript Preparation

The section-by-section discussion of the content of an article is designed to convey the flow or logic of the study and the interplay of description, explanation, and contextualization. The study ought to have a thematic line throughout and all sections ought to reflect that in a logical way. The thematic line consists of the substantive issues guiding the hypotheses and decisions of the investigator (e.g., about procedures and analyses) that are used to elaborate these hypotheses. I mentioned that one way to check this is to read sections together like the Introduction and Results (by skipping the Method section). These sections ought to follow a similar flow. Analyses should be connected logically but with sentences about what is being tested and what was found in relation to the ideas or hypotheses presented in the Introduction. Skipping the Method section for this reading helps one to consider the flow. Similarly, one could push this further and read the Introduction and then Discussion – are they connected? The opening of the Discussion can address issues that were written at the end of the Introduction, i.e., the purpose of this study. This is not a repeat of the purpose but a summary of the main results that addressed those purposes and goals. All these little tools are designed to help us as authors convey a thematic and logical flow that the reader can easily see.

A more concrete and hence perhaps more helpful way of aiding preparation of the manuscript is to consider our task as authors as that of answering many questions. There are questions for the authors to ask themselves or, on the other hand, questions reviewers and consumers of the research are likely to ask as they read the manuscript. These questions ought to be addressed suitably within the manuscript. Table 13.1 presents questions according to the different sections of a manuscript. The questions emphasize the descriptive information, as well as the rationale for procedures, decisions, and practices in the design and execution. The set of questions is useful as a way of checking to see that many important facets of the study have not been overlooked. As a cautionary note, the questions alert one to the parts rather than the whole; the manuscript in its entirety or as a whole is evaluated to see how the substantive question and methodology interrelate and how decisions regarding subject selection, control conditions, measures, and data analyses relate in a coherent fashion to the guiding question.

Table 13.1 Major questions to guide journal article preparation

Abstract

  • What are the main purposes of the study?

  • Who was studied (sample, sample size, special characteristics)?

  • How were participants selected and assigned to conditions?

  • To what conditions, if any, were participants exposed?

  • What type of design was used?

  • What are the main findings and conclusions?

  • What are one or two specific implications or future directions of the study?

Introduction

  • What is the background and context for the study?

  • What in current theory or research makes this study useful, important, or of interest?

  • What is different or special about the study in focus, methods, or design to address a need in the area?

  • Is the rationale clear regarding the constructs (independent and dependent variables) to be assessed?

  • What specifically are the purposes, predictions, or hypotheses?

  • Are there ancillary or exploratory goals that can be distinguished as well?

Method

  • Participants

  • Who are the participants and how many of them are there in this study?

  • Why was this sample selected in light of the research goals?

  • How was this sample obtained, recruited, and selected?

  • What are the subject and demographic characteristics of the sample (e.g., sex, age, ethnicity, race, socioeconomic status)?

  • What, if any, inclusion and exclusion criteria were invoked, i.e., selection rules to obtain participants?

  • How many of those subjects eligible or recruited actually were selected and participated in the study?

  • In light of statistical power considerations, how was the sample size determined?

  • Was informed consent solicited? How and from whom (e.g., child and parent), if special populations were used?

  • If non-human animals are the participants, what protections were in place to ensure their humane care and adherence to ethical guidelines for their protection?

  • Are they any professional, personal, or business interests or connections, financial or otherwise (e.g., service on boards) that might be or be perceived as a conflict of interest in relation to the focus of the study or direction of the findings?

Design

  • What is the design (e.g., group, true-experiment) and how does the design relate to the goals?

  • How were participants assigned to groups or conditions?

  • How many groups were included in the design?

  • How are the groups similar and different?

  • If groups are “control” groups, for what is the group intended to control?

  • Why are these groups critical to address the questions of interest?

Procedures

  • Where was the study conducted (setting)?

  • What measures, materials, equipment, or apparatus were used?

  • What is the chronological sequence of events to which participants were exposed?

  • What intervals elapsed between different aspects of the study (e.g., assessment, exposure to the manipulation, follow-up)?

  • If assessments involved novel measures created for this study, what data can be brought to bear regarding pertinent types of reliability and validity?

  • What checks were made to ensure that the conditions were carried out as intended?

  • What other information does one need to know to understand how participants were treated and what conditions were provided to facilitate replication of this study?

Results

  • What are the primary measures and data upon which the hypotheses or predictions depend?

  • What analyses are to be used and how specifically do these address the original hypotheses and purposes?

  • Are the assumptions of the statistical analyses met?

  • If multiple tests are used, what means are provided to control error rates (increased likelihood of finding significant differences in light of using many tests)?

  • If more than one group is delineated (e.g., through experimental manipulation or subject selection), are they similar on variables that might otherwise explain the results (e.g., diagnosis, age)?

  • Are data missing due to incomplete measures (not filled out completely by the participants) or due to loss of subjects? If so, how are these handled in the data analyses?

  • Are there ancillary analyses that might further inform the primary analyses or exploratory analyses that might stimulate further work?

Discussion

  • What are the major findings of the study?

  • Specifically, how do these findings add to research and support, refute, or inform current theory?

  • What alternative interpretations, theoretical or methodological, can be placed on the data?

  • What limitations or qualifiers are necessary, given methodology and design issues?

  • What research follows from the study to move the field forward?

  • Specifically, what ought to be done next (e.g., next study, career change of the author)?

More generally

  • What were the sources of support (e.g., grants, contracts) for this specific study?

  • If there is any real or potentially perceived conflict of interest, what might that be?

  • Are you or any coauthors or a funding agency likely to profit from the findings or materials (e.g., drugs, equipment) that are central to the study?

Note: These questions capture many of the domains that ought to be included, but they do not exhaust information that a given topic, type of research, or journal might require. Even so, the questions convey the scope of the challenge in preparing a manuscript for publication.

4. Guidelines for Research

4.1 Impetus for Reporting Guidelines

There have been a long-series of guidelines on how to conduct and report research and these are directly related to preparation of a study for publication. The history includes special emphasis on ethical treatment of participants. Regulations followed in response to atrocities of the Nazi regime during World War II and the resulting development of the Nuremburg Code (1940). Since then, many other codes have developed (e.g., Declaration of Helsinki of the World Medical Association, Belmont Report), beyond the scope of the present discussion (see Reference KazdinKazdin, 2017). Protection of participant rights remain as important as ever and even of greater concern considering new opportunities to obtain and combine data sources (“big data”), often with information that is public in some way (e.g., medical records, social media, tracking locations and purchases). Individuals may not be aware of the collection and use of the information. Even when participants may be anonymous, in fact often groups (e.g., by ethnic, culture or setting) can be readily singled out and identified in ways that can reflect quite negatively on them (e.g., Reference Metcalf and CrawfordMetcalf & Crawford, 2016; Reference ZimmerZimmer, 2010).

The need for guidelines for reporting research has emerged from multiple additional concerns. First, collaborative research currently is more the rule rather than the exception in science. Collaborations often involve scores of authors, from multiple disciplines, and from many different countries. There is interest across nations in reaching common standards in relation to the openness of research, access to information, the merit-review process, and ethical issues (e.g., Reference SureshSuresh, 2011). That has provided a critical context for providing guidelines for research and the reporting of research that span multiple disciplines, countries, and journals.

Second, lapses in what is reported in research have been well documented. For example, information often is omitted such as exactly who the participants are (e.g., subject and demographic variables) and how they were recruited, who administered treatment or experimental procedures, the extent of their training, whether the integrity or execution of treatment was assessed, fundamental characteristics of the data evaluation, and more, as reflected in citations noted previously.

Third, selective reporting of results and data analyses has been raised as a critical issue that introduces biases in individual studies and entire literatures. The selective reporting of results of some data analyses or some of the dependent measures can increase the likelihood of more chance findings in the literature (Reference Simmons, Nelson and SimonsohnSimmons et al., 2011). For example, in identifying evidence-based treatment, often there is a clear bias in how authors report the data by not presenting the full range of measures, some of which would not support conclusions about the impact of treatment (Reference De Los Reyes and KazdinDe Los Reyes & Kazdin, 2008). Guidelines are intended to foster consistency and clarity in how the study will be reported to minimize the biases in reporting that emerge.

Fourth, more flagrant than “mere” omission of information and selective reporting has been fraud and fabrication of data in science. Fraud is not new in science. However, both the visibility of fraud to the public, including the circulation through social media, and direct and disastrous implications from fraudulent studies have been more evident than ever before (see Reference Levelt and CommitteesLevelt, Noort, & Drenth Committees, 2012; Reference Watanabe and AokiWatanabe & Aoki, 2014).Footnote 2

Finally, there has been renewed concern about the replicability of research. Replication as a general tenet, if not practice, has always been the backbone of science. Given many of the points I have already mentioned (publication biases, statistical analyses, and selective reporting), various authors have reached the dramatic conclusion, occasionally supplemented with mathematical proofs and simulations, that that many and even most published research findings are not correct, i.e., are false (see Reference FrancisFrancis, 2012; Reference IoannidisIoannidis, 2005; Reference Moonesinghe, Khoury and JanssensMoonesinghe et al., 2007). Several calls for increased replication have been made. Psychology has taken the lead in calling for and supporting replications and underscoring the importance of transparency of procedures (Center for Open Science, https://cos.io/). Central to replication, of course, is making the procedures explicit and the materials and results available. Guidelines for conducting research to increase the likelihood that a study can be replicated are obviously important. Many journals, national and international, require providing information about a study and the data so they are freely available to others to facilitate re-evaluation of the data and replication of the entire study.

Overall, science has come under increased scrutiny both from within the sciences, government, and the public at large. Even though the assorted problems I highlighted are seemingly infrequent, the circulation of information (e.g., Web, news and social media) is more extensive than ever before and retractions (when authors and journals make some effort to “take back” and renounce what was published) are more visible and available as well. And news media more routinely comment on scientific findings and reflect skepticism about replication and replicability of effects (e.g., Reference LehrerLehrer, 2010). The points I have raised have served in part as the impetus for improving research, especially focusing on transparency and accountability of investigators. Guidelines have been helpful in fostering greater consistency in reporting of the research and in the process sensitizing researchers of what to attend to in advance of a study.

4.2 Sample Guidelines Briefly Noted

Several organizations and groups have developed standards for reporting research and in the process convey the need to address many of the issues I have highlighted previously. The scope of guidelines that are available is enormous. An international umbrella organization that collects, oversees, and promotes the use of research guidelines is the Equator Network. The network maintains a comprehensive database of guidelines, numbering over 400, as of this writing (www.equator-network.org/reporting-guidelines/). The network is nicely organized by type of paper (guidelines for empirical studies, literature reviews, meta-analyses) and by different methodologies (e.g., randomized trials, observational studies), and so on. With so many guidelines, with enormous overlap in what they cover, one can see this is not a minor movement to improve research.

Examples of such standards are the:

Most of the guidelines include some combination of checklists, flow charts, and narrative explanations of what specific items are to be included in a report and what the information is designed to accomplish. I mention two briefly.

First, the CONSORT standards, mentioned above, are arguably the most familiar set of guidelines. They have been adopted by hundreds of professional journals encompassing many disciplines and countries (see www.consort-statement.org/). The CONSORT guidelines have been devised primarily for clinical trials in medical research but have extended well beyond that and are routinely used in clinical trials of psychosocial interventions. As noted in the most recent version, clinical trials have a history of omitting significant information such as description of who was included in the study; sample size calculation (e.g., why a specific size was included in relation to statistical power issues), descriptions of procedures; and presentation of procedures (e.g., randomization) that were not really invoked; and so on with other lapses (Reference Moher, Schulz and AltmanMoher et al., 2001). Beginning in the early 1990s, efforts began to make recommendations for reporting of studies and from that the CONSORT guidelines emerged.

The guidelines consist of a checklist of essential items that ought to be included in any randomized controlled trial of treatment. The checklist displays what is needed, but along with the checklist is a detailed explanation of the items and their rationale for inclusion (Reference Moher, Schulz and AltmanMoher et al., 2001). In addition, the website provides educational material and a database of materials related to reporting of randomized controlled trials (e.g., examples from real trials). In preparation of journal articles, the CONSORT criteria include a list of what to cover and how. This is more concrete than my general statements of ensuring there is a logical flow and underlying theme to the journal article. Yet, the details are basic and critical and hence these guidelines are valuable and widely adopted by many journals.

Second, ClinicalTrials.Gov (https://clinicaltrials.gov/) provides another model to guide research. This consists of preregistration of a study that requires authors to convey their plan for conducting research and analyzing the data. Preregistration allows for the range of participants in research (investigators, peer reviewers, journal editors, funding agencies, policy makers, the public at large) to determine whether the research, when completed, has deviated from the pre-registered plan. Pre-registration of research is now common across many funding agencies and journals (Reference Nosek, Ebersole, DeHaven and MellorNosek et al., 2018). ClinicalTrials.Gov is a large database that includes privately and publicly funded studies of investigations throughout the world. Indeed, this is the largest clinical trials database and as of this writing over 327,000 studies are registered and include studies from all 50 states in the United States and 209 countries (as of January 2020). When clinical trials are comparing interventions or an intervention against a control group, funding agencies (e.g., National Institutes of Health), organizations (e.g., World Health Organization), and a consortium of journal editors (the International Committee of Medical Journal Editors) require investigators to register their clinical trials in advance of the study.

In registration of one’s study, information covers diverse facets of the project. Indeed, there is a multi-page template that includes identification of the investigators of the study, the design, what the interventions are, what will be the outcome criteria (e.g., primary and secondary), the number of anticipated subjects, criteria for inclusion of the subjects, status of procedures to protect clients, and much more. Merely mentioning some of the domains that are included in the register does not underscore their significance in the conduct and reporting of the study. Consider three examples to convey the point.

First, the guidelines require specification of the outcome criteria and which outcomes or measure will be primary and secondary. This is a pre-commitment of the investigator to be clear about the outcome. This does not mean that investigators cannot look at all outcome measures or derive new ones based on interesting findings, pre-specified or not, as the study is completed. However, pre-specification can reduce the tendency in written reports to underscore, emphasize, and consider as primary, those measures that “come out,” i.e., support the hypotheses.

Second, in many studies there are multiple investigators whose roles vary in the design, execution, analysis and other facets of the study and these investigators are likely to be listed as authors. Specification of the investigators and their roles clarifies accountability for the final manuscript. Also, this requires that people in fact have a role in the study before being included. All the expected dynamics of human interactions (e.g., who does and does not get to be an author, where they are placed in the list of authors) and human emotions (e.g., indignation, disappointment, rage, helplessness) surround authorship. War stories here could fill volumes. The guidelines can help a little. At the end of a study, there is accountability of who is in charge, who had what role, and who was involved. If there is fraud or faked data or questionable practices or a manuscript retraction (once there is a question about practices or a finding), the team involved in the study and their roles can be delineated. This can enhance the integrity of the research progress by making clear that one is accountable for the study and its conduct. Third, the guidelines specify whether, how, and where the data will be stored and whether other materials critical to the study will be available. Occasionally journals or funding agencies require that the data are deposited and made available. This practice fostered by the guidelines increases the transparency of the research but also helps replication efforts.

A few comments in passing. To begin, pre-registration does not fix the research so that no further changes can be made. In fact, one can update changes that occur in the course of a study. “Pre”-registrations can be updated after participant enrollment or even after data collection has begun to document any changes that occur in the course of a study (Reference DeHavenDeHaven, 2017; Reference Nosek, Ebersole, DeHaven and MellorNosek et al., 2018). All that is required is to make sure the changes are clear, transparent, and explained. The registration still thwarts post-hoc decision making based on how the data come out or switching some measures and ignoring others, some of the sins of research. An additional point, the many guidelines are designed to improve reporting of research. However, so many facets need to be considered ahead of time in these guidelines that they necessarily influence and guide the design of a study. This article is about preparing a manuscript for journal publication. Consulting and following many widely adopted guidelines underscores the point that key issues about the publication of a journal article emerge before the first subject is run in the study.

5. Selecting a Journal

Preparation of the manuscript logically occurs before selecting a journal and submitting to the journal for publication. However, investigators occasionally have the journal or a couple of journals in mind before the manuscript is prepared. Journals have different emphases and research with specific foci (e.g., theory, application), samples (e.g., non-human animals, college students, community samples), settings (laboratory, field), and research designs (cross-sectional, longitudinal, experimental, observational). Consequently, it is not odd for the investigator to plan/hope that a study when completed will be appropriate for a journal he or she targeted well before preparing the manuscript for publication. In my own case, I prefer to see the final or almost final write up to consider what journals might be reasonable outlets for the article. I mention selecting a journal here on the assumption that this logically follows in the sequence of completing a study, preparing the write up, and submitting the article for publication. Selecting a journal is part of submitting the article.

Thousands of journals are available in the behavioral and social sciences and the resources and potential relevance to your study are easily obtained from the Web (Reference GuntherGunther, 2011; Thomson Reuters, 2011; Reference ThursbyThursby, 2011). These sources can be searched by topic and keywords in relation to how you view your study (e.g., clinical psychology, candidate for Nobel prize). It is beneficial to skip the search among the thousands of journals and begin the search more narrowly. There are many professional organizations within psychology that have their own publications. The two major professional organizations whose journal programs are widely recognized and emulated are American Psychological Association (APA, 2020a) and the Association for Psychological Science (2020).

Each source I have noted here provides information about the editorial policy, content area or domain, type of article (e.g., investigations, literature reviews, case studies), guidelines for manuscript preparation, and access to tables of contents of current and past issues. I have emphasized journals in the English language. Psychology is an active discipline internationally and psychological associations in many countries and regions (e.g., European Union, Scandinavia) have many excellent journals as well.

Many criteria are invoked to select a journal to which one will submit a manuscript, including the relevance of the journal in relation to the topic, the prestige value of the journal in an implicit hierarchy of journals in the field, the likelihood of acceptance, the breadth and number of readers or subscribers, and the discipline and audience one wishes to reach (e.g., psychology, psychiatry, medicine, social work, health, education). As for the prestige value, clearly some journals are regarded as more selective than others. For example, some of the APA journals are premier journal outlets in their respective areas (e.g., Journal of Consulting and Clinical Psychology, Journal of Personality and Social Psychology). Yet, journals from other organizations, journals not sponsored by an organization, and journals from other professions or disciplines can be as or more highly regarded. Indeed, in some areas (e.g., behavioral neuroscience), some of the most discriminating and selective publication outlets are not psychology journals (Science, Nature Neuroscience). One can identify the best outlets by familiarity with the literature (e.g., where do the best studies seem to be published) and by chatting with colleagues.

Word of mouth and reputation of a journal often are well recognized and their status within professional organizations is known. There has been an enduring interest in having more objective measures and they are available. The impact of a journal is primary among these measures (Web of Science, 2020) and includes the extent to which articles in a journal are cited by others. Journals with articles that are heavily cited are those with much higher impact. Information is available for journals in virtually all areas of science. Within the social sciences alone over 3400 journals are covered. There are reasons not to be wedded to journal impact.Footnote 3 The impact of one’s work is very important, and it appears that is not really related to the journal impact measure.

Some journals are not very selective and, indeed, must hustle (e.g., invite, accept many) articles so they can fill their pages. Indeed, it is not difficult at all to get one’s work published in the genre referred to as predatory journals (e.g., Reference BrainardBrainard, 2020). These are journals that send countless emails to professionals seeking their manuscripts and with little and sometimes no evaluation of merit. The journals are primarily business ventures and charge high author fees. The journal landscape is intricate because some journals with a peer-review process offer the option of open access (article available to anyone on line) if the author pays a publication fee.

Be a little wary of journals in psychology that charge authors for publishing their papers. For these journals, when one’s paper is accepted, the author is charged based on how many journal pages the article will require. These outlets do not necessarily take all submissions, but they often take most. These journals tend not to be as carefully peer-reviewed and hence publications in such journals are commensurately much less well regarded. Within psychology, career advice is to focus on peer-reviewed and well-regarded journals, leaving aside other issues (e.g., who publishes the journal, whether there are charges). Knowledge of the area of research, journal citation impact, and contact with one’s colleagues can readily identify the ideal outlets for one’s research. Early in my career, I asked a senior colleague about a journal and he gave me a sharp NEVER publish there. Decades later I can see that was sound advice. If in doubt, seek advice. If you have no doubts but you are early in your career, perhaps also seek advice.

The audience one wishes to reach may be a critical and indeed primary consideration in selecting a journal. Who might be interested in this study (beyond blood relatives)? One way to answer this is to consider the Reference section of one’s article. Are one or two journals emphasized in the Reference section of the manuscript? If so, one of these journals might be the most appropriate outlet. Citation of the journal on multiple occasions indicates that the journal publishes work on the topic and readers who are likely to be interested in the topic are also likely to see the article. Also relevant, journals vary markedly in their readership and subscription base. Some journals have relatively few subscribers (e.g., 200–600 up to several thousand) or are omitted from easily accessed databases. The visibility of one’s study and the chance that others will see it are influenced by these considerations. Fortunately, most professional journals have their abstracts included in databases that can be accessed from the Web. This makes even the most obscure study accessible.

Most journals are in print (hard copy) and electronic form, but many are only Web-based and are sometimes referred to as electronic journals or e-journals. This is not the place to discuss that topic except to note often publication on the Web is much faster (less delay in review of the manuscript and acceptance of the manuscript) than is publication in a printed journal. There are still dynamic changes in how journals will be published and disseminated and print versions may be on borrowed time. The central issue for one’s career is the extent to which the publication outlet is well regarded by one’s peers and the care with which manuscripts are reviewed before they are accepted and published. Electronic versus printed journal format is not as critical as the quality of the publication. If publication in the journal requires little or no peer review, if most manuscripts are accepted, and if manuscripts are accepted largely as they are (without revision), quality of the research and the value of the publication to one’s career may be commensurately reduced.

6. Manuscript Submission and Review

6.1 Overview of the Journal Review Process

Alas, through careful deliberation and 30 minutes with your coauthor at a Ouija board, you select a journal and are ready to submit your manuscript for publication. Before you do, consult the Instructions to Authors written in the journal to make sure you submit the manuscript correctly. Usually manuscripts are submitted through a journal portal, i.e., electronically, in which the manuscript file and a letter of submission are uploaded to the journal website. In some cases, you may be required to include sentences or paragraphs in the letter you submit that say this study is not being considered elsewhere in another journal, has not been published before, has met ethical guidelines specified by university or institutional policy and various laws, and that you will give the copyright to the publisher if the manuscript is accepted. Processing of the manuscript could be delayed if your letter does not meet the guidelines provided in the journal.

Once the manuscript is submitted, the journal editor usually sends the electronic file to two or more reviewers who are selected because of their knowledge and special expertise in the area of the study or because of familiarity with selected features of the study (e.g., novel methods of data analyses). Reviewers may be selected from the names of authors whose articles you included in your Introduction. Some reviewers are consulting editors who review often for the journal and presumably have a perspective of the type and quality of papers the journal typically publishes; other reviewers are ad-hoc reviewers and are selected less regularly than consulting editors. Reviewers are asked to evaluate the manuscript critically and to examine whether or the extent to which:

  • The question(s) is important for the field;

  • The design and methodology are appropriate to the question;

  • The results are suitably analyzed;

  • The interpretations follow from the design and findings; and

  • The knowledge yield contributes in an incremental way to what is known already.

(You may note that these bulleted points encompass the explanation and contextualization features I noted in relation to manuscript preparation. Each point is one that can be readily addressed by the author in preparing the manuscript.) Typically, reviewers are asked to give a summary recommendation (e.g., reject or accept the manuscript). All recommendations to an editor are advisory and not binding in any way. At the same time, the editor sought experts and usually follows their recommendations. Yet reviewers too must make the case for their comments.

Once the paper is reviewed, the editor evaluates the manuscript and the comments of the reviewers. In some cases, the editor may provide his or her own independent review of the paper; in other cases he or she may not review the paper at all but defer to the comments and recommendations of the reviewers. The editor writes the author and notes the editorial decision. Usually, one of three decisions is reached: the manuscript is accepted pending a number of revisions that address points of concern in the reviewers’ comments; the manuscript is rejected and will not be considered further by the journal; or the manuscript is rejected but the author is invited to resubmit an extensively revised version of the paper for reconsideration.

The accept decision usually means that the overall study was judged to provide important information and was well done. However, reviewers and the editor may have identified several points for further clarification and analysis. The author is asked to revise the paper to address these points. The revised paper would be accepted for publication.

The reject decision means that the reviewers and/or editor considered the paper to include flaws in conception, design, or execution or that the research problem, focus, and question did not address a very important issue. For journals with high rejection rates, papers are usually not rejected because they are flagrantly flawed in design. Rather, the importance of the study, the suitability of the methods for the questions, and specific methodological and design decisions conspire to serve as the basis for the decision.

The reject–resubmit decision may be used if several issues emerged that raise questions about the research and the design. In a sense, the study may be viewed as basically sound and important but many significant questions preclude definitive evaluation. The author may be invited to prepare an extensively revised version that includes further procedural details, additional data analyses, and clarification of many decision points pivotal to the findings and conclusions. The revised manuscript may be re-entered into the review process and evaluated again as if it were new. On other occasions, the manuscript may be resent to reviewers familiar with the prior version. Less often the editor may make an executive decision and accept or reject the manuscript without outside input.

Of the three letters, clearly a rejection letter is the most commonly received. Authors, and perhaps new authors in particular, are insufficiently prepared for this feature of the journal publication business.Footnote 4 Journals often publish their rejection rates, i.e., proportion of submitted manuscripts that are rejected, and this figure can be quite high (e.g., 70–90 percent). Often the prestige value of the journal is in part based on the high rejection rate. Yet, the rate is ambiguous at best because of self-screening among potential authors. For example, for very prestigious publication outlets (e.g., Psychological Review, Science) where psychological papers are published, the rejection rates cannot consider the fact that most authors are not likely to even try that outlet if they have a contribution that falls within the topic and format domain. Rejection rates across journals are not directly comparable. Even so, the rates give the would-be author the approximate odds if one enters the fray.

Although beyond our purpose, the review process deserves passing comment. The entire process of manuscript submission, review, and publication has been heavily lamented, debated, and criticized. The peer-review process has a long history as an effort of quality control over the content and standards of what is published (Reference SpierSpier, 2002). The topic is central to science broadly and continues to be assessed, commented on, and evaluated with efforts to alter or improve the processes (e.g., Reference Elson, Huff and UtzElson et al., 2020; Reference Kirman, Simon and HaysKirman et al., 2019). The alternatives to peer review (e.g., no review, judgment by one person such as the editor) have their own liabilities. Many journals invoke procedures where the identity of the authors and the reviewers is masked, i.e., names are not included on the manuscript sent to reviewers or the reviews sent to authors. The goal is to try to limit some of the human factors that can operate about responses to a person, name, or other facet and to allow reviewers to be candid in their evaluations without worrying about facing the colleague who will never speak to them again. The peer-review system is far from perfect. The imperfections and biases of peer review, the lack of agreement between reviewers of a given paper, the influence of variables (e.g., prestige value of the author’s institution, number of citations of one’s prior work within the manuscript) on decisions of reviewers, and the control that reviewers and editors exert over authors have been endlessly vigorously discussed (e.g., Reference Bailar and PattersonBailar & Patterson, 1985; Reference Benos, Bashari, Chaves, Gaggar, Kapoor, LaFrance, Mans, Mayhew, McGowan, Polter, Qadri, Sarfare, Schultz, Splittgerber, Stephenson, Tower, Walton and ZotovBenos et al., 2007; Reference CicchettiCicchetti, 1991; Reference SmithSmith, 2006; Reference Stahel and MooreStahel & Moore, 2014).

Understanding the review process can be aided by underscoring the one salient characteristic that authors, reviewers, and editors share, to wit, they are all human. This means that they (we) vary widely in skills, expertise, perspectives, sensitivities, motives, and abilities to communicate. Science is an enterprise of people and hence cannot be divorced from subjectivity and judgment. In noting subjectivity in the manuscript review and evaluation process, there is a false implication of arbitrariness and fiat. Quality research often rises to the top and opinions of quality over time are not idiosyncratic. Think of the peer-review process as the home-plate umpire in a baseball game. Any given call (e.g., strike) may be incorrect, arguable, and misguided. And any given pitcher or batter suffers unfairly as a result of that call. As reviewers (the umpires) make the call on your manuscript (rejection, you strike out), you too may have that occasional bad call. But over time, it is unlikely that all manuscripts an author submits receive a misguided call. Pitchers and batters earn their reputations by seeing how they perform over time, across many umpires, and many games. One looks for patterns to emerge, and this can be seen in the publication record of an active researcher.

6.2 You Receive the Reviews

Alas, the editorial process is completed (typically within three months after manuscript submission) and the reviews are in. You receive an email (or in olden days a printed letter) from the editor noting whether the paper is accepted for publication and if not whether it might be if suitably revised. It is possible that the letter will say the manuscript is accepted as is (no further changes) and praise you for your brilliance. The letter may comment further that the reviewers were awed by how the study was executed and how well the manuscript was written. If this occurs, it is the middle of the night and you are dreaming. Remain in this wonderfully pleasant state as long as you can. When you awake, your spouse, partner, or significant (p < .05) other reads the email and you read one of the three decisions noted previously.

If the manuscript is accepted, usually some changes are needed. These do not raise problems. More often than not, the manuscript is rejected. There are individual differences in how one reacts to this decision. Typically, one feels at least one of these: miffed, misunderstood, frustrated, or angry at the reviewers. Usually one has only the email comments and has limited avenues (e.g., scrutiny of the phrasing and language) for trying to identify who could have possibly rejected the manuscript. If a hard (printed) version of the reviews was sent to you, one can scrutinize the font style, key words, possible DNA remnants of the reviewers’ comments sheets, and molecules on the pages that might reveal pollutants associated with a specific city in the country. (I myself would never stoop to such behaviors but I have a “friend” who, over the years, was able to identify two not-so-friendly reviewers who unwittingly left clues that I – I mean my friend – was able to decipher.) To handle a rejection verdict, some authors select one of the very effective psychotherapies or medications for depression; others use coping strategies (e.g., anger management training, stress inoculation therapy) or complementary and integrative medicines (e.g., acupuncture, mineral baths, vegan enemas). (I myself use all these routinely with their order balanced in a Hyper-Graeco-Latin Square Design.)

The task is to publish one’s work. Consequently, it is useful and important to take from the reviews all one can to revise the manuscript. Maladaptive cognitions can harm the process. For example, when reading a review, the author might say, the reviewer misunderstood what I did or did not read this or that critical part. These claims may be true, but the onus is always on the author to make the study, its rationale and procedures, patently clear. A misunderstanding by a reviewer is likely to serve as a preview of the reactions of many other readers of the article. Indeed, most readers may not read with the care and scrutiny of the reviewers. If the author feels a rejected manuscript can be revised to address the key concerns, by all means write to the editor and explain this in detail and without righteous indignation and affect.

Authors often are frustrated at the reactions of reviewers. In reading the reactions of reviewers, the authors usually recognize and acknowledge the value of providing more details (e.g., further information about the participants or procedures). This is the descriptive facet of manuscript preparation I discussed previously. However, when the requests pertain to explanation and contextualization, authors are more likely to be baffled or defensive. This reaction may be reasonable because much less attention is given to these facets in graduate training and explanation and contextualization are much less straightforward. Also, reviewers’ comments and editorial decision letters may not be explicit about the need for explanation and contextualization. For example, some of the more general reactions of reviewers are often reflected in comments such as: “Nothing in the manuscript is new,” “I fail to see the importance of the study,” or “This study has already been done in a much better way by others.”Footnote 5 In fact, the characterizations may be true. Authors (e.g., me) often feel like they are victims of reviewers who wore sleep masks when they read the manuscript, did not grasp key points, and have had little exposure to, let alone mastery of, the pertinent literature. Occasionally two or more of these are true.

As often as not, it is the reviewers who might more appropriately give the victim speech. The comments I noted are great signs that the author has not made the connections among the extant literature and this study and integrated the substantive, methodological, and data-analytic features in a cohesive and thematic way. Reviewers’ comments and less than extravagant praise often reflect the extent to which the author has failed to contextualize the study to mitigate these reactions. The lesson for preparing and evaluating research reports is clear. Describing a study does not establish its contribution to the field, no matter how strongly the author feels that the study is a first.

Let us assume that the manuscript was rejected with an invitation to resubmit. As a rule, I try to incorporate as many of the reviewers’ and editor’s recommendations as possible. My view is that the reviewer may be idiosyncratic, but more likely represents a constituency that might read the article. If I can address several or all issues, clarify procedures that I thought were already perfectly clear, and elaborate a rationale or two, it is advisable to do so. Free advice from reviewers can and ought to be used to one’s advantage.

There are likely to be aspects of the reviews one cannot address. Perhaps reviewers provide conflicting recommendations, or a manuscript page limit precludes addressing or elaborating a specific point. Even more importantly, perhaps as an author one strongly disagrees with the point. Mention these in the letter to the editor that accompanies the revised manuscript. Explain what revisions were or were not made and why. If there are large revisions that alter the text (few sentences), methods or data analyses, help the editor by noting where the change can be found in the manuscript and even submit an extra copy of the manuscript in which the changes are tracked in some editing/word-processing system.

The investigator may receive a rejection letter and decide simply to submit the manuscript as is to another journal. I believe this is generally unwise. If there are fairly detailed reviews, it is to the author’s advantage to incorporate key and often not-so-key points, even if the manuscript is to go to another journal. I have often seen the same manuscript (not mine) rejected from two different journals in which there were no changes after the rejection from the first journal. The authors could have greatly improved the likelihood of publication in the second journal but were a bit stubborn about making any revisions. Even if the manuscript were to be accepted as is in the second journal, it is still likely the author missed an opportunity to make improvements after the first set of reviews was provided. In general, try to take all the recommendations and criticisms from the reviews and convert them to facets that can improve the manuscript. Obstacles to this process may stem from our natural defensive reactions as authors or a negativity bias and the occasional brutish way in which reviewers convey cogent points. (I remember being highly offended the first two or three times reviewers noted such comments, “the author [me] would not recognize a hypothesis if it fell on his lap” and “the design of this study raises very important issues, such as whether it is too late for the author [me] to consider a career change.” I have come to refer to all this as the pier-review process to underscore how often reviewers have made me want to jump off one.)

There is an additional reason to encourage taking advantage of the review process and trying to improve a manuscript we might think is perfect. For those researchers who remain in academia, one’s published studies occasionally are read as part of a promotion process. As an author we might feel relieved that a study or two was published and view that automatically as things are great. In some ways it does, but as this study is read later we still want to be sure the case was made in a compelling fashion and reviewer suggestions might help. My view is to incorporate as many recommendations, changes, and comments as possible. I begin with the view that reviewers are experts and their recommendations, concerns, and misunderstandings are facets of the manuscript it behooves me to address.

It is worthwhile and highly rewarding to publish one’s research. The process takes time and persistence. Also, contact with others through the review process can greatly improve one’s work. In my own case, reading the reviews occasionally has stimulated the next studies that I carried out. In one case, I befriended a person who had been a reviewer of my work earlier in my career. Over time and from following his work, it was very clear that he was behind an influential review although his identity had been masked. Years later over dinner, I mentioned his review in a distant past, the study it generated, and the very interesting results and, of course, expressed my gratitude. His suggestion actually led to a few studies. (His review of my manuscript was not entirely positive, which probably is the main reason I hid in the bathroom of the restaurant until he paid the check for dinner.) The lesson is more than getting one’s manuscript published. Reviews can be very educational and it is useful to let the comments sit for a while until the rage over rejection subsides.

The journal review process is not the only way to obtain input on one’s manuscript. Once in a while, I send a penultimate draft of a manuscript to experts in the field whom I do not know. I convey in a letter what I am trying to accomplish and ask if they would provide feedback. I have done this on several occasions and cannot recall any colleague who has refused to provide comments. The comments are usually detailed and quite constructive and have a different tone from those that emanate from the journal review process. The comments in turn can be used to devise the version that is submitted for publication.

5. Closing Comments

Designing and completing a study requires many skills. Publication and communication of results of research represent a separate set of skills and most of these skills are not mentioned or detailed in graduate training. I have mentioned three tasks that are involved in preparing a manuscript for journal publication: description, explanation, and contextualization of the study. The writing we are routinely taught in science focuses on description, but the other portions are central as well and determine whether a study not only appears to be important but also in fact actually is. Recommendations were made in what to address and how to incorporate description, explanation, and contextualization within the different sections of a manuscript (e.g., Introduction, Method).

It is often useful to identify a model study from one’s own reading that nicely integrates description, explanation, and contextualization. Read this paper for content and then evaluate sections and paragraphs from a higher level of abstraction. What does this paragraph accomplish in leading to the next section, what did the author do to make the case for the study, how did she keep the same story line of the Introduction, Results, and Discussion very clear, and so on? These meta-level questions can help identify a template to better operationalize points I have emphasized.

Another way to approach the task of preparing the manuscript is to consider the set of questions that ought to be addressed. Questions were provided to direct the researcher to the types of issues reviewers are likely to ask about a manuscript. I mentioned the many guidelines now that govern research. These guidelines sometimes must be followed as a matter of policy for various journals. The guidelines are useful in relation to identifying key facets of a study and a report that need to be addressed including clarity of all facets of the study, transparency of procedures, ethical issues and attention to participants, and others. All these facets are obviously important, but they are more focused on description rather than explanation and contextualization. As you prepare the manuscript, give great attention to these latter components because these areas are likely to be the Achilles heel as the manuscript is evaluated for publication.

Publication of one’s research has many rewards. Certainly salient among them are generating new knowledge. There is a canvas of ignorance that is still mostly blank and one’s research can paint one stroke. That is hugely rewarding. Added external rewards are often available as well. Fame and fortune are not likely, but one’s publication record can contribute directly to job and job promotion and the opportunity to work with students at all levels and postdoctoral researchers who join in and improve the work by their ideas. Research also helps one’s own thinking that began with conceptualization of the study and an effort to better understand the phenomenon. Writing up the results often helps to extend one’s own thinking further and hence is a critical step in the next study or in conceptualization of the topic or area. This is a reciprocal process where we too are influenced by the publications of others and hopefully exert influence with our own publications. In short, publication is not just about publication but is a gateway experience that fosters many additional fulfilling activities including participation in the larger scientific agenda and community.

Footnotes

1 Preparing a manuscript for publication entails several format requirements, such as print style and size, citations of sources, use of abbreviations, structure of tables and figures, and order in which sections of the article appears. These are detailed in the Publication Manual of the American Psychological Association (APA, 2020b) and are not covered in this chapter.

2 Arguably the most visible case in the past 20 years was the fabricated report that a commonly used vaccine (measles, mumps, rubella) in young children caused autism among well-functioning children (Reference Wakefield, Murch, Anthony, Linnell, Casson, Malik, Berelowitz, Dhillon, Thomson, Harvey, Valentine, Davies and Walker-SmithWakefield et al., 1998). That the data were faked eventually came to light, but not after far-reaching consequences, including an enormous international backlash against vaccines and the unnecessary deaths of many children whose parents refused vaccinations. Antivaccination movements antedate this report, but social media and the Internet permitted this one to spread widely and over an extended period that continues to this day (Reference Yang, Broniatowski and ReissYang et al., 2019).

3 A quantitative measure to evaluate journals is referred to as the “impact factor,” and is based on the frequency with which articles appear in the journal in a given time period (2 years) in proportion to the total number articles published in the journal. An objective quantitative measure of impact has multiple uses for different parties who have interest in the impact of a journal (e.g., libraries making subscription decisions, publisher evaluating the status of a specific journal they have published). Administrators and faculty peers often use impact of the journals in which a colleague publishes as well as how often their work is cited by others among the criteria used for job appointments and promotions in academic rank, and salary adjustments. There has been a strong movement to no longer use the impact factor to evaluate research or merit of an investigator (see Reference AlbertsAlberts, 2013). Impact was not designed to measure that and is subject to all sorts of influences (e.g., that vary by discipline, artifacts of publishing practices of individual journals). Moreover, that impact factor bears little relation to expert views of scientific quality. In 2012, an organization (San Francisco Declaration of Research Assessment, abbreviated as DORA), initiated at a meeting of the American Society for Cell Biology and including many editors and publishers, examined the ways in which journals are evaluated. Among the consequences was agreement that “impact factor” might be useful for the purposes for which it was intended, but not for evaluating the merit of scientific research. Consequently, DORA urged journals and scientific organizations to drop the use of impact factor as an index of quality of the journal or articles in which the journal appears. Now many scientific and professional organizations (>1800 at the time of this writing) and researchers (~15,000) have signed on to this recommendation to not use or flaunt impact factor as an index of quality (https://sfdora.org/read/). Even so, many journals still flaunt their “impact factor” and occasionally researchers promote their own work based on the impact factor of the journal in which their work has appeared. It is important to mention here in case the reader is considering this as a main or major reason for submitting a manuscript to one journal rather than another.

4 Excellent readings are available to prepare the author for the journal review process (The Trial by Kafka, The Myth of Sisyphus by Camus, Inferno by Dante, and Nausea by Sartre). Some experiences (e.g., root canal without an anesthetic, bungee jumping with a cord that does not stretch in any way, income tax audit) also are touted to be helpful because they evoke reactions that mimic those experienced when reading reviews of one manuscript.

5 Thanks to my dissertation committee for letting me quote from their comments.

References

Aalbersberg, I. J., Appleyard, T., Brookhart, S., Carpenter, T., Clarke, M., Curry, S., Dahl, J., DeHaven, A., Eich, E., Franko, M., Freedman, L., Graf, C., Grant, S., Hanson, B., Joseph, H., Kiermer, V., Kramer, B., Kraut, A., Karn, R. K., … Vazire, S. (2018, February 15). Making science transparent by default; Introducing the TOP Statement. https://osf.io/sm78t/?_ga=2.66881262.1762683141.1579096697-214340795.1579096697Google Scholar
Alberts, B. (2013). Editorial: Impact factor distortions. Science, 340, 787.Google Scholar
American Educational Research Association. (2006). Standards for reporting on empirical social science research in AERA publications. Educational Researcher, 35, 3340.Google Scholar
American Psychological Association. (2020a). APA and affiliated journals. Online at www.apa.org/pubs/journalsGoogle Scholar
American Psychological Association. (2020b). Publication manual of the American Psychological Association (7th ed.). Washington, DC: American Psychological Association.Google Scholar
Appelbaum, M., Cooper, H., Kline, R. B., Mayo-Wilson, E., Nezu, A. M., & Rao, S. M. (2018). Journal article reporting standards for quantitative research in psychology: The APA Publications and Communications Board task force report. American Psychologist, 73(1), 325.Google Scholar
Association for Psychological Science. (2020). APS journals. On line at www.psychologicalscience.org/publicationsGoogle Scholar
Bailar, J. C. III., & Patterson, K. (1985). Journal of peer review: The need for a research agenda. New England Journal of Medicine, 312, 654657.Google Scholar
Benos, D. J., Bashari, E., Chaves, J. M., Gaggar, A., Kapoor, N., LaFrance, M., Mans, R., Mayhew, D., McGowan, S., Polter, A., Qadri, Y., Sarfare, S., Schultz, K., Splittgerber, R., Stephenson, J., Tower, C., Walton, A. G., & Zotov, A. (2007). The ups and downs of peer review. Advances in Physiology Education, 31(2), 145152.Google Scholar
Brainard, J. (2020). Articles in ‘predatory’ journals receive few or no citations. Science, 367(6474), 139.Google Scholar
Case, L., & Smith, T. B. (2000). Ethnic representation in a sample of the literature of applied psychology. Journal of Consulting and Clinical Psychology, 68, 11071110.CrossRefGoogle Scholar
Cicchetti, D. V. (1991). The reliability of the peer review for manuscript and grant submissions: A cross-disciplinary investigation. Behavioral and Brain Sciences, 14, 119186.Google Scholar
Cooper, H. (2020). Reporting quantitative research in psychology: How to meet APA style journal article reporting standards (2nd ed.). Washington, DC: American Psychological Association.CrossRefGoogle Scholar
De Los Reyes, A., & Kazdin, A. E. (2008). When the evidence says, “yes, no, and maybe so”: Attending to and interpreting inconsistent findings among evidence-based interventions. Current Directions in Psychological Science, 17, 4751.CrossRefGoogle ScholarPubMed
DeHaven, A. (2017, May 23). Preregistration: A plan, not a prison. Retrieved from https://cos.io/blog/preregistration-plan-not-prison/Google Scholar
Des Jarlais, D. C., Lyles, C., Crepaz, N., & the TREND Group. (2004). Improving the reporting quality of nonrandomized evaluations of behavioral and public health interventions: The TREND statement. American Journal of Public Health, 94, 361366.Google Scholar
Elson, M., Huff, M., & Utz, S. (2020). Meta science on peer review: testing the effects of study originality and statistical significance in a field experiment. Advances in Methods and Practices in Psychological Science, 3(1), 5365.Google Scholar
Francis, G. (2012). The psychology of replication and replication in psychology. Perspectives on Psychological Science, 7(6), 585594.Google Scholar
Gerber, A., Arceneaux, K., Boudreau, C., Dowling, C., Hillygus, S., Palfrey, T., Biggers, D. R., & Hendry, D. J. (2014). Reporting guidelines for experimental research: A report from the experimental research section standards committee. Journal of Experimental Political Science, 1(1), 8198.CrossRefGoogle Scholar
Gunther, A. (2011). PSYCLINE: Your guide to psychology and social science journals on the web. Retrieved August 2011 from www.psycline.org/journals/psycline.htmlGoogle Scholar
Hawkins, E. (2019, October 21). Journals test the Materials Design Analysis Reporting (MDAR) checklist of shemes and memes community. Blog from Nature.com. http://blogs.nature.com/ofschemesandmemes/author/lizhGoogle Scholar
Ioannidis, J. P. (2005). Why most published research findings are false. PLoS Medicine, 2(8), e124.Google Scholar
Kazdin, A. E. (2017). Research design in clinical psychology (5th ed.). Boston, MA: Pearson.Google Scholar
Kepes, S., McDaniel, M. A., Brannick, M. T., & Banks, G. C. (2013). Meta-analytic reviews in the organizational sciences: Two meta-analytic schools on the way to MARS (the Meta-Analytic Reporting Standards). Journal of Business and Psychology, 28(2), 123143.Google Scholar
Kirman, C. R., Simon, T. W., & Hays, S. M. (2019). Science peer review for the 21st century: Assessing scientific consensus for decision-making while managing conflict of interests, reviewer and process bias. Regulatory Toxicology and Pharmacology, 103, 7385.Google Scholar
Lehrer, J. (2010). The truth wears off. The New Yorker. Available at www.newyorker.com/reporting/2010/12/13/101213fa_fact_lehrerGoogle Scholar
Levelt, Noort, and Committees, Drenth (2012, November). Flawed science: The fraudulent research practices of social psychologist Diederik Stapel. Available at www.commissielevelt.nl/wp-content/uploads_per_blog/commissielevelt/2012/11/120695_Rapp_nov_2012_UK_web.pdfGoogle Scholar
Levitt, H. M., Bamberg, M., Creswell, J. W., Frost, D., Josselson, R., & Suárez-Orozco, C. (2018). Journal article reporting standards for qualitative research in psychology: The APA publications and communications board task force report. American Psychologist, 73, 2646.CrossRefGoogle Scholar
Ma, C., Liu, Y., Neumann, S., & Gao, X. (2017). Nicotine from cigarette smoking and diet and Parkinson disease: A review. Translational Neurodegeneration, 6(1), 18. https://translationalneurodegeneration.biomedcentral.com/track/pdf/10.1186/s40035-017-0090-8Google Scholar
Metcalf, J., & Crawford, K. (2016). Where are human subjects in big data research? The emerging ethics divide. Big Data & Society, 3(1), 2053951716650211.Google Scholar
Miller, L. R., & Das, S. K. (2007). Cigarette smoking and Parkinson’s disease. Experimental and Clinical Sciences International, 6, 9399.Google Scholar
Moher, D., Schulz, K.F., & Altman, D. (2001). The CONSORT statement: Revised recommendations for improving the quality of reports of parallel-group randomized trials. Journal of American Medical Association, 285, 19871991.CrossRefGoogle Scholar
Moonesinghe, R., Khoury, M. J., & Janssens, A. C. J. W (2007). Most published research findings are false – But a little replication goes a long way. PLoS Medicine, 4(2), e28.CrossRefGoogle ScholarPubMed
Nosek, B. A., Ebersole, C. R., DeHaven, A. C., & Mellor, D. T. (2018). The preregistration revolution. Proceedings of the National Academy of Sciences, 115, 26002606.CrossRefGoogle ScholarPubMed
Pryczak, F. (2017). Writing empirical research reports: A basic guide for students of the social and behavioral sciences (8th ed.). Abingdon: Routledge.Google Scholar
Simmons, J. P., Nelson, L. D., & Simonsohn, U. (2011). False-positive psychology: Undisclosed flexibility in data collection and analysis allows presenting anything as significant. Psychological Science, 22, 13591366.CrossRefGoogle ScholarPubMed
Smith, R. (2006). Peer review: A flawed process at the heart of science and journals. Journal of the Royal Society of Medicine, 99, 178182.CrossRefGoogle ScholarPubMed
Spier, R. (2002). The history of the peer-review process. Trends in Biotechnology, 20, 357358.Google Scholar
Stahel, P. F., & Moore, E. E. (2014). Peer review for biomedical publications: We can improve the system. BMC Medicine, 12(1), 179.CrossRefGoogle ScholarPubMed
Suresh, S. (2011). Moving toward global science. Science, 333, 802.Google Scholar
Tate, R. L., Perdices, M., Rosenkoetter, U., McDonald, S., Togher, L., Shadish, W., Horner, R., Kratochwill, T., Barlow, D. H., Kazdin, A., Sampson, M., Shamseer, L., & Sampson, M. (2016). The Single-Case Reporting Guideline In BEhavioural Interventions (SCRIBE) 2016: Explanation and elaboration. Archives of Scientific Psychology, 4(1), 1031.Google Scholar
Thomson Reuters. (2011). Journal search: Psychology. New York: Thomson Reuters. Retrieved August 2011 from http://science.thomsonreuters.com/cgi-bin/jrnlst/jlresults.cgi?PC=MASTER&Word=psychologyGoogle Scholar
Thursby, G. (2011). Psychology virtual library: Journals (electronic and print). Retrieved August 2011 from www.vl-site.org/psychology/journals.htmlGoogle Scholar
Von Elm, E., Altman, D. G., Egger, M., Pocock, S. J., Gøtzsche, P. C., & Vandenbroucke, J. P. (2007). The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) statement: Guidelines for reporting observational studies. Annals of Internal Medicine, 147(8), 573577.Google Scholar
Wakefield, A. J., Murch, S. H., Anthony, A., Linnell, J., Casson, D. M., Malik, M., Berelowitz, M., Dhillon, A. P., Thomson, M. A., Harvey, P., Valentine, A., Davies, S. E., & Walker-Smith, J. A. (1998). RETRACTED: Ileal-lymphoid-nodular hyperplasia, non-specific colitis, and pervasive developmental disorder in children. The Lancet, 351(9103), 637641.Google Scholar
Watanabe, M., & Aoki, M. (2014, January 10). Researcher: Test data falsified in major Alzheimer’s disease project. The Ashai Shimbun. Available online at https://ajw.asahi.com/article/behind_news/social_affairs/AJ201401100085Google Scholar
Web of Science. (2011). 2010 Journal citation reports. New York: Thomson Reuters. Retrieved August 2011 from http://wokinfo.com/products_tools/analytical/jcr/Google Scholar
Yang, Y. T., Broniatowski, D. A., & Reiss, D. R. (2019). Government role in regulating vaccine misinformation on social media platforms. JAMA Pediatrics. Published online September 03, 2019.Google Scholar
Zimmer, M. (2010). “But the data is already public”: On the ethics of research in Facebook. Ethics and Information Technology, 12(4), 313325.Google Scholar
Figure 0

Table 13.1 Major questions to guide journal article preparation

Save book to Kindle

To save this book to your Kindle, first ensure coreplatform@cambridge.org is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about saving to your Kindle.

Note you can select to save to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.

Find out more about the Kindle Personal Document Service.

Available formats
×

Save book to Dropbox

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Dropbox.

Available formats
×

Save book to Google Drive

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Google Drive.

Available formats
×