• Privacy Policy

Research Method

Home » Evaluating Research – Process, Examples and Methods

Evaluating Research – Process, Examples and Methods

Table of Contents

Evaluating Research

Evaluating Research

Definition:

Evaluating Research refers to the process of assessing the quality, credibility, and relevance of a research study or project. This involves examining the methods, data, and results of the research in order to determine its validity, reliability, and usefulness. Evaluating research can be done by both experts and non-experts in the field, and involves critical thinking, analysis, and interpretation of the research findings.

Research Evaluating Process

The process of evaluating research typically involves the following steps:

Identify the Research Question

The first step in evaluating research is to identify the research question or problem that the study is addressing. This will help you to determine whether the study is relevant to your needs.

Assess the Study Design

The study design refers to the methodology used to conduct the research. You should assess whether the study design is appropriate for the research question and whether it is likely to produce reliable and valid results.

Evaluate the Sample

The sample refers to the group of participants or subjects who are included in the study. You should evaluate whether the sample size is adequate and whether the participants are representative of the population under study.

Review the Data Collection Methods

You should review the data collection methods used in the study to ensure that they are valid and reliable. This includes assessing the measures used to collect data and the procedures used to collect data.

Examine the Statistical Analysis

Statistical analysis refers to the methods used to analyze the data. You should examine whether the statistical analysis is appropriate for the research question and whether it is likely to produce valid and reliable results.

Assess the Conclusions

You should evaluate whether the data support the conclusions drawn from the study and whether they are relevant to the research question.

Consider the Limitations

Finally, you should consider the limitations of the study, including any potential biases or confounding factors that may have influenced the results.

Evaluating Research Methods

Evaluating Research Methods are as follows:

  • Peer review: Peer review is a process where experts in the field review a study before it is published. This helps ensure that the study is accurate, valid, and relevant to the field.
  • Critical appraisal : Critical appraisal involves systematically evaluating a study based on specific criteria. This helps assess the quality of the study and the reliability of the findings.
  • Replication : Replication involves repeating a study to test the validity and reliability of the findings. This can help identify any errors or biases in the original study.
  • Meta-analysis : Meta-analysis is a statistical method that combines the results of multiple studies to provide a more comprehensive understanding of a particular topic. This can help identify patterns or inconsistencies across studies.
  • Consultation with experts : Consulting with experts in the field can provide valuable insights into the quality and relevance of a study. Experts can also help identify potential limitations or biases in the study.
  • Review of funding sources: Examining the funding sources of a study can help identify any potential conflicts of interest or biases that may have influenced the study design or interpretation of results.

Example of Evaluating Research

Example of Evaluating Research sample for students:

Title of the Study: The Effects of Social Media Use on Mental Health among College Students

Sample Size: 500 college students

Sampling Technique : Convenience sampling

  • Sample Size: The sample size of 500 college students is a moderate sample size, which could be considered representative of the college student population. However, it would be more representative if the sample size was larger, or if a random sampling technique was used.
  • Sampling Technique : Convenience sampling is a non-probability sampling technique, which means that the sample may not be representative of the population. This technique may introduce bias into the study since the participants are self-selected and may not be representative of the entire college student population. Therefore, the results of this study may not be generalizable to other populations.
  • Participant Characteristics: The study does not provide any information about the demographic characteristics of the participants, such as age, gender, race, or socioeconomic status. This information is important because social media use and mental health may vary among different demographic groups.
  • Data Collection Method: The study used a self-administered survey to collect data. Self-administered surveys may be subject to response bias and may not accurately reflect participants’ actual behaviors and experiences.
  • Data Analysis: The study used descriptive statistics and regression analysis to analyze the data. Descriptive statistics provide a summary of the data, while regression analysis is used to examine the relationship between two or more variables. However, the study did not provide information about the statistical significance of the results or the effect sizes.

Overall, while the study provides some insights into the relationship between social media use and mental health among college students, the use of a convenience sampling technique and the lack of information about participant characteristics limit the generalizability of the findings. In addition, the use of self-administered surveys may introduce bias into the study, and the lack of information about the statistical significance of the results limits the interpretation of the findings.

Note*: Above mentioned example is just a sample for students. Do not copy and paste directly into your assignment. Kindly do your own research for academic purposes.

Applications of Evaluating Research

Here are some of the applications of evaluating research:

  • Identifying reliable sources : By evaluating research, researchers, students, and other professionals can identify the most reliable sources of information to use in their work. They can determine the quality of research studies, including the methodology, sample size, data analysis, and conclusions.
  • Validating findings: Evaluating research can help to validate findings from previous studies. By examining the methodology and results of a study, researchers can determine if the findings are reliable and if they can be used to inform future research.
  • Identifying knowledge gaps: Evaluating research can also help to identify gaps in current knowledge. By examining the existing literature on a topic, researchers can determine areas where more research is needed, and they can design studies to address these gaps.
  • Improving research quality : Evaluating research can help to improve the quality of future research. By examining the strengths and weaknesses of previous studies, researchers can design better studies and avoid common pitfalls.
  • Informing policy and decision-making : Evaluating research is crucial in informing policy and decision-making in many fields. By examining the evidence base for a particular issue, policymakers can make informed decisions that are supported by the best available evidence.
  • Enhancing education : Evaluating research is essential in enhancing education. Educators can use research findings to improve teaching methods, curriculum development, and student outcomes.

Purpose of Evaluating Research

Here are some of the key purposes of evaluating research:

  • Determine the reliability and validity of research findings : By evaluating research, researchers can determine the quality of the study design, data collection, and analysis. They can determine whether the findings are reliable, valid, and generalizable to other populations.
  • Identify the strengths and weaknesses of research studies: Evaluating research helps to identify the strengths and weaknesses of research studies, including potential biases, confounding factors, and limitations. This information can help researchers to design better studies in the future.
  • Inform evidence-based decision-making: Evaluating research is crucial in informing evidence-based decision-making in many fields, including healthcare, education, and public policy. Policymakers, educators, and clinicians rely on research evidence to make informed decisions.
  • Identify research gaps : By evaluating research, researchers can identify gaps in the existing literature and design studies to address these gaps. This process can help to advance knowledge and improve the quality of research in a particular field.
  • Ensure research ethics and integrity : Evaluating research helps to ensure that research studies are conducted ethically and with integrity. Researchers must adhere to ethical guidelines to protect the welfare and rights of study participants and to maintain the trust of the public.

Characteristics Evaluating Research

Characteristics Evaluating Research are as follows:

  • Research question/hypothesis: A good research question or hypothesis should be clear, concise, and well-defined. It should address a significant problem or issue in the field and be grounded in relevant theory or prior research.
  • Study design: The research design should be appropriate for answering the research question and be clearly described in the study. The study design should also minimize bias and confounding variables.
  • Sampling : The sample should be representative of the population of interest and the sampling method should be appropriate for the research question and study design.
  • Data collection : The data collection methods should be reliable and valid, and the data should be accurately recorded and analyzed.
  • Results : The results should be presented clearly and accurately, and the statistical analysis should be appropriate for the research question and study design.
  • Interpretation of results : The interpretation of the results should be based on the data and not influenced by personal biases or preconceptions.
  • Generalizability: The study findings should be generalizable to the population of interest and relevant to other settings or contexts.
  • Contribution to the field : The study should make a significant contribution to the field and advance our understanding of the research question or issue.

Advantages of Evaluating Research

Evaluating research has several advantages, including:

  • Ensuring accuracy and validity : By evaluating research, we can ensure that the research is accurate, valid, and reliable. This ensures that the findings are trustworthy and can be used to inform decision-making.
  • Identifying gaps in knowledge : Evaluating research can help identify gaps in knowledge and areas where further research is needed. This can guide future research and help build a stronger evidence base.
  • Promoting critical thinking: Evaluating research requires critical thinking skills, which can be applied in other areas of life. By evaluating research, individuals can develop their critical thinking skills and become more discerning consumers of information.
  • Improving the quality of research : Evaluating research can help improve the quality of research by identifying areas where improvements can be made. This can lead to more rigorous research methods and better-quality research.
  • Informing decision-making: By evaluating research, we can make informed decisions based on the evidence. This is particularly important in fields such as medicine and public health, where decisions can have significant consequences.
  • Advancing the field : Evaluating research can help advance the field by identifying new research questions and areas of inquiry. This can lead to the development of new theories and the refinement of existing ones.

Limitations of Evaluating Research

Limitations of Evaluating Research are as follows:

  • Time-consuming: Evaluating research can be time-consuming, particularly if the study is complex or requires specialized knowledge. This can be a barrier for individuals who are not experts in the field or who have limited time.
  • Subjectivity : Evaluating research can be subjective, as different individuals may have different interpretations of the same study. This can lead to inconsistencies in the evaluation process and make it difficult to compare studies.
  • Limited generalizability: The findings of a study may not be generalizable to other populations or contexts. This limits the usefulness of the study and may make it difficult to apply the findings to other settings.
  • Publication bias: Research that does not find significant results may be less likely to be published, which can create a bias in the published literature. This can limit the amount of information available for evaluation.
  • Lack of transparency: Some studies may not provide enough detail about their methods or results, making it difficult to evaluate their quality or validity.
  • Funding bias : Research funded by particular organizations or industries may be biased towards the interests of the funder. This can influence the study design, methods, and interpretation of results.

About the author

' src=

Muhammad Hassan

Researcher, Academic Writer, Web developer

You may also like

Research Questions

Research Questions – Types, Examples and Writing...

Institutional Review Board (IRB)

Institutional Review Board – Application Sample...

References in Research

References in Research – Types, Examples and...

Conceptual Framework

Conceptual Framework – Types, Methodology and...

Limitations in Research

Limitations in Research – Types, Examples and...

Research Gap

Research Gap – Types, Examples and How to...

  • EXPLORE Random Article
  • Happiness Hub

How to Evaluate a Research Paper

Last Updated: July 4, 2023 Fact Checked

This article was co-authored by Matthew Snipp, PhD . C. Matthew Snipp is the Burnet C. and Mildred Finley Wohlford Professor of Humanities and Sciences in the Department of Sociology at Stanford University. He is also the Director for the Institute for Research in the Social Science’s Secure Data Center. He has been a Research Fellow at the U.S. Bureau of the Census and a Fellow at the Center for Advanced Study in the Behavioral Sciences. He has published 3 books and over 70 articles and book chapters on demography, economic development, poverty and unemployment. He is also currently serving on the National Institute of Child Health and Development’s Population Science Subcommittee. He holds a Ph.D. in Sociology from the University of Wisconsin—Madison. There are 14 references cited in this article, which can be found at the bottom of the page. This article has been fact-checked, ensuring the accuracy of any cited facts and confirming the authority of its sources. This article has been viewed 57,265 times.

While writing a research paper can be tough, it can be even tougher to evaluate the strength of one. Whether you’re giving feedback to a fellow student or learning how to grade, you can judge a paper’s content and formatting to determine its excellence. By looking at the key attributes of successful research papers in the humanities and sciences, you can evaluate a given work work against high standards.

Examining a Humanities Research Paper

Step 1 Look for the thesis statement on page 1 of the paper.

  • A strong thesis statement might be: “Single-sex education helps girls develop more self-confidence than co-education. Using scholarly research, I will illustrate how young girls develop greater self-confidence in single-sex schools, why their self-confidence becomes more developed in this setting, and what this means for the future of education in America.”
  • A poor thesis statement might be: “Single-sex education is a schooling choice that separates the sexes.” This is a description of single-sex education rather than an opinion that can be supported with research.

Step 2 Judge if the thesis is debatable.

  • In the example outlined above, one could theoretically create a valid argument for single-sex education negatively impacting the self-esteem of girls. This depth makes the topic worthy of examination.

Step 3 Assess whether the thesis is original.

  • Ask yourself if the thesis feels obvious. If it does, it is probably not a strong choice.

Step 4 Find at least 3 points supporting the thesis statement.

  • The paragraphs should each start with a topic sentence so you feel introduced to the research at hand.
  • A typical 5-paragraph essay will have 3 supporting points of 1 paragraph each. Most good research papers are longer than 5 paragraphs, though, and may have multiple paragraphs about just one point of many that support the thesis.
  • The more points and correlating research that support the thesis statement, the better.

Step 5 Identify research quotations that reinforce the points.

  • A paper with little supporting research to bolster its points likely falls short of adequately illustrating its thesis.
  • Quotations longer than 4 lines should be set off in block format for readability.

Step 6 Identify context and analysis for each research quotation.

  • For example, if a thesis statement were that cats are smarter than dogs. A good supporting point might be that cats are better hunters than dogs. The author could introduce a source well in support of this by saying, “In animal expert Linda Smith’s book Cats are King , Smith describes cats’ superior hunting abilities. On page 8 she says, ‘Cats are the most developed hunters in the civilized world.’ Because hunting requires incredible mental focus and skill, this statement supports my view that cats are smarter than dogs.”
  • Any quotations that are used to summarize a text are likely not pulling their weight in the research paper. All quotations should serve as direct supports.

Step 7 Find an acknowledgement of potential objections.

  • Doing this is a mark of a strong research paper and helps convince the reader that the author’s thesis is good and valid.

Step 8 Look for a conclusion that discusses larger implication of the thesis.

  • A good research paper shows that the thesis is important beyond just the narrow context of its question.

Evaluating a Sciences Research Paper

Step 1 Look for an abstract of 300 words or less.

  • An effective abstract should also include a brief interpretation of those results, which will be explored further later.
  • The author should note any overall trends or major revelations discovered in their research.

Step 2 Identify an introduction that provides a guide for the reader.

  • The author of a successful scientific research paper should note any boundaries that limit their research.
  • For example: If the author conducted a study of pregnant women, but only women over 35 responded to a call for participants, the author should mention it. An element like this could affect the conclusions the author draws from their research.

Step 3 Look for a methodology section that describes the author’s approach.

  • An effective methodology section should be written in the past tense, as the author has already decided and executed their research.

Step 4 Read for a results section that confirms or rejects the original theory.

  • A good results section shouldn’t include raw data from the research; it should explain the findings of the research.

Step 5 Look for a discussion section that illuminates new insights.

  • A strong discussion section could present new solutions to the original problem or suggest further areas of study given the results.
  • A good discussion moves beyond interpreting findings and offers some subjective analysis.

Step 6 Read for a conclusion that demonstrates the importance of the findings.

  • The author might think about the potential real-world consequences of ignoring the research results, for example.

Checking the Formatting of a Research Paper

Step 1 Check for the author’s name, course, instructor name, and date.

  • For example: “Growing up Stronger: Why Girls’ Schools Create More Confident Women” is more interesting and concise than “Single-Sex Schools are Better than Co-Ed Schools at Developing Self Confidence for Girls.”

Step 3 Verify that the font is standard and readable.

  • The author should clearly identify any ideas that are not their own and reference other works that make up the research landscape for context.
  • Common style guides for research papers include the APA style guide, the Chicago Manual of Style, the MLA Handbook, and the Turabian citation guide.

Expert Q&A

Matthew Snipp, PhD

You Might Also Like

Get a Loan Even With Bad Credit

  • ↑ https://writingcenter.unc.edu/tips-and-tools/thesis-statements/
  • ↑ https://owl.purdue.edu/owl/general_writing/academic_writing/establishing_arguments/index.html
  • ↑ https://owl.english.purdue.edu/owl/resource/658/02
  • ↑ https://advice.writing.utoronto.ca/planning/thesis-statements/
  • ↑ https://writingcenter.unc.edu/tips-and-tools/quotations/
  • ↑ https://wts.indiana.edu/writing-guides/writing-conclusions.html
  • ↑ https://libguides.usc.edu/writingguide/abstract
  • ↑ https://libguides.usc.edu/writingguide/introduction
  • ↑ https://libguides.usc.edu/writingguide/methodology
  • ↑ https://libguides.usc.edu/writingguide/results
  • ↑ https://libguides.usc.edu/writingguide/discussion
  • ↑ https://libguides.usc.edu/writingguide/conclusion
  • ↑ https://writing.umn.edu/sws/assets/pdf/quicktips/academicessaystructures.pdf
  • ↑ https://subjectguides.library.american.edu/citation

About this article

Matthew Snipp, PhD

Did this article help you?

Get a Loan Even With Bad Credit

  • About wikiHow
  • Terms of Use
  • Privacy Policy
  • Do Not Sell or Share My Info
  • Not Selling Info

how to evaluate a research paper

site header image

Evaluating Research Articles

Understanding research statistics, critical appraisal, help us improve the libguide.

how to evaluate a research paper

Imagine for a moment that you are trying to answer a clinical (PICO) question regarding one of your patients/clients. Do you know how to determine if a research study is of high quality? Can you tell if it is applicable to your question? In evidence based practice, there are many things to look for in an article that will reveal its quality and relevance. This guide is a collection of resources and activities that will help you learn how to evaluate articles efficiently and accurately.

Is health research new to you? Or perhaps you're a little out of practice with reading it? The following questions will help illuminate an article's strengths or shortcomings. Ask them of yourself as you are reading an article:

  • Is the article peer reviewed?
  • Are there any conflicts of interest based on the author's affiliation or the funding source of the research?
  • Are the research questions or objectives clearly defined?
  • Is the study a systematic review or meta analysis?
  • Is the study design appropriate for the research question?
  • Is the sample size justified? Do the authors explain how it is representative of the wider population?
  • Do the researchers describe the setting of data collection?
  • Does the paper clearly describe the measurements used?
  • Did the researchers use appropriate statistical measures?
  • Are the research questions or objectives answered?
  • Did the researchers account for confounding factors?
  • Have the researchers only drawn conclusions about the groups represented in the research?
  • Have the authors declared any conflicts of interest?

If the answer to these questions about an article you are reading are mostly YESes , then it's likely that the article is of decent quality. If the answers are most NOs , then it may be a good idea to move on to another article. If the YESes and NOs are roughly even, you'll have to decide for yourself if the article is good enough quality for you. Some factors, like a poor literature review, are not as important as the researchers neglecting to describe the measurements they used. As you read more research, you'll be able to more easily identify research that is well done vs. that which is not well done.

North Carolina State University Libraries has a short video that describes how to evaluate sources for credibility. 

how to evaluate a research paper

Determining if a research study has used appropriate statistical measures is one of the most critical and difficult steps in evaluating an article. The following links are great, quick resources for helping to better understand how to use statistics in health research.

how to evaluate a research paper

  • How to read a paper: Statistics for the non-statistician. II: “Significant” relations and their pitfalls This article continues the checklist of questions that will help you to appraise the statistical validity of a paper. Greenhalgh Trisha. How to read a paper: Statistics for the non-statistician. II: “Significant” relations and their pitfalls BMJ 1997; 315 :422 *On the PMC PDF, you need to scroll past the first article to get to this one.*
  • A consumer's guide to subgroup analysis The extent to which a clinician should believe and act on the results of subgroup analyses of data from randomized trials or meta-analyses is controversial. Guidelines are provided in this paper for making these decisions.

Statistical Versus Clinical Significance

When appraising studies, it's important to consider both the clinical and statistical significance of the research. This video offers a quick explanation of why.

If you have a little more time, this video explores statistical and clinical significance in more detail, including examples of how to calculate an effect size.

  • Statistical vs. Clinical Significance Transcript Transcript document for the Statistical vs. Clinical Significance video.
  • Effect Size Transcript Transcript document for the Effect Size video.
  • P Values, Statistical Significance & Clinical Significance This handout also explains clinical and statistical significance.
  • Absolute versus relative risk – making sense of media stories Understanding the difference between relative and absolute risk is essential to understanding statistical tests commonly found in research articles.

Critical appraisal is the process of systematically evaluating research using established and transparent methods. In critical appraisal, health professionals use validated checklists/worksheets as tools to guide their assessment of the research. It is a more advanced way of evaluating research than the more basic method explained above. To learn more about critical appraisal or to access critical appraisal tools, visit the websites below.

how to evaluate a research paper

  • Last Updated: Sep 19, 2024 8:04 PM
  • URL: https://libguides.massgeneral.org/evaluatingarticles

how to evaluate a research paper

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

The PMC website is updating on October 15, 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • Perspect Clin Res
  • v.12(2); Apr-Jun 2021

Critical appraisal of published research papers – A reinforcing tool for research methodology: Questionnaire-based study

Snehalata gajbhiye.

Department of Pharmacology and Therapeutics, Seth GS Medical College and KEM Hospital, Mumbai, Maharashtra, India

Raakhi Tripathi

Urwashi parmar, nishtha khatri, anirudha potey.

1 Department of Clinical Trials, Serum Institute of India, Pune, Maharashtra, India

Background and Objectives:

Critical appraisal of published research papers is routinely conducted as a journal club (JC) activity in pharmacology departments of various medical colleges across Maharashtra, and it forms an important part of their postgraduate curriculum. The objective of this study was to evaluate the perception of pharmacology postgraduate students and teachers toward use of critical appraisal as a reinforcing tool for research methodology. Evaluation of performance of the in-house pharmacology postgraduate students in the critical appraisal activity constituted secondary objective of the study.

Materials and Methods:

The study was conducted in two parts. In Part I, a cross-sectional questionnaire-based evaluation on perception toward critical appraisal activity was carried out among pharmacology postgraduate students and teachers. In Part II of the study, JC score sheets of 2 nd - and 3 rd -year pharmacology students over the past 4 years were evaluated.

One hundred and twenty-seven postgraduate students and 32 teachers participated in Part I of the study. About 118 (92.9%) students and 28 (87.5%) faculties considered the critical appraisal activity to be beneficial for the students. JC score sheet assessments suggested that there was a statistically significant improvement in overall scores obtained by postgraduate students ( n = 25) in their last JC as compared to the first JC.

Conclusion:

Journal article criticism is a crucial tool to develop a research attitude among postgraduate students. Participation in the JC activity led to the improvement in the skill of critical appraisal of published research articles, but this improvement was not educationally relevant.

INTRODUCTION

Critical appraisal of a research paper is defined as “The process of carefully and systematically examining research to judge its trustworthiness, value and relevance in a particular context.”[ 1 ] Since scientific literature is rapidly expanding with more than 12,000 articles being added to the MEDLINE database per week,[ 2 ] critical appraisal is very important to distinguish scientifically useful and well-written articles from imprecise articles.

Educational authorities like the Medical Council of India (MCI) and Maharashtra University of Health Sciences (MUHS) have stated in pharmacology postgraduate curriculum that students must critically appraise research papers. To impart training toward these skills, MCI and MUHS have emphasized on the introduction of journal club (JC) activity for postgraduate (PG) students, wherein students review a published original research paper and state the merits and demerits of the paper. Abiding by this, pharmacology departments across various medical colleges in Maharashtra organize JC at frequent intervals[ 3 , 4 ] and students discuss varied aspects of the article with teaching faculty of the department.[ 5 ] Moreover, this activity carries a significant weightage of marks in the pharmacology university examination. As postgraduate students attend this activity throughout their 3-year tenure, it was perceived by the authors that this activity of critical appraisal of research papers could emerge as a tool for reinforcing the knowledge of research methodology. Hence, a questionnaire-based study was designed to find out the perceptions from PG students and teachers.

There have been studies that have laid emphasis on the procedure of conducting critical appraisal of research papers and its application into clinical practice.[ 6 , 7 ] However, there are no studies that have evaluated how well students are able to critically appraise a research paper. The Department of Pharmacology and Therapeutics at Seth GS Medical College has developed an evaluation method to score the PG students on this skill and this tool has been implemented for the last 5 years. Since there are no research data available on the performance of PG Pharmacology students in JC, capturing the critical appraisal activity evaluation scores of in-house PG students was chosen as another objective of the study.

MATERIALS AND METHODS

Description of the journal club activity.

JC is conducted in the Department of Pharmacology and Therapeutics at Seth GS Medical College once in every 2 weeks. During the JC activity, postgraduate students critically appraise published original research articles on their completeness and aptness in terms of the following: study title, rationale, objectives, study design, methodology-study population, inclusion/exclusion criteria, duration, intervention and safety/efficacy variables, randomization, blinding, statistical analysis, results, discussion, conclusion, references, and abstract. All postgraduate students attend this activity, while one of them critically appraises the article (who has received the research paper given by one of the faculty members 5 days before the day of JC). Other faculties also attend these sessions and facilitate the discussions. As the student comments on various sections of the paper, the same predecided faculty who gave the article (single assessor) evaluates the student on a total score of 100 which is split per section as follows: Introduction –20 marks, Methodology –20 marks, Discussion – 20 marks, Results and Conclusion –20 marks, References –10 marks, and Title, Abstract, and Keywords – 10 marks. However, there are no standard operating procedures to assess the performance of students at JC.

Methodology

After seeking permission from the Institutional Ethics Committee, the study was conducted in two parts. Part I consisted of a cross-sectional questionnaire-based survey that was conducted from October 2016 to September 2017. A questionnaire to evaluate perception towards the activity of critical appraisal of published papers as research methodology reinforcing tool was developed by the study investigators. The questionnaire consisted of 20 questions: 14 questions [refer Figure 1 ] graded on a 3-point Likert scale (agree, neutral, and disagree), 1 multiple choice selection question, 2 dichotomous questions, 1 semi-open-ended questions, and 2 open-ended questions. Content validation for this questionnaire was carried out with the help of eight pharmacology teachers. The content validity ratio per item was calculated and each item in the questionnaire had a CVR ratio (CVR) of >0.75.[ 8 ] The perception questionnaire was either E-mailed or sent through WhatsApp to PG pharmacology students and teaching faculty in pharmacology departments at various medical colleges across Maharashtra. Informed consent was obtained on E-mail from all the participants.

An external file that holds a picture, illustration, etc.
Object name is PCR-12-100-g001.jpg

Graphical representation of the percentage of students/teachers who agreed that critical appraisal of research helped them improve their knowledge on various aspects of research, perceived that faculty participation is important in this activity, and considered critical appraisal activity beneficial for students. The numbers adjacent to the bar diagrams indicate the raw number of students/faculty who agreed, while brackets indicate %

Part II of the study consisted of evaluating the performance of postgraduate students toward skills of critical appraisal of published papers. For this purpose, marks obtained by 2 nd - and 3 rd -year residents during JC sessions conducted over a period of 4 years from October 2013 to September 2017 were recorded and analyzed. No data on personal identifiers of the students were captured.

Statistical analysis

Marks obtained by postgraduate students in their first and last JC were compared using Wilcoxon signed-rank test, while marks obtained by 2 nd - and 3 rd -year postgraduate students were compared using Mann–Whitney test since the data were nonparametric. These statistical analyses were performed using GraphPad Prism statistical software, San Diego, Calfornia, USA, Version 7.0d. Data obtained from the perception questionnaire were entered in Microsoft Excel sheet and were expressed as frequencies (percentages) using descriptive statistics.

Participants who answered all items of the questionnaire were considered as complete responders and only completed questionnaires were analyzed. The questionnaire was sent through an E-mail to 100 students and through WhatsApp to 68 students. Out of the 100 students who received the questionnaire through E-mail, 79 responded completely and 8 were incomplete responders, while 13 students did not revert back. Out of the 68 students who received the questionnaire through WhatsApp, 48 responded completely, 6 gave an incomplete response, and 14 students did not revert back. Hence, of the 168 postgraduate students who received the questionnaire, 127 responded completely (student response rate for analysis = 75.6%). The questionnaire was E-mailed to 33 faculties and was sent through WhatsApp to 25 faculties. Out of the 33 faculties who received the questionnaire through E-mail, 19 responded completely, 5 responded incompletely, and 9 did not respond at all. Out of the 25 faculties who received the questionnaire through WhatsApp, 13 responded completely, 3 were incomplete responders, and 9 did not respond at all. Hence, of a total of 58 faculties who were contacted, 32 responded completely (faculty response rate for analysis = 55%). For Part I of the study, responses on the perception questionnaire from 127 postgraduate students and 32 postgraduate teachers were recorded and analyzed. None of the faculty who participated in the validation of the questionnaire participated in the survey. Number of responses obtained region wise (Mumbai region and rest of Maharashtra region) have been depicted in Table 1 .

Region-wise distribution of responses

Students ( =127)Faculty ( =32)
Mumbai colleges58 (45.7)18 (56.3)
Rest of Maharashtra colleges69 (54.3)14 (43.7)

Number of responses obtained from students/faculty belonging to Mumbai colleges and rest of Maharashtra colleges. Brackets indicate percentages

As per the data obtained on the Likert scale questions, 102 (80.3%) students and 29 (90.6%) teachers agreed that critical appraisal trains the students in doing a review of literature before selecting a particular research topic. Majority of the participants, i.e., 104 (81.9%) students and 29 (90.6%) teachers also believed that the activity increases student's knowledge regarding various experimental evaluation techniques. Moreover, 112 (88.2%) students and 27 (84.4%) faculty considered that critical appraisal activity results in improved skills of writing and understanding methodology section of research articles in terms of inclusion/exclusion criteria, endpoints, and safety/efficacy variables. About 103 (81.1%) students and 24 (75%) teachers perceived that this activity results in refinement of the student's research work. About 118 (92.9%) students and 28 (87.5%) faculty considered the critical appraisal activity to be beneficial for the students. Responses to 14 individual Likert scale items of the questionnaire have been depicted in Figure 1 .

With respect to the multiple choice selection question, 66 (52%) students and 16 (50%) teachers opined that faculty should select the paper, 53 (41.7%) students and 9 (28.1%) teachers stated that the papers should be selected by the presenting student himself/herself, while 8 (6.3%) students and 7 (21.9%) teachers expressed that some other student should select the paper to be presented at the JC.

The responses to dichotomous questions were as follows: majority of the students, that is, 109 (85.8%) and 23 (71.9%) teachers perceived that a standard checklist for article review should be given to the students before critical appraisal of journal article. Open-ended questions of the questionnaire invited suggestions from the participants regarding ways of getting trained on critical appraisal skills and of improving JC activity. Some of the suggestions given by faculty were as follows: increasing the frequency of JC activity, discussion of cited articles and new guidelines related to it, selecting all types of articles for criticism rather than only randomized controlled trials, and regular yearly exams on article criticism. Students stated that regular and frequent article criticism activity, practice of writing letter to the editor after criticism, active participation by peers and faculty, increasing weightage of marks for critical appraisal of papers in university examinations (at present marks are 50 out of 400), and a formal training for research criticism from 1 st year of postgraduation could improve critical appraisal program.

In Part II of this study, performance of the students on the skill of critical appraisal of papers was evaluated. Complete data of the first and last JC scores of a total of 25 students of the department were available, and when these scores were compared, it was seen that there was a statistically significant improvement in the overall scores ( P = 0.04), as well as in the scores obtained in methodology ( P = 0.03) and results section ( P = 0.02). This is depicted in Table 2 . Although statistically significant, the differences in scores in the methodology section, results section, and overall scores were 1.28/20, 1.28/20, and 4.36/100, respectively, amounting to 5.4%, 5.4%, and 4.36% higher scores in the last JC, which may not be considered educationally relevant (practically significant). The quantum of difference that would be considered practically significant was not decided a priori .

Comparison of marks obtained by pharmacology residents in their first and last journal club

SectionMarks obtained by pharmacology residents in their first journal club ( =25) Marks obtained by pharmacology residents in their last journal club ( =25) Wilcoxon signed-rank test
Mean±SDMedian (IQR)Mean±SDMedian (IQR) value
Introduction (maximum: 20 marks)13.48±2.5214 (12-16)14.28±2.3214 (13-16)0.22
Methodology (maximum: 20 marks)13.36±3.1114 (12-16)14.64±2.4014 (14-16.5)0.03*
Results and conclusion (maximum: 20 marks)13.60±2.4214 (12-15.5)14.88±2.6415 (13.5-16.5)0.02*
Discussion (maximum: 20 marks)13.44±3.2014 (11-16)14.16±2.7814 (12.5-16)0.12
References (maximum: 10 marks)7.12±1.207 (6.5-8)7.06±1.287 (6-8)0.80
Title, abstract, and keywords (maximum: 10 marks)7.44±0.927 (7-8)7.78±1.128 (7-9)0.17
Overall score (maximum: 100 marks)68.44±11.3972 (64-76)72.80±11.3271 (68-82.5)0.04*

Marks have been represented as mean±SD. The maximum marks that can be obtained in each section have been stated as maximum. *Indicates statistically significant ( P <0.05). IQR=Interquartile range, SD=Standard deviation

Scores of two groups, one group consisting of 2 nd -year postgraduate students ( n = 44) and second group consisting of 3 rd -year postgraduate students ( n = 32) were compared and revealed no statistically significant difference in overall score ( P = 0.84). This is depicted in Table 3 . Since the quantum of difference in the overall scores was meager 0.84/100 (0.84%), it cannot be considered practically significant.

Comparison of marks obtained by 2 nd - and 3 rd -year pharmacology residents in the activity of critical appraisal of research articles

SectionMarks obtained by 2 -year pharmacology students ( =44) Marks obtained by 3 -year pharmacology students ( =32) Mann-Whitney test, value
Mean±SDMedian (IQR)Mean±SDMedian (IQR)
Introduction (maximum: 20 marks)14.09±2.4114 (13-16)14.28±2.1414 (13-16)0.7527
Methodology (maximum: 20 marks)14.30±2.9014.5 (13-16)14.41±2.2414 (13-16)0.8385
Results and conclusion (maximum: 20 marks)14.09±2.4414 (12.5-16)14.59±2.6114.5 (13-16)0.4757
Discussion (maximum: 20 marks)13.86±2.7314 (12-16)14.16±2.7114.5 (12.5-16)0.5924
References (maximum: 10 marks)7.34±1.168 (7-8)7.05±1.407 (6-8)0.2551
Title, abstract, and keywords (maximum: 10 marks)7.82±0.908 (7-8.5)7.83±1.118 (7-8.5)0.9642
Overall score (maximum: 100 marks)71.50±10.7171.5 (66.5-79.5)72.34±10.8573 (66-79.5)0.8404

Marks have been represented as mean±SD. The maximum marks that can be obtained in each section have been stated as maximum. P <0.05 was considered to be statistically significant. IQR=Interquartile range, SD=Standard deviation

The present study gauged the perception of the pharmacology postgraduate students and teachers toward the use of critical appraisal activity as a reinforcing tool for research methodology. Both students and faculties (>50%) believed that critical appraisal activity increases student's knowledge on principles of ethics, experimental evaluation techniques, CONSORT guidelines, statistical analysis, concept of conflict of interest, current trends and recent advances in Pharmacology and trains on doing a review of literature, and improves skills on protocol writing and referencing. In the study conducted by Crank-Patton et al ., a survey on 278 general surgery program directors was carried out and more than 50% indicated that JC was important to their training program.[ 9 ]

The grading template used in Part II of the study was based on the IMRaD structure. Hence, equal weightage was given to the Introduction, Methodology, Results, and Discussion sections and lesser weightage was given to the references and title, abstract, and keywords sections.[ 10 ] While evaluating the scores obtained by 25 students in their first and last JC, it was seen that there was a statistically significant improvement in the overall scores of the students in their last JC. However, the meager improvement in scores cannot be considered educationally relevant, as the authors expected the students to score >90% for the upgrade to be considered educationally impactful. The above findings suggest that even though participation in the JC activity led to a steady increase in student's performance (~4%), the increment was not as expected. In addition, the students did not portray an excellent performance (>90%), with average scores being around 72% even in the last JC. This can be probably explained by the fact that students perform this activity in a routine setting and not in an examination setting. Unlike the scenario in an examination, students were aware that even if they performed at a mediocre level, there would be no repercussions.

A separate comparison of scores obtained by 44 students in their 2 nd year and 32 students in their 3 rd year of postgraduation students was also done. The number of student evaluation sheets reviewed for this analysis was greater than the number of student evaluation sheets reviewed to compare first and last JC scores. This can be spelled out by the fact that many students were still in 2 nd year when this analysis was done and the score data for their last JC, which would take place in 3 rd year, was not available. In addition, few students were asked to present at JC multiple times during the 2 nd /3 rd year of their postgraduation.

While evaluating the critical appraisal scores obtained by 2 nd - and 3 rd -year postgraduate students, it was found that although the 3 rd -year students had a mean overall score greater than the 2 nd -year students, this difference was not statistically significant. During the 1 st year of MD Pharmacology course, students at the study center attend JC once in every 2 weeks. Even though the 1 st -year students do not themselves present in JC, they listen and observe the criticism points stated by senior peers presenting at the JC, and thereby, incur substantial amount of knowledge required to critically appraise papers. By the time, they become 2 nd -year students, they are already well versed with the program and this could have led to similar overall mean scores between the 2 nd -year students (71.50 ± 10.71) and 3 rd -year students (72.34 ± 10.85). This finding suggests that attentive listening is as important as active participation in the JC. Moreover, although students are well acquainted with the process of criticism when they are in their 3 rd year, there is certainly a scope for improvement in terms of the mean overall scores.

Similar results were obtained in a study conducted by Stern et al ., in which 62 students in the internal medicine program at the New England Medical Center were asked to respond to a questionnaire, evaluate a sample article, and complete a self-assessment of competence in evaluation of research. Twenty-eight residents returned the questionnaire and the composite score for the resident's objective assessment was not significantly correlated with the postgraduate year or self-assessed critical appraisal skill.[ 11 ]

Article criticism activity provides the students with practical experience of techniques taught in research methodology workshop. However, this should be supplemented with activities that assess the improvement of designing and presenting studies, such as protocol and paper writing. Thus, critical appraisal plays a significant role in reinforcing good research practices among the new generation of physicians. Moreover, critical appraisal is an integral part of PG assessment, and although the current format of conducting JCs did not portray a clinically meaningful improvement, the authors believe that it is important to continue this activity with certain modifications suggested by students who participated in this study. Students suggested that an increase in the frequency of critical appraisal activity accompanied by the display of active participation by peers and faculty could help in the betterment of this activity. This should be brought to attention of the faculty, as students seem to be interested to learn. Critical appraisal should be a two-way teaching–learning process between the students and faculty and not a dire need for satisfying the students' eligibility criteria for postgraduate university examinations. This activity is not only for the trainee doctors but also a part of the overall faculty development program.[ 12 ]

In the present era, JCs have been used as a tool to not only teach critical appraisal skills but also to teach other necessary aspects such as research design, medical statistics, clinical epidemiology, and clinical decision-making.[ 13 , 14 ] A study conducted by Khan in 2013 suggested that success of JC program can be ensured if institutes develop a defined JC objective for the development of learning capability of students and also if they cultivate more skilled faculties.[ 15 ] A good JC is believed to facilitate relevant, meaningful scientific discussion, and evaluation of the research updates that will eventually benefit the patient care.[ 12 ]

Although there is a lot of literature emphasizing the importance of JC, there is a lack of studies that have evaluated the outcome of such activity. One such study conducted by Ibrahim et al . assessed the importance of critical appraisal as an activity in surgical trainees in Nigeria. They reported that 92.42% trainees considered the activity to be important or very important and 48% trainees stated that the activity helped in improving literature search.[ 16 ]

This study is unique since it is the first of its kind to evaluate how well students are able to critically appraise a research paper. Moreover, the study has taken into consideration the due opinions of the students as well as faculties, unlike the previous literature which has laid emphasis on only student's perception. A limitation of this study is that sample size for faculties was smaller than the students, as it was not possible to convince the distant faculty in other cities to fill the survey. Besides, there may be a variation in the manner of conduct of the critical appraisal activity in pharmacology departments across the various medical colleges in the country. Another limitation of this study was that a single assessor graded a single student during one particular JC. Nevertheless, each student presented at multiple JC and thereby came across multiple assessors. Since the articles addressed at different JC were disparate, interobserver variability was not taken into account in this study. Furthermore, the authors did not make an a priori decision on the quantum of increase in scores that would be considered educationally meaningful.

Pharmacology students and teachers acknowledge the role of critical appraisal in improving the ability to understand the crucial concepts of research methodology and research conduct. In our institute, participation in the JC activity led to an improvement in the skill of critical appraisal of published research articles among the pharmacology postgraduate students. However, this improvement was not educationally relevant. The scores obtained by final-year postgraduate students in this activity were nearly 72% indicating that there is still scope of betterment in this skill.

Financial support and sponsorship

Conflicts of interest.

There are no conflicts of interest.

Acknowledgments

We would like to acknowledge the support rendered by the entire Department of Pharmacology and Therapeutics at Seth GS Medical College.

Banner

Research Paper: A step-by-step guide: 7. Evaluating Sources

  • 1. Getting Started
  • 2. Topic Ideas
  • 3. Thesis Statement & Outline
  • 4. Appropriate Sources
  • 5. Search Techniques
  • 6. Taking Notes & Documenting Sources
  • 7. Evaluating Sources
  • 8. Citations & Plagiarism
  • 9. Writing Your Research Paper

alt=" "

Evaluation Criteria

It's very important to evaluate the materials you find to make sure they are appropriate for a research paper.  It's not enough that the information is relevant; it must also be credible.  You will want to find more than enough resources, so that you can pick and choose the best for your paper.   Here are some helpful criteria you can apply to the information you find:

C urrency :

  • When was the information published?
  • Is the source out-of-date for the topic? 
  • Are there new discoveries or important events since the publication date?

R elevancy:

  • How is the information related to your argument? 
  • Is the information too advanced or too simple? 
  • Is the audience focus appropriate for a research paper? 
  • Are there better sources elsewhere?

A uthority :

  • Who is the author? 
  • What is the author's credential in the related field? 
  • Is the publisher well-known in the field? 
  • Did the information go through the peer-review process or some kind of fact-checking?

A ccuracy :

  • Can the information be verified? 
  • Are sources cited? 
  • Is the information factual or opinion based?
  • Is the information biased? 
  • Is the information free of grammatical or spelling errors?
  • What is the motive of providing the information: to inform? to sell? to persuade? to entertain?
  • Does the author or publisher make their intentions clear? Who is the intended audience?

Evaluating Web Sources

Most web pages are not fact-checked or anything like that, so it's especially important to evaluate information you find on the web.  Many articles on websites are fine for information, and many others are distorted or made up.  Check out our media evaluation guide for tips on evaluating what you see on social media, news sites, blogs, and so on.

This three-part video series, in which university students, historians, and pro fact-checkers go head-to-head in checking out online information, is also helpful.

  • << Previous: 6. Taking Notes & Documenting Sources
  • Next: 8. Citations & Plagiarism >>
  • Last Updated: Apr 18, 2023 12:12 PM
  • URL: https://butte.libguides.com/ResearchPaper
  • Your Science & Health Librarians
  • How To Find Articles with Databases
  • Video Learning
  • Artificial Intelligence Tools
  • Industry Reports
  • How To Evaluate Articles
  • Search Tips, General
  • Develop a Research Question
  • How To Read A Scientific Paper
  • How To Interpret Data
  • How To Write A Scientific Paper
  • Teaching Materials
  • Systematic & Evideced-Based Reviews
  • Get More Help
  • New Books, STEM areas

Useful Sources

  • How to Critically Evaluate the Quality of a Research Article?
  • How to Evaluate Journal Articles

Evaluating Articles

Not all articles are created equally. Evaluating sources for relevancy and usefulness is one the most important steps in research. This helps researchers in the STEM disciplines to gather the information they need. There are several tools to use when evaluating an article. 

  • Purposes can include persuasion, informing, or proving something to the reader. Depending on the topic of your paper, you will want to evaluate the article's purpose to support your own position. 

Publication

  • Most sources for college papers should come from scholarly journals. Scholarly journals are journals that are peer-reviewed before publication. This means that experts in the field read and approve the material in the article before it is published.
  • The date of publication is especially relevant to those in the STEM disciplines. Research in science, technology, engineering, and mathematics moves very quickly, so have an article that was published recently will be more useful. 
  • Many universities operate their own press, Indiana University included! Books and articles that are published by a reputable institution are generally accepted as valuable information. Other notable presses included Harvard University Press and The MIT Press. 
  • Good sources of information come from experts in the field. This is called authority. Many times these individuals will be employed at research institutions such as universities, labs, or founded associations. 
  • Many times, authors with authority have written more than one article about a topic in their field. Not only does this add support to their reputation, but can also be a great source for more articles. 
  • Citation is a great indicator to the effectiveness of the article. If other experts are citing the article, it is a good sign this source is trusted and relevant. 

Bibliography

  • Scholarly works will always contain a bibliography of the sources used. Trusted articles will have sources that are also scholarly in nature and authored by individuals with authority in their field. 
  • Much like evaluating the publication of an article, the bibliography of the source should also contain sources that are up-to-date. 
  • << Previous: Industry Reports
  • Next: Search Tips, General >>
  • Last Updated: Sep 12, 2024 2:44 PM
  • URL: https://guides.libraries.indiana.edu/STEM

Social media

Additional resources, featured databases.

  • OneSearch@IU
  • Google Scholar
  • HathiTrust Digital Library

IU Libraries

  • Our Departments
  • Intranet SharePoint (Staff)

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings
  • My Bibliography
  • Collections
  • Citation manager

Save citation to file

Email citation, add to collections.

  • Create a new collection
  • Add to an existing collection

Add to My Bibliography

Your saved search, create a file for external citation management software, your rss feed.

  • Search in PubMed
  • Search in NLM Catalog
  • Add to Search

A simplified approach to critically appraising research evidence

Affiliation.

  • 1 School of Health and Life Sciences, Teesside University, Middlesbrough, England.
  • PMID: 33660465
  • DOI: 10.7748/nr.2021.e1760

Background Evidence-based practice is embedded in all aspects of nursing and care. Understanding research evidence and being able to identify the strengths, weaknesses and limitations of published primary research is an essential skill of the evidence-based practitioner. However, it can be daunting and seem overly complex.

Aim: To provide a single framework that researchers can use when reading, understanding and critically assessing published research.

Discussion: To make sense of published research papers, it is helpful to understand some key concepts and how they relate to either quantitative or qualitative designs. Internal and external validity, reliability and trustworthiness are discussed. An illustration of how to apply these concepts in a practical way using a standardised framework to systematically assess a paper is provided.

Conclusion: The ability to understand and evaluate research builds strong evidence-based practitioners, who are essential to nursing practice.

Implications for practice: This framework should help readers to identify the strengths, potential weaknesses and limitations of a paper to judge its quality and potential usefulness.

Keywords: literature review; qualitative research; quantitative research; research; systematic review.

©2021 RCN Publishing Company Ltd. All rights reserved. Not to be copied, transmitted or recorded in any way, in whole or part, without prior permission of the publishers.

PubMed Disclaimer

Conflict of interest statement

None declared

Similar articles

  • The Effectiveness of Integrated Care Pathways for Adults and Children in Health Care Settings: A Systematic Review. Allen D, Gillen E, Rixson L. Allen D, et al. JBI Libr Syst Rev. 2009;7(3):80-129. doi: 10.11124/01938924-200907030-00001. JBI Libr Syst Rev. 2009. PMID: 27820426
  • The future of Cochrane Neonatal. Soll RF, Ovelman C, McGuire W. Soll RF, et al. Early Hum Dev. 2020 Nov;150:105191. doi: 10.1016/j.earlhumdev.2020.105191. Epub 2020 Sep 12. Early Hum Dev. 2020. PMID: 33036834
  • Promoting and supporting self-management for adults living in the community with physical chronic illness: A systematic review of the effectiveness and meaningfulness of the patient-practitioner encounter. Rees S, Williams A. Rees S, et al. JBI Libr Syst Rev. 2009;7(13):492-582. doi: 10.11124/01938924-200907130-00001. JBI Libr Syst Rev. 2009. PMID: 27819974
  • Health professionals' experience of teamwork education in acute hospital settings: a systematic review of qualitative literature. Eddy K, Jordan Z, Stephenson M. Eddy K, et al. JBI Database System Rev Implement Rep. 2016 Apr;14(4):96-137. doi: 10.11124/JBISRIR-2016-1843. JBI Database System Rev Implement Rep. 2016. PMID: 27532314 Review.
  • The nursing contribution to qualitative research in palliative care 1990-1999: a critical evaluation. Bailey C, Froggatt K, Field D, Krishnasamy M. Bailey C, et al. J Adv Nurs. 2002 Oct;40(1):48-60. doi: 10.1046/j.1365-2648.2002.02339.x. J Adv Nurs. 2002. PMID: 12230529 Review.
  • Special Collection Editorial: The digital movement in nursing. Cronin C. Cronin C. J Res Nurs. 2022 Aug;27(5):411-420. doi: 10.1177/17449871221117437. Epub 2022 Sep 17. J Res Nurs. 2022. PMID: 36131703 Free PMC article. No abstract available.
  • Search in MeSH

LinkOut - more resources

Other literature sources.

  • scite Smart Citations
  • Citation Manager

NCBI Literature Resources

MeSH PMC Bookshelf Disclaimer

The PubMed wordmark and PubMed logo are registered trademarks of the U.S. Department of Health and Human Services (HHS). Unauthorized use of these marks is strictly prohibited.

  • Writing Rules
  • Running Head & Page numbersÂ
  • Using Quotations
  • Citing Sources
  • Reference List
  • General Reference List Principles
  • Structure of the Report

Introduction

  • References & Appendices
  • Unpacking the Assignment Topic
  • Planning and Structuring the Assignment
  • Writing the Assignment
  • Writing Concisely
  • Developing Arguments

Critically Evaluating Research

  • Editing the Assignment
  • Writing in the Third Person
  • Directive Words
  • Before You Submit
  • Cover Sheet & Title Page
  • Academic Integrity
  • Marking Criteria
  • Word Limit Rules
  • Submitting Your WorkÂ
  • Writing Effective E-mails
  • Writing Concisely Exercise
  • About Redbook

Some research reports or assessments will require you critically evaluate a journal article or piece of research. Below is a guide with examples of how to critically evaluate research and how to communicate your ideas in writing.

To develop the skill of being able to critically evaluate, when reading research articles in psychology read with an open mind and be active when reading. Ask questions as you go and see if the answers are provided. Initially skim through the article to gain an overview of the problem, the design, methods, and conclusions. Then read for details and consider the questions provided below for each section of a journal article.

  • Did the title describe the study?
  • Did the key words of the title serve as key elements of the article?
  • Was the title concise, i.e., free of distracting or extraneous phrases?
  • Was the abstract concise and to the point?
  • Did the abstract summarise the study’s purpose/research problem, the independent and dependent variables under study, methods, main findings, and conclusions?
  • Did the abstract provide you with sufficient information to determine what the study is about and whether you would be interested in reading the entire article?
  • Was the research problem clearly identified?
  • Is the problem significant enough to warrant the study that was conducted?
  • Did the authors present an appropriate theoretical rationale for the study?
  • Is the literature review informative and comprehensive or are there gaps?
  • Are the variables adequately explained and operationalised?
  • Are hypotheses and research questions clearly stated? Are they directional? Do the author’s hypotheses and/or research questions seem logical in light of the conceptual framework and research problem?
  • Overall, does the literature review lead logically into the Method section?
  • Is the sample clearly described, in terms of size, relevant characteristics (gender, age, SES, etc), selection and assignment procedures, and whether any inducements were used to solicit subjects (payment, subject credit, free therapy, etc)?
  • What population do the subjects represent (external validity)?
  • Are there sufficient subjects to produce adequate power (statistical validity)?
  • Have the variables and measurement techniques been clearly operationalised?
  • Do the measures/instruments seem appropriate as measures of the variables under study (construct validity)?
  • Have the authors included sufficient information about the psychometric properties (eg. reliability and validity) of the instruments?
  • Are the materials used in conducting the study or in collecting data clearly described?
  • Are the study’s scientific procedures thoroughly described in chronological order?
  • Is the design of the study identified (or made evident)?
  • Do the design and procedures seem appropriate in light of the research problem, conceptual framework, and research questions/hypotheses?
  • Are there other factors that might explain the differences between groups (internal validity)?
  • Were subjects randomly assigned to groups so there was no systematic bias in favour of one group? Was there a differential drop-out rate from groups so that bias was introduced (internal validity and attrition)?
  • Were all the necessary control groups used? Were participants in each group treated identically except for the administration of the independent variable?
  • Were steps taken to prevent subject bias and/or experimenter bias, eg, blind or double blind procedures?
  • Were steps taken to control for other possible confounds such as regression to the mean, history effects, order effects, etc (internal validity)?
  • Were ethical considerations adhered to, eg, debriefing, anonymity, informed consent, voluntary participation?
  • Overall, does the method section provide sufficient information to replicate the study?
  • Are the findings complete, clearly presented, comprehensible, and well organised?
  • Are data coding and analysis appropriate in light of the study’s design and hypotheses? Are the statistics reported correctly and fully, eg. are degrees of freedom and p values given?
  • Have the assumptions of the statistical analyses been met, eg. does one group have very different variance to the others?
  • Are salient results connected directly to hypotheses? Are there superfluous results presented that are not relevant to the hypotheses or research question?
  • Are tables and figures clearly labelled? Well-organised? Necessary (non-duplicative of text)?
  • If a significant result is obtained, consider effect size. Is the finding meaningful? If a non-significant result is found, could low power be an issue? Were there sufficient levels of the IV?
  • If necessary have appropriate post-hoc analyses been performed? Were any transformations performed; if so, were there valid reasons? Were data collapsed over any IVs; if so, were there valid reasons? If any data was eliminated, were valid reasons given?

Discussion and Conclusion

  • Are findings adequately interpreted and discussed in terms of the stated research problem, conceptual framework, and hypotheses?
  • Is the interpretation adequate? i.e., does it go too far given what was actually done or not far enough? Are non-significant findings interpreted inappropriately?
  • Is the discussion biased? Are the limitations of the study delineated?
  • Are implications for future research and/or practical application identified?
  • Are the overall conclusions warranted by the data and any limitations in the study? Are the conclusions restricted to the population under study or are they generalised too widely?
  • Is the reference list sufficiently specific to the topic under investigation and current?
  • Are citations used appropriately in the text?

General Evaluation

  • Is the article objective, well written and organised?
  • Does the information provided allow you to replicate the study in all its details?
  • Was the study worth doing? Does the study provide an answer to a practical or important problem? Does it have theoretical importance? Does it represent a methodological or technical advance? Does it demonstrate a previously undocumented phenomenon? Does it explore the conditions under which a phenomenon occurs?

How to turn your critical evaluation into writing

Example from a journal article.

how to evaluate a research paper

  • The Open University
  • Accessibility hub
  • Guest user / Sign out
  • Study with The Open University

My OpenLearn Profile

Personalise your OpenLearn profile, save your favourite content and get recognition for your learning

About this free course

Become an ou student, download this course, share this free course.

Succeeding in postgraduate study

Start this free course now. Just create an account and sign in. Enrol and complete the course for a free statement of participation or digital badge if available.

1 Important points to consider when critically evaluating published research papers

Simple review articles (also referred to as ‘narrative’ or ‘selective’ reviews), systematic reviews and meta-analyses provide rapid overviews and ‘snapshots’ of progress made within a field, summarising a given topic or research area. They can serve as useful guides, or as current and comprehensive ‘sources’ of information, and can act as a point of reference to relevant primary research studies within a given scientific area. Narrative or systematic reviews are often used as a first step towards a more detailed investigation of a topic or a specific enquiry (a hypothesis or research question), or to establish critical awareness of a rapidly-moving field (you will be required to demonstrate this as part of an assignment, an essay or a dissertation at postgraduate level).

The majority of primary ‘empirical’ research papers essentially follow the same structure (abbreviated here as IMRAD). There is a section on Introduction, followed by the Methods, then the Results, which includes figures and tables showing data described in the paper, and a Discussion. The paper typically ends with a Conclusion, and References and Acknowledgements sections.

The Title of the paper provides a concise first impression. The Abstract follows the basic structure of the extended article. It provides an ‘accessible’ and concise summary of the aims, methods, results and conclusions. The Introduction provides useful background information and context, and typically outlines the aims and objectives of the study. The Abstract can serve as a useful summary of the paper, presenting the purpose, scope and major findings. However, simply reading the abstract alone is not a substitute for critically reading the whole article. To really get a good understanding and to be able to critically evaluate a research study, it is necessary to read on.

While most research papers follow the above format, variations do exist. For example, the results and discussion sections may be combined. In some journals the materials and methods may follow the discussion, and in two of the most widely read journals, Science and Nature, the format does vary from the above due to restrictions on the length of articles. In addition, there may be supporting documents that accompany a paper, including supplementary materials such as supporting data, tables, figures, videos and so on. There may also be commentaries or editorials associated with a topical research paper, which provide an overview or critique of the study being presented.

Box 1 Key questions to ask when appraising a research paper

  • Is the study’s research question relevant?
  • Does the study add anything new to current knowledge and understanding?
  • Does the study test a stated hypothesis?
  • Is the design of the study appropriate to the research question?
  • Do the study methods address key potential sources of bias?
  • Were suitable ‘controls’ included in the study?
  • Were the statistical analyses appropriate and applied correctly?
  • Is there a clear statement of findings?
  • Does the data support the authors’ conclusions?
  • Are there any conflicts of interest or ethical concerns?

There are various strategies used in reading a scientific research paper, and one of these is to start with the title and the abstract, then look at the figures and tables, and move on to the introduction, before turning to the results and discussion, and finally, interrogating the methods.

Another strategy (outlined below) is to begin with the abstract and then the discussion, take a look at the methods, and then the results section (including any relevant tables and figures), before moving on to look more closely at the discussion and, finally, the conclusion. You should choose a strategy that works best for you. However, asking the ‘right’ questions is a central feature of critical appraisal, as with any enquiry, so where should you begin? Here are some critical questions to consider when evaluating a research paper.

Look at the Abstract and then the Discussion : Are these accessible and of general relevance or are they detailed, with far-reaching conclusions? Is it clear why the study was undertaken? Why are the conclusions important? Does the study add anything new to current knowledge and understanding? The reasons why a particular study design or statistical method were chosen should also be clear from reading a research paper. What is the research question being asked? Does the study test a stated hypothesis? Is the design of the study appropriate to the research question? Have the authors considered the limitations of their study and have they discussed these in context?

Take a look at the Methods : Were there any practical difficulties that could have compromised the study or its implementation? Were these considered in the protocol? Were there any missing values and, if so, was the number of missing values too large to permit meaningful analysis? Was the number of samples (cases or participants) too small to establish meaningful significance? Do the study methods address key potential sources of bias? Were suitable ‘controls’ included in the study? If controls are missing or not appropriate to the study design, we cannot be confident that the results really show what is happening in an experiment. Were the statistical analyses appropriate and applied correctly? Do the authors point out the limitations of methods or tests used? Were the methods referenced and described in sufficient detail for others to repeat or extend the study?

Take a look at the Results section and relevant tables and figures : Is there a clear statement of findings? Were the results expected? Do they make sense? What data supports them? Do the tables and figures clearly describe the data (highlighting trends etc.)? Try to distinguish between what the data show and what the authors say they show (i.e. their interpretation).

Moving on to look in greater depth at the Discussion and Conclusion : Are the results discussed in relation to similar (previous) studies? Do the authors indulge in excessive speculation? Are limitations of the study adequately addressed? Were the objectives of the study met and the hypothesis supported or refuted (and is a clear explanation provided)? Does the data support the authors’ conclusions? Maybe there is only one experiment to support a point. More often, several different experiments or approaches combine to support a particular conclusion. A rule of thumb here is that if multiple approaches and multiple lines of evidence from different directions are presented, and all point to the same conclusion, then the conclusions are more credible. But do question all assumptions. Identify any implicit or hidden assumptions that the authors may have used when interpreting their data. Be wary of data that is mixed up with interpretation and speculation! Remember, just because it is published, does not mean that it is right.

O ther points you should consider when evaluating a research paper : Are there any financial, ethical or other conflicts of interest associated with the study, its authors and sponsors? Are there ethical concerns with the study itself? Looking at the references, consider if the authors have preferentially cited their own previous publications (i.e. needlessly), and whether the list of references are recent (ensuring that the analysis is up-to-date). Finally, from a practical perspective, you should move beyond the text of a research paper, talk to your peers about it, consult available commentaries, online links to references and other external sources to help clarify any aspects you don’t understand.

The above can be taken as a general guide to help you begin to critically evaluate a scientific research paper, but only in the broadest sense. Do bear in mind that the way that research evidence is critiqued will also differ slightly according to the type of study being appraised, whether observational or experimental, and each study will have additional aspects that would need to be evaluated separately. For criteria recommended for the evaluation of qualitative research papers, see the article by Mildred Blaxter (1996), available online. Details are in the References.

Activity 1 Critical appraisal of a scientific research paper

A critical appraisal checklist, which you can download via the link below, can act as a useful tool to help you to interrogate research papers. The checklist is divided into four sections, broadly covering:

  • some general aspects
  • research design and methodology
  • the results
  • discussion, conclusion and references.

Science perspective – critical appraisal checklist [ Tip: hold Ctrl and click a link to open it in a new tab. ( Hide tip ) ]

  • Identify and obtain a research article based on a topic of your own choosing, using a search engine such as Google Scholar or PubMed (for example).
  • The selection criteria for your target paper are as follows: the article must be an open access primary research paper (not a review) containing empirical data, published in the last 2–3 years, and preferably no more than 5–6 pages in length.
  • Critically evaluate the research paper using the checklist provided, making notes on the key points and your overall impression.

Critical appraisal checklists are useful tools to help assess the quality of a study. Assessment of various factors, including the importance of the research question, the design and methodology of a study, the validity of the results and their usefulness (application or relevance), the legitimacy of the conclusions, and any potential conflicts of interest, are an important part of the critical appraisal process. Limitations and further improvements can then be considered.

Previous

  • Library Home
  • Research Guides

Writing a Research Paper

  • Evaluate Sources

Library Research Guide

  • Choose Your Topic
  • Organize Your Information
  • Draft Your Paper
  • Revise, Review, Refine

How Will This Help Me?

Evaluating your sources will help you:

  • Determine the credibility of information
  • Rule out questionable information
  • Check for bias in your sources

In general, websites are hosted in domains that tell you what type of site it is.

  • .com = commercial
  • .net = network provider
  • .org = organization
  • .edu = education
  • .mil = military
  • .gov = U.S. government

Commercial sites want to persuade you to buy something, and organizations may want to persuade you to see an issue from a particular viewpoint. 

Useful information can be found on all kinds of sites, but you must consider carefully whether the source is useful for your purpose and for your audience.

Content Farms

Content farms are websites that exist to host ads. They post about popular web searches to try to drive traffic to their sites. They are rarely good sources for research.

  • Web’s “Content Farms” Grow Audiences For Ads This article by Zoe Chace at National Public Radio describes the ways How To sites try to drive more traffic to their sites to see the ads they host.

Fact Checking

Fact checking can help you verify the reliability of a source. The following sites may not have all the answers, but they can help you look into the sources for statements made in U.S. politics.

  • FactCheck.org This site monitors the accuracy of statements made in speeches, debates, interviews, and more and links to sources so readers can see the information for themselves. The site is a project of the Annenberg Public Policy Center of the University of Pennsylvania.
  • PolitiFact This resource evaluates the accuracy of statements made by elected officials, lobbyists, and special interest groups and provides sources for their evaluations. PolitiFact is currently run by the nonprofit Poynter Institute for Media Studies.

Evaluate Sources With the Big 5 Criteria

The Big 5 Criteria can help you evaluate your sources for credibility:

  • Currency: Check the publication date and determine whether it is sufficiently current for your topic.
  • Coverage (relevance): Consider whether the source is relevant to your research and whether it covers the topic adequately for your needs.
  • Authority: Discover the credentials of the authors of the source and determine their level of expertise and knowledge about the subject.
  • Accuracy: Consider whether the source presents accurate information and whether you can verify that information. 
  • Objectivity (purpose): Think about the author's purpose in creating the source and consider how that affects its usefulness to your research. 

Evaluate Sources With the CRAAP Test

Another way to evaluate your sources is the CRAAP Test, which means evaluating the following qualities of your sources:

This video (2:17) from Western Libraries explains the CRAAP Test. 

Video transcript

Evaluating Sources ( Western Libraries ) CC BY-NC-ND 3.0

Evaluate Websites

Evaluating websites follows the same process as for other sources, but finding the information you need to make an assessment can be more challenging with websites. The following guidelines can help you decide if a website is a good choice for a source for your paper. 

  • Currency . A useful site is updated regularly and lets visitors know when content was published on the site. Can you tell when the site was last updated? Can you see when the content you need was added? Does the site show signs of not being maintained (broken links, out-of-date information, etc.)?
  • Relevance . Think about the target audience for the site. Is it appropriate for you or your paper's audience?
  • Authority . Look for an About Us link or something similar to learn about the site's creator. The more you know about the credentials and mission of a site's creators, as well as their sources of information, the better idea you will have about the site's quality. 
  • Accuracy. Does the site present references or links to the sources of information it presents? Can you locate these sources so that you can read and interpret the information yourself?
  • Purpose. Consider the reason why the site was created. Can you detect any bias? Does the site use emotional language? Is the site trying to persuade you about something? 

Identify Political Perspective

News outlets, think tanks, organizations, and individual authors can present information from a particular political perspective. Consider this fact to help determine whether sources are useful for your paper. 

how to evaluate a research paper

Check a news outlet's website, usually under About Us or Contact Us , for information about their reporters and authors. For example, USA Today has the USA Today Reporter Index , and the LA Times has an Editorial & Newsroom Contacts . Reading a profile or bio for a reporter or looking at other articles by the author may tell you whether that person favors a particular viewpoint. 

If a particular organization is mentioned in an article, learn more about the organization to identify potential biases. Think tanks and other associations usually exist for a reason. Searching news articles about the organization can help you determine their political leaning. 

Bias is not always bad, but you must be aware of it. Knowing the perspective of a source helps contextualize the information presented. 

  • << Previous: Databases
  • Next: Organize Your Information >>
  • Last Updated: Sep 4, 2024 2:42 PM
  • URL: https://guides.lib.k-state.edu/writingresearchpaper

K-State Libraries

1117 Mid-Campus Drive North, Manhattan, KS 66506

785-532-3014 | [email protected]

  • Statements and Disclosures
  • Accessibility
  • © Kansas State University

Have a language expert improve your writing

Run a free plagiarism check in 10 minutes, generate accurate citations for free.

  • Knowledge Base

Methodology

  • How to Write a Literature Review | Guide, Examples, & Templates

How to Write a Literature Review | Guide, Examples, & Templates

Published on January 2, 2023 by Shona McCombes . Revised on September 11, 2023.

What is a literature review? A literature review is a survey of scholarly sources on a specific topic. It provides an overview of current knowledge, allowing you to identify relevant theories, methods, and gaps in the existing research that you can later apply to your paper, thesis, or dissertation topic .

There are five key steps to writing a literature review:

  • Search for relevant literature
  • Evaluate sources
  • Identify themes, debates, and gaps
  • Outline the structure
  • Write your literature review

A good literature review doesn’t just summarize sources—it analyzes, synthesizes , and critically evaluates to give a clear picture of the state of knowledge on the subject.

Instantly correct all language mistakes in your text

Upload your document to correct all your mistakes in minutes

upload-your-document-ai-proofreader

Table of contents

What is the purpose of a literature review, examples of literature reviews, step 1 – search for relevant literature, step 2 – evaluate and select sources, step 3 – identify themes, debates, and gaps, step 4 – outline your literature review’s structure, step 5 – write your literature review, free lecture slides, other interesting articles, frequently asked questions, introduction.

  • Quick Run-through
  • Step 1 & 2

When you write a thesis , dissertation , or research paper , you will likely have to conduct a literature review to situate your research within existing knowledge. The literature review gives you a chance to:

  • Demonstrate your familiarity with the topic and its scholarly context
  • Develop a theoretical framework and methodology for your research
  • Position your work in relation to other researchers and theorists
  • Show how your research addresses a gap or contributes to a debate
  • Evaluate the current state of research and demonstrate your knowledge of the scholarly debates around your topic.

Writing literature reviews is a particularly important skill if you want to apply for graduate school or pursue a career in research. We’ve written a step-by-step guide that you can follow below.

Literature review guide

Don't submit your assignments before you do this

The academic proofreading tool has been trained on 1000s of academic texts. Making it the most accurate and reliable proofreading tool for students. Free citation check included.

how to evaluate a research paper

Try for free

Writing literature reviews can be quite challenging! A good starting point could be to look at some examples, depending on what kind of literature review you’d like to write.

  • Example literature review #1: “Why Do People Migrate? A Review of the Theoretical Literature” ( Theoretical literature review about the development of economic migration theory from the 1950s to today.)
  • Example literature review #2: “Literature review as a research methodology: An overview and guidelines” ( Methodological literature review about interdisciplinary knowledge acquisition and production.)
  • Example literature review #3: “The Use of Technology in English Language Learning: A Literature Review” ( Thematic literature review about the effects of technology on language acquisition.)
  • Example literature review #4: “Learners’ Listening Comprehension Difficulties in English Language Learning: A Literature Review” ( Chronological literature review about how the concept of listening skills has changed over time.)

You can also check out our templates with literature review examples and sample outlines at the links below.

Download Word doc Download Google doc

Before you begin searching for literature, you need a clearly defined topic .

If you are writing the literature review section of a dissertation or research paper, you will search for literature related to your research problem and questions .

Make a list of keywords

Start by creating a list of keywords related to your research question. Include each of the key concepts or variables you’re interested in, and list any synonyms and related terms. You can add to this list as you discover new keywords in the process of your literature search.

  • Social media, Facebook, Instagram, Twitter, Snapchat, TikTok
  • Body image, self-perception, self-esteem, mental health
  • Generation Z, teenagers, adolescents, youth

Search for relevant sources

Use your keywords to begin searching for sources. Some useful databases to search for journals and articles include:

  • Your university’s library catalogue
  • Google Scholar
  • Project Muse (humanities and social sciences)
  • Medline (life sciences and biomedicine)
  • EconLit (economics)
  • Inspec (physics, engineering and computer science)

You can also use boolean operators to help narrow down your search.

Make sure to read the abstract to find out whether an article is relevant to your question. When you find a useful book or article, you can check the bibliography to find other relevant sources.

You likely won’t be able to read absolutely everything that has been written on your topic, so it will be necessary to evaluate which sources are most relevant to your research question.

For each publication, ask yourself:

  • What question or problem is the author addressing?
  • What are the key concepts and how are they defined?
  • What are the key theories, models, and methods?
  • Does the research use established frameworks or take an innovative approach?
  • What are the results and conclusions of the study?
  • How does the publication relate to other literature in the field? Does it confirm, add to, or challenge established knowledge?
  • What are the strengths and weaknesses of the research?

Make sure the sources you use are credible , and make sure you read any landmark studies and major theories in your field of research.

You can use our template to summarize and evaluate sources you’re thinking about using. Click on either button below to download.

Take notes and cite your sources

As you read, you should also begin the writing process. Take notes that you can later incorporate into the text of your literature review.

It is important to keep track of your sources with citations to avoid plagiarism . It can be helpful to make an annotated bibliography , where you compile full citation information and write a paragraph of summary and analysis for each source. This helps you remember what you read and saves time later in the process.

Here's why students love Scribbr's proofreading services

Discover proofreading & editing

To begin organizing your literature review’s argument and structure, be sure you understand the connections and relationships between the sources you’ve read. Based on your reading and notes, you can look for:

  • Trends and patterns (in theory, method or results): do certain approaches become more or less popular over time?
  • Themes: what questions or concepts recur across the literature?
  • Debates, conflicts and contradictions: where do sources disagree?
  • Pivotal publications: are there any influential theories or studies that changed the direction of the field?
  • Gaps: what is missing from the literature? Are there weaknesses that need to be addressed?

This step will help you work out the structure of your literature review and (if applicable) show how your own research will contribute to existing knowledge.

  • Most research has focused on young women.
  • There is an increasing interest in the visual aspects of social media.
  • But there is still a lack of robust research on highly visual platforms like Instagram and Snapchat—this is a gap that you could address in your own research.

There are various approaches to organizing the body of a literature review. Depending on the length of your literature review, you can combine several of these strategies (for example, your overall structure might be thematic, but each theme is discussed chronologically).

Chronological

The simplest approach is to trace the development of the topic over time. However, if you choose this strategy, be careful to avoid simply listing and summarizing sources in order.

Try to analyze patterns, turning points and key debates that have shaped the direction of the field. Give your interpretation of how and why certain developments occurred.

If you have found some recurring central themes, you can organize your literature review into subsections that address different aspects of the topic.

For example, if you are reviewing literature about inequalities in migrant health outcomes, key themes might include healthcare policy, language barriers, cultural attitudes, legal status, and economic access.

Methodological

If you draw your sources from different disciplines or fields that use a variety of research methods , you might want to compare the results and conclusions that emerge from different approaches. For example:

  • Look at what results have emerged in qualitative versus quantitative research
  • Discuss how the topic has been approached by empirical versus theoretical scholarship
  • Divide the literature into sociological, historical, and cultural sources

Theoretical

A literature review is often the foundation for a theoretical framework . You can use it to discuss various theories, models, and definitions of key concepts.

You might argue for the relevance of a specific theoretical approach, or combine various theoretical concepts to create a framework for your research.

Like any other academic text , your literature review should have an introduction , a main body, and a conclusion . What you include in each depends on the objective of your literature review.

The introduction should clearly establish the focus and purpose of the literature review.

Depending on the length of your literature review, you might want to divide the body into subsections. You can use a subheading for each theme, time period, or methodological approach.

As you write, you can follow these tips:

  • Summarize and synthesize: give an overview of the main points of each source and combine them into a coherent whole
  • Analyze and interpret: don’t just paraphrase other researchers — add your own interpretations where possible, discussing the significance of findings in relation to the literature as a whole
  • Critically evaluate: mention the strengths and weaknesses of your sources
  • Write in well-structured paragraphs: use transition words and topic sentences to draw connections, comparisons and contrasts

In the conclusion, you should summarize the key findings you have taken from the literature and emphasize their significance.

When you’ve finished writing and revising your literature review, don’t forget to proofread thoroughly before submitting. Not a language expert? Check out Scribbr’s professional proofreading services !

This article has been adapted into lecture slides that you can use to teach your students about writing a literature review.

Scribbr slides are free to use, customize, and distribute for educational purposes.

Open Google Slides Download PowerPoint

If you want to know more about the research process , methodology , research bias , or statistics , make sure to check out some of our other articles with explanations and examples.

  • Sampling methods
  • Simple random sampling
  • Stratified sampling
  • Cluster sampling
  • Likert scales
  • Reproducibility

 Statistics

  • Null hypothesis
  • Statistical power
  • Probability distribution
  • Effect size
  • Poisson distribution

Research bias

  • Optimism bias
  • Cognitive bias
  • Implicit bias
  • Hawthorne effect
  • Anchoring bias
  • Explicit bias

A literature review is a survey of scholarly sources (such as books, journal articles, and theses) related to a specific topic or research question .

It is often written as part of a thesis, dissertation , or research paper , in order to situate your work in relation to existing knowledge.

There are several reasons to conduct a literature review at the beginning of a research project:

  • To familiarize yourself with the current state of knowledge on your topic
  • To ensure that you’re not just repeating what others have already done
  • To identify gaps in knowledge and unresolved problems that your research can address
  • To develop your theoretical framework and methodology
  • To provide an overview of the key findings and debates on the topic

Writing the literature review shows your reader how your work relates to existing research and what new insights it will contribute.

The literature review usually comes near the beginning of your thesis or dissertation . After the introduction , it grounds your research in a scholarly field and leads directly to your theoretical framework or methodology .

A literature review is a survey of credible sources on a topic, often used in dissertations , theses, and research papers . Literature reviews give an overview of knowledge on a subject, helping you identify relevant theories and methods, as well as gaps in existing research. Literature reviews are set up similarly to other  academic texts , with an introduction , a main body, and a conclusion .

An  annotated bibliography is a list of  source references that has a short description (called an annotation ) for each of the sources. It is often assigned as part of the research process for a  paper .  

Cite this Scribbr article

If you want to cite this source, you can copy and paste the citation or click the “Cite this Scribbr article” button to automatically add the citation to our free Citation Generator.

McCombes, S. (2023, September 11). How to Write a Literature Review | Guide, Examples, & Templates. Scribbr. Retrieved September 27, 2024, from https://www.scribbr.com/dissertation/literature-review/

Is this article helpful?

Shona McCombes

Shona McCombes

Other students also liked, what is a theoretical framework | guide to organizing, what is a research methodology | steps & tips, how to write a research proposal | examples & templates, get unlimited documents corrected.

✔ Free APA citation check included ✔ Unlimited document corrections ✔ Specialized in correcting academic texts

Shapiro Library

Evaluating Sources

Source evaluation rubric.

This rubric can help you determine if a source is a "good" source; one that is reliable to use in your research or paper. It can help you weed out "bad" sources and defend your "good" sources to your instructor.

  • C.R.A.A.P.O Source Evaluation Rubric NOTE: You may need to download the PDF to fill in the form fields electronically.

How to use this Rubric:

  • Enter information about the source at the top of the page, i.e. title, url, author, dates
  • For each line, starting with Currency, read each box from left to right and choose the one that matches your source the best
  • Enter the column number, 1-4 that corresponds to the box that matches your source the best in the right hand column
  • Once every line has a number, tally the numbers in the right hand column and write the score at the bottom of the page

The score you tally is out of 24 total points. You must determine what is the lowest score you will accept.

An acceptable score for a source to be used in a research paper for college is between 20 to 24.

C.R.A.A.P.O Source Evaluation Rubric
  ONE (1) TWO (2) THREE (3) FOUR (4) SCORE

No publish date listed -- or No revisions in the last eighteen months*.

No updates in the past year*.

Updated in the last six months*.

Publish date included -- or Updated in the last three months*.

 

Content is unrelated to your topic -- and / or level is too simple / too advanced.

Content is either related and incorrect level – Or unrelated and correct level.

Content is related -- And at correct level -- But you are not comfortable using the source in your research.

Content is related -- And at correct level – And you are comfortable using the source in your research.

 

No author is listed -- and No contact info provided.

No author is listed – but includes contact information.

Author is listed without credentials -- You are unsure if the author is the creator of the material.

Author is listed with credentials -- Is the originator of the information-- Contact information provided

 

Information is not verifiable -- Resources not documented.

Some resources are not documented -- some links do not work*.

Most resources are documented -- links work*.

Well organized source -- Resources documented -- links work*

 

A lot of advertising makes the content unclear.

Purpose is to sell, entertain, or persuade -- Source contains a lot of advertising and bias.

Purpose is to inform and teach-- Contains some advertising -- Minimal bias.

Purpose is to inform and teach -- Contains little advertising -- Bias free.

 

It is unclear what institution published and support the source.

It is unclear if author has any connection with a larger institution -- Source is .com, .org, or other generic domain type*

Source is supported by larger institution -- But some bias is apparent.

It is clear the source was published and supported by a reputable institution -- Bias free.

 

       
(of possible 24)

 

The C.R.A.A.P. Test was created by Sarah Blakeslee (University of California at Chico, Meriam Library). With her permission, this content was based off her  original text  with some modification.

  • << Previous: Purpose & Objectivity
  • Next: What About Websites? >>

Purdue Online Writing Lab Purdue OWL® College of Liberal Arts

Evaluating Sources: General Guidelines

OWL logo

Welcome to the Purdue OWL

This page is brought to you by the OWL at Purdue University. When printing this page, you must include the entire legal notice.

Copyright ©1995-2018 by The Writing Lab & The OWL at Purdue and Purdue University. All rights reserved. This material may not be published, reproduced, broadcast, rewritten, or redistributed without permission. Use of this site constitutes acceptance of our terms and conditions of fair use.

Once you have an idea of the types of sources you need for your research, you can spend time evaluating individual sources. If a bibliographic citation seems promising, it’s a good idea to spend a bit more time with the source before you determine its credibility. Below are some questions to ask and things to consider as you read through a source. 

Find Out What You Can about the Author

One of the first steps in evaluating a source is to locate more information about the author. Sometimes simply typing an author’s name into a search engine will give you an initial springboard for information. Finding the author’s educational background and areas of expertise will help determine whether the author has experience in what they’re writing about. You should also examine whether the author has other publications and if they are with well-known publishers or organizations.

Read the Introduction / Preface

Begin by reading the Introduction or the Preface—What does the author want to accomplish? Browse through the Table of Contents and the Index. This will give you an overview of the source. Is your topic covered in enough depth to be helpful? If you don't find your topic discussed, try searching for some synonyms in the Index.

If your source does not contain any of these elements, consider reading the first few paragraphs of the source and determining whether it includes enough information on your topic for it to be relevant.

Determine the Intended Audience

Consider the tone, style, vocabulary, level of information, and assumptions the author makes about the reader. Are they appropriate for your needs? Remember that scholarly sources often have a very particular audience in mind, and popular sources are written for a more general audience. However, some scholarly sources may be too dense for your particular research needs, so you may need to turn to sources with a more general audience in mind.

Determine whether the Information is Fact, Opinion, or Propaganda

Information can usually be divided into three categories: fact, opinion, and propaganda. Facts are objective, while opinions and propaganda are subjective. A fact is something that is known to be true. An opinion gives the thoughts of a particular individual or group. Propaganda is the (usually biased) spreading of information for a specific person, group, event, or cause. Propaganda often relies on slogans or emotionally-charged images to influence an audience. It can also involve the selective reporting of true information in order to deceive an audience.

  • Fact: The Purdue OWL was launched in 1994.
  • Opinion: The Purdue OWL is the best website for writing help.
  • Propaganda: Some students have gone on to lives of crime after using sites that compete with the Purdue OWL. The Purdue OWL is clearly the only safe choice for student writers.

The last example above uses facts in a bad-faith way to take advantage of the audience's fear. Even if the individual claim is true, the way it is presented helps the author tell a much larger lie. In this case, the lie is that there is a link between the websites students visit for writing help and their later susceptibility to criminal lifstyles. Of course, there is no such link. Thus, when examining sources for possible propaganda, be aware that sometimes groups may deploy pieces of true information in deceptive ways.

Note also that the difference between an opinion and propaganda is that propaganda usually has a specific agenda attached—that is, the information in the propaganda is being spread for a certain reason or to accomplish a certain goal. If the source appears to represent an opinion, does the author offer legitimate reasons for adopting that stance? If the opinion feels one-sided, does the author acknowledge opposing viewpoints? An opinion-based source is not necessarily unreliable, but it’s important to know whether the author recognizes that their opinion is not the only opinion.

Identify the Language Used

Is the language objective or emotional? Objective language sticks to the facts, but emotional language relies on garnering an emotional response from the reader. Objective language is more commonly found in fact-based sources, while emotional language is more likely to be found in opinion-based sources and propaganda.

Evaluate the Evidence Listed

If you’re just starting your research, you might look for sources that include more general information. However, the deeper you get into your topic, the more comprehensive your research will need to be.

If you’re reading an opinion-based source, ask yourself whether there’s enough evidence to back up the opinions. If you’re reading a fact-based source, be sure that it doesn’t oversimplify the topic.

The more familiar you become with your topic, the easier it will be for you to evaluate the evidence in your sources.

Cross-Check the Information

When you verify the information in one source with information you find in another source, this is called cross-referencing or cross-checking. If the author lists specific dates or facts, can you find that same information somewhere else? Having information listed in more than one place increases its credibility.

Check the Timeliness of the Source

How timely is the source? Is the source twenty years out of date? Some information becomes dated when new research is available, but other older sources of information can still be useful and reliable fifty or a hundred years later. For example, if you are researching a scientific topic, you will want to be sure you have the most up-to-date information. However, if you are examining an historical event, you may want to find primary documents from the time of the event, thus requiring older sources.

Examine the List of References

Check for a list of references or other citations that look as if they will lead you to related material that would be good sources. If a source has a list of references, it often means that the source is well-researched and thorough.

As you continue to encounter more sources, evaluating them for credibility will become easier.

  • Prospective Students
  • Current Students
  • Faculty & Staff

Housatonic Campus Library

Housatonic student research paper award.

  • HOUSATONIC STUDENT RESEARCH PAPER AWARD
  • SUBMIT APPLICATION This link opens in a new window

Submission Deadline

 

 

 

EVALUATION COMMITTEE

, Reference/Instruction Librarian , Faculty, English Lecturer to Professor , Director of Library Services

 

Award Description

how to evaluate a research paper

In partnership with Housatonic Community College faculty and the Tutoring Center, the Housatonic Campus Library is excited to introduce the Housatonic Student Research Paper Award! This award recognizes outstanding student research, with two winning papers each receiving a prize of $200. The winners will also enjoy the honor of having their names inscribed on an award plaque displayed in the library, along with publication of their papers on the library website.

ELIGIBILITY REQUIREMENTS

  • Students must be enrolled at Housatonic Community College at least part-time during the following semesters: Summer 2024, Fall 2024, and Spring 2025.
  • The paper or project must be the student's original work and submitted for credit in a course taken during the 2024-2025 academic year.
  • All sources must be cited using a standard citation format (MLA, APA, AMA, or Chicago/Turabian).
  • Web resources cited should be of high quality.
  • All submissions must include a reflective paragraph discussing what was learned about research and writing/editing strategies, as well as how library resources and tools were utilized.

SUBMISSION GUIDELINES To successfully apply, please include the following:

  • Your name, student email address, and phone number
  • Your paper or project, submitted as a Word document
  • The course name at Housatonic for which the paper was written
  • The name of the professor who assigned and graded your paper
  • A copy of the assignment
  • A paragraph reflecting on what you learned about research and writing/editing strategies, and how you applied research resources and tools while working on your paper.

We look forward to celebrating your hard work and exceptional research efforts!

  • Next: FLYER >>
  • Last Updated: Sep 27, 2024 1:59 PM
  • URL: https://library.ctstate.edu/HCStudentResearchPaperAward

Evaluating Policy Counterfactuals: A VAR-Plus Approach

In a rich family of linearized structural macroeconomic models, the counterfactual evolution of the macro-economy under alternative policy rules is pinned down by just two objects: first, reduced-form projections with respect to a large information set; and second, the dynamic causal effects of policy shocks. In particular, no assumptions about the structural shocks affecting the economy are needed. We propose to recover these two sufficient statistics using a ``VAR-Plus'' approach, and apply it to evaluate several monetary policy counterfactuals.

We received helpful comments from Marios Angeletos, Régis Barnichon, Marco Bassetto, Anmol Bhandari, Francesco Bianchi, Ricardo Caballero, Gabriel Chodorow-Reich, Martin Eichenbaum, Simon Gilchrist, Cosmin Ilut, Giuseppe Moscarini, Mikkel Plagborg-Møller, Giorgio Primiceri, Valerie Ramey, Matt Rognlie, Ben Schumann, Ludwig Straub, Iván Werning, and seminar participants at various venues. We also thank Seungki Hong, Klodiana Istrefi, and Diego Känzig for valuable discussions, and Valeria Morales Vasquez for superb research assistance. Wolf acknowledges that this material is based upon work supported by the NSF under Grant #2314736. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the NSF, the Federal Reserve Bank of Minneapolis, or the Federal Reserve System. The views expressed herein are those of the authors and do not necessarily reflect the views of the National Bureau of Economic Research.

MARC RIS BibTeΧ

Download Citation Data

  • Replication files

More from NBER

In addition to working papers , the NBER disseminates affiliates’ latest findings through a range of free periodicals — the NBER Reporter , the NBER Digest , the Bulletin on Retirement and Disability , the Bulletin on Health , and the Bulletin on Entrepreneurship  — as well as online conference reports , video lectures , and interviews .

2024, 16th Annual Feldstein Lecture, Cecilia E. Rouse," Lessons for Economists from the Pandemic" cover slide

  • Feldstein Lecture
  • Presenter: Cecilia E. Rouse

 2024 Methods Lecture, Susan Athey, "Analysis and Design of Multi-Armed Bandit Experiments and Policy Learning"

  • Methods Lectures
  • Presenter: Susan Athey

2024, Economics of Social Security Panel, "Earnings Inequality and Payroll Tax Revenues"

  • Panel Discussion
  • Presenters: Karen Dynan , Karen Glenn, Stephen Goss, Fatih Guvenen & James Pearce

Help | Advanced Search

Computer Science > Computation and Language

Title: towards a realistic long-term benchmark for open-web research agents.

Abstract: We present initial results of a forthcoming benchmark for evaluating LLM agents on white-collar tasks of economic value. We evaluate agents on real-world "messy" open-web research tasks of the type that are routine in finance and consulting. In doing so, we lay the groundwork for an LLM agent evaluation suite where good performance directly corresponds to a large economic and societal impact. We built and tested several agent architectures with o1-preview, GPT-4o, Claude-3.5 Sonnet, Llama 3.1 (405b), and GPT-4o-mini. On average, LLM agents powered by Claude-3.5 Sonnet and o1-preview substantially outperformed agents using GPT-4o, with agents based on Llama 3.1 (405b) and GPT-4o-mini lagging noticeably behind. Across LLMs, a ReAct architecture with the ability to delegate subtasks to subagents performed best. In addition to quantitative evaluations, we qualitatively assessed the performance of the LLM agents by inspecting their traces and reflecting on their observations. Our evaluation represents the first in-depth assessment of agents' abilities to conduct challenging, economically valuable analyst-style research on the real open web.
Subjects: Computation and Language (cs.CL); Information Retrieval (cs.IR); Machine Learning (cs.LG)
Cite as: [cs.CL]
  (or [cs.CL] for this version)
  Focus to learn more arXiv-issued DOI via DataCite

Submission history

Access paper:.

  • HTML (experimental)
  • Other Formats

license icon

References & Citations

  • Google Scholar
  • Semantic Scholar

BibTeX formatted citation

BibSonomy logo

Bibliographic and Citation Tools

Code, data and media associated with this article, recommenders and search tools.

  • Institution

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs .

IMAGES

  1. 3 Ways to Evaluate a Research Paper

    how to evaluate a research paper

  2. Use This to Evaluate Research Paper

    how to evaluate a research paper

  3. 3 Ways to Evaluate a Research Paper

    how to evaluate a research paper

  4. Evaluation Research Examples

    how to evaluate a research paper

  5. Critical Analysis Of A Research Paper

    how to evaluate a research paper

  6. 3 Ways to Evaluate a Research Paper

    how to evaluate a research paper

VIDEO

  1. How to Evaluate Research Sources

  2. Criticality in Reviewing Literature

  3. How to evaluate research proposal?

  4. Leading and Staying on Track: Leveraging Data to Monitor and Implement Informed Programs

  5. AI-Powered Learning: The New Era of Student Development

  6. Critical evaluation of research papers, thesis & projects

COMMENTS

  1. Evaluating Research

    Evaluating Research refers to the process of assessing the quality, credibility, and relevance of a research study or project. This involves examining the methods, data, and results of the research in order to determine its validity, reliability, and usefulness. Evaluating research can be done by both experts and non-experts in the field, and ...

  2. Evaluating Research in Academic Journals: A Practical Guide to

    New to this edition: New chapters on: - evaluating mixed methods research - evaluating systematic reviews and meta-analyses - program evaluation research Updated chapters and appendices that ...

  3. 3 Ways to Evaluate a Research Paper

    1. Check for the author's name, course, instructor name, and date. Look for the important information that helps identify the paper's owner and intended recipient on page 1. This information helps both the author and instructor keep track of the research paper. [14] 2. Look for an intriguing title on the first page.

  4. Evaluating Research Articles

    Critical Appraisal. Critical appraisal is the process of systematically evaluating research using established and transparent methods. In critical appraisal, health professionals use validated checklists/worksheets as tools to guide their assessment of the research. It is a more advanced way of evaluating research than the more basic method ...

  5. Critical appraisal of published research papers

    INTRODUCTION. Critical appraisal of a research paper is defined as "The process of carefully and systematically examining research to judge its trustworthiness, value and relevance in a particular context."[] Since scientific literature is rapidly expanding with more than 12,000 articles being added to the MEDLINE database per week,[] critical appraisal is very important to distinguish ...

  6. How do you determine the quality of a journal article?

    1. Where is the article published? The journal (academic publication) where the article is published says something about the quality of the article. Journals are ranked in the Journal Quality List (JQL). If the journal you used is ranked at the top of your professional field in the JQL, then you can assume that the quality of the article is high.

  7. Evaluating Sources

    The sources you use are an important component of your research. It's important to evaluate the sources you're considering using, in order to: Ensure that they're credible. Determine whether they're relevant to your topic. Assess the quality of their arguments. You should have a clear idea of your own research question or topic and ...

  8. Tips for Evaluating Your Research

    Abstract. When evaluating Quality of Science (QoS) in the context of development initiatives, it is essential to define adequate criteria. The objective of this perspective paper is to show how altmetric and bibliometric indicators have been used to support the evaluation of QoS in the 2020 Review of the Phase 2-CGIAR Research Programs (CRPs, 2017-2022), where, for the first time, the ...

  9. Understanding and Evaluating Research: A Critical Guide

    Understanding and Evaluating Research: A Critical Guide shows students how to be critical consumers of research and to appreciate the power of methodology as it shapes the research question, the use of theory in the study, the methods used, and how the outcomes are reported. The book starts with what it means to be a critical and uncritical ...

  10. Scientific Articles: How to Evaluate a Research Article

    General Tips for Evaluating Articles. There are a couple of general questions that are worth asking of any article before reading, so that you can see how important it might be. You'll want to start with the journal. Head to the journal website, look for the "about us" or "journal information" link and think about the following questions:

  11. Evaluating Sources

    Before you decide to rely on a source, you should evaluate the source and decide whether it is appropriate to use in your paper. You should always determine the qualifications of the author, the purpose of the source (that is, in what context it was created), the scope of the source (what it covers and in what depth), and, where relevant, the ...

  12. Research Paper: A step-by-step guide: 7. Evaluating Sources

    Evaluation Criteria. It's very important to evaluate the materials you find to make sure they are appropriate for a research paper. It's not enough that the information is relevant; it must also be credible. You will want to find more than enough resources, so that you can pick and choose the best for your paper.

  13. Library Research Guides: STEM: How To Evaluate Articles

    Evaluating Articles. Not all articles are created equally. Evaluating sources for relevancy and usefulness is one the most important steps in research. This helps researchers in the STEM disciplines to gather the information they need. There are several tools to use when evaluating an article. Purpose.

  14. A simplified approach to critically appraising research evidence

    Conclusion: The ability to understand and evaluate research builds strong evidence-based practitioners, who are essential to nursing practice. Implications for practice: This framework should help readers to identify the strengths, potential weaknesses and limitations of a paper to judge its quality and potential usefulness.

  15. Evaluating research: A multidisciplinary approach to assessing research

    However, before describing and discussing our study and the resulting model, we need to clarify the terminology used. In some publications and websites on this topic, there seems to be some confusion between the term research and the term science, and these terms seem to be used interchangeably.In our view the term science is broader, and research is more like the practice of working in a ...

  16. Writing Tips: Critically Evaluating Research

    To develop the skill of being able to critically evaluate, when reading research articles in psychology read with an open mind and be active when reading. Ask questions as you go and see if the answers are provided. Initially skim through the article to gain an overview of the problem, the design, methods, and conclusions.

  17. 1 Important points to consider when critically evaluating published

    Critically evaluate the research paper using the checklist provided, making notes on the key points and your overall impression. Discussion. Critical appraisal checklists are useful tools to help assess the quality of a study. Assessment of various factors, including the importance of the research question, the design and methodology of a study ...

  18. Research Guides: Writing a Research Paper: Evaluate Sources

    Evaluate Sources With the Big 5 Criteria. The Big 5 Criteria can help you evaluate your sources for credibility: Currency: Check the publication date and determine whether it is sufficiently current for your topic. Coverage (relevance): Consider whether the source is relevant to your research and whether it covers the topic adequately for your ...

  19. Measuring research: A guide to research evaluation frameworks and tools

    In addition, a detailed overview of six research evaluation frameworks is provided, along with a brief overview of a further eight frameworks, and discussion of the main tools used in research evaluation. The report is likely to be of interest to policymakers, research funders, institutional leaders and research managers.

  20. How to Write a Literature Review

    Show how your research addresses a gap or contributes to a debate; Evaluate the current state of research and demonstrate your knowledge of the scholarly debates around your topic. Writing literature reviews is a particularly important skill if you want to apply for graduate school or pursue a career in research. We've written a step-by-step ...

  21. Research Guides: Evaluating Sources: Evaluation Tools & Rubric

    This Library Research Guide will provide you with information that you can use to critically evaluate sources including websites, articles and reports, books and ebooks, etc. Skip to Main Content. ... An acceptable score for a source to be used in a research paper for college is between 20 to 24. C.R.A.A.P.O Source Evaluation Rubric ONE (1) ...

  22. 6.2.6: Evaluating and Processing Your Sources

    An article published in 2002 about this topic will not provide current information. On the other hand, a research paper on elementary education practices might refer to studies published decades ago by influential child psychologists. When using websites for research, look on the webpage to see when the site was last updated.

  23. PDF Evaluating Sources

    Before you decide to rely on a source, you should evaluate the source and decide whether it is appropriate to use in your paper. You should always determine the qualifications of the author, the purpose of the source (that is, in what context it was created), the scope of the source (what it covers and in what depth), and, where relevant, the ...

  24. Evaluating Sources: General Guidelines

    Once you have an idea of the types of sources you need for your research, you can spend time evaluating individual sources. If a bibliographic citation seems promising, it's a good idea to spend a bit more time with the source before you determine its credibility. Below are some questions to ask and things to consider as you read through a ...

  25. Housatonic Student Research Paper Award

    The name of the professor who assigned and graded your paper; A copy of the assignment; A paragraph reflecting on what you learned about research and writing/editing strategies, and how you applied research resources and tools while working on your paper. We look forward to celebrating your hard work and exceptional research efforts!

  26. Evaluating Policy Counterfactuals: A VAR-Plus Approach

    In a rich family of linearized structural macroeconomic models, the counterfactual evolution of the macro-economy under alternative policy rules is pinned down by just two objects: first, reduced-form projections with respect to a large information set; and second, the dynamic causal effects of ...

  27. Towards a Realistic Long-Term Benchmark for Open-Web Research Agents

    We present initial results of a forthcoming benchmark for evaluating LLM agents on white-collar tasks of economic value. We evaluate agents on real-world "messy" open-web research tasks of the type that are routine in finance and consulting. In doing so, we lay the groundwork for an LLM agent evaluation suite where good performance directly corresponds to a large economic and societal impact ...