Practitioner's Guide to Using Research for Evidence-Informed Practice
eBook - ePub

Practitioner's Guide to Using Research for Evidence-Informed Practice

Allen Rubin, Jennifer Bellamy

Share book
  1. English
  2. ePUB (mobile friendly)
  3. Available on iOS & Android
eBook - ePub

Practitioner's Guide to Using Research for Evidence-Informed Practice

Allen Rubin, Jennifer Bellamy

Book details
Book preview
Table of contents
Citations

About This Book

The latest edition of an essential text to help students and practitioners distinguish between research studies that should and should not influence practice decisions Now in its third edition, Practitioner's Guide to Using Research for Evidence-Informed Practice delivers an essential and practical guide to integrating research appraisal into evidence-informed practice. The book walks you through the skills, knowledge, and strategies you can use to identify significant strengths and limitations in research. The ability to appraise the veracity and validity of research will improve your service provision and practice decisions. By teaching you to be a critical consumer of modern research, this book helps you avoid treatments based on fatally flawed research and methodologies. Practitioner's Guide to Using Research for Evidence-Informed Practice, Third Edition offers:

  • An extensive introduction to evidence-informed practice, including explorations of unethical research and discussions of social justice in the context of evidence-informed practice.
  • Explanations of how to appraise studies on intervention efficacy, including the criteria for inferring effectiveness and critically examining experiments.
  • Discussions of how to critically appraise studies for alternative evidence-informed practice questions, including nonexperimental quantitative studies and qualitative studies.

A comprehensive and authoritative blueprint for critically assessing research studies, interventions, programs, policies, and assessment tools, Practitioner's Guide to Using Research for Evidence-Informed Practice belongs in the bookshelves of students and practitioners of the social sciences.

Frequently asked questions

How do I cancel my subscription?
Simply head over to the account section in settings and click on “Cancel Subscription” - it’s as simple as that. After you cancel, your membership will stay active for the remainder of the time you’ve paid for. Learn more here.
Can/how do I download books?
At the moment all of our mobile-responsive ePub books are available to download via the app. Most of our PDFs are also available to download and we're working on making the final remaining ones downloadable now. Learn more here.
What is the difference between the pricing plans?
Both plans give you full access to the library and all of Perlego’s features. The only differences are the price and subscription period: With the annual plan you’ll save around 30% compared to 12 months on the monthly plan.
What is Perlego?
We are an online textbook subscription service, where you can get access to an entire online library for less than the price of a single book per month. With over 1 million books across 1000+ topics, we’ve got you covered! Learn more here.
Do you support text-to-speech?
Look out for the read-aloud symbol on your next book to see if you can listen to it. The read-aloud tool reads text aloud for you, highlighting the text as it is being read. You can pause it, speed it up and slow it down. Learn more here.
Is Practitioner's Guide to Using Research for Evidence-Informed Practice an online PDF/ePUB?
Yes, you can access Practitioner's Guide to Using Research for Evidence-Informed Practice by Allen Rubin, Jennifer Bellamy in PDF and/or ePUB format, as well as other popular books in Psychology & Clinical Psychology. We have over one million books available in our catalogue for you to explore.

Information

Publisher
Wiley
Year
2022
ISBN
9781119858584
Edition
3

PART 1
OVERVIEW OF EVIDENCE-INFORMED PRACTICE

1
Introduction to Evidence-Informed Practice (EIP)

  1. 1.1 Emergence of EIP
  2. 1.2 Defining EIP
  3. 1.3 Types of EIP Questions
    1. 1.3.1 What Factors Best Predict Desirable or Undesirable Outcomes?
    2. 1.3.2 What Can I Learn about Clients, Service Delivery, and Targets of Intervention from the Experiences of Others?
    3. 1.3.3 What Assessment Tool Should Be Used?
    4. 1.3.4 Which Intervention, Program, or Policy Has the Best Effects?
    5. 1.3.5 What Are the Costs of Interventions, Policies, and Tools?
    6. 1.3.6 What about Potential Harmful Effects?
  4. 1.4 EIP Practice Regarding Policy and Social Justice
  5. 1.5 EIP and Black Lives Matter
  6. 1.6 Developing an EIP Practice Process Outlook
    1. 1.6.1 Critical Thinking
  7. 1.7 EIP as a Client-Centered, Compassionate Means, Not an End unto Itself
  8. 1.8 EIP and Professional Ethics
    1. 1.8.1 What about Unethical Research?
    2. Key Chapter Concepts
    3. Review Exercises
    4. Additional Readings
You've started reading the third edition of a book about using scientific research evidence to inform your practice decisions that is being written at a time when the COVID-19 pandemic is on the verge of being conquered by vaccines that were developed and evaluated by means of research. So, despite the ambivalence – perhaps even disdain – that some might have about studying research or using it to inform their practice decisions, we must ask: Especially during this post-COVID era, why would any ethical, caring helping professional choose NOT to be informed by research?
If you approach this topic with an open mind, and if you actually look for research evidence that can enhance your practice, you'll find many scientific studies that can help you to become more effective in your practice and to avoid doing harm. Seeking those studies and critically appraising them are part of what is called evidence-informed practice(EIP).
The term evidence-informed practice was more commonly called evidence-based practice when it became fashionable near the end of the last century. The main ideas behind it, however, are really quite old. As early as 1917, for example, in her classic text on social casework, Mary Richmond discussed the use of research-generated facts to guide the provision of direct clinical services as well as social reform efforts.
Also quite old is the skepticism about the notion that your practice experience and expertise – that is, your practice wisdom – are by themselves a sufficient foundation for effective practice. That skepticism does not imply that your practice experience and expertise are irrelevant and unnecessary – just that they alone are not enough.
Perhaps you don't share that skepticism. In fact, it's understandable if you even resent it. Despite the existence of research studies showing that some intervention approaches are ineffective and perhaps harmful, students learning about clinical practice have long been taught that to be an effective practitioner, they must believe in their own effectiveness as well as the effectiveness of the interventions they employed. Chances are that you have learned this, too, either in your training or through your own practice experience. It stands to reason that clients will react differently depending on whether they are being served by practitioners who are skeptical about the effectiveness of the interventions they provide versus practitioners who believe in the effectiveness of the interventions and are enthusiastic about them.
But it's hard to maintain optimism about your effectiveness if influential sources – like research-oriented scholars or managed care companies – express skepticism about the services you provide. Such skepticism was catalyzed by a notorious research study by Eysenck (1952), which concluded that psychotherapy was not effective (at least not in those days). Although various critiques of Eysenck's analysis later emerged that supported the effectiveness of psychotherapy, maintaining optimism was not easy in the face of various subsequent research reviews that shared Eysenck's conclusions about different forms of human services (Fischer, 1973; Mullen & Dumpson, 1972). Those reviews, in part, helped usher in what was then called an age of accountability – a precursor of the current EIP era.
The main idea behind this so-called age was the need to evaluate the effectiveness of all human services. It was believed that doing so would help the public learn “what bang it was getting for its buck” and, in turn, lead to discontinued funding for ineffective programs and continued funding for effective ones. Thus, that era was also known as the program evaluation movement. It eventually became apparent, however, that many of the ensuing evaluations lacked credibility due to serious flaws in their research designs and methods – flaws that often stemmed from biases connected to the vested interests of program stakeholders. Nevertheless, many scientifically rigorous evaluations were conducted, and many had encouraging results supporting the effectiveness of certain types of interventions.
In addition to studies supporting the effectiveness of particular intervention modalities, perhaps most encouraging to clinicians were studies that found that one of the most important factors influencing service effectiveness is the quality of the practitioner-client relationship. Some studies even concluded that the quality of practitioners' clinical relationship skills has more influence on treatment outcome than the choices practitioners make about what particular interventions to employ. Although that conclusion continues to be debated, some studies show that practitioner effectiveness is influenced by both the type of intervention employed and certain common relationship factors (Cuijpers et al., 2019; Nathan, 2004).

1.1 Emergence of EIP

The accumulation of scientifically rigorous studies showing that some interventions appear to be more effective than others helped spawn the EIP movement. In simple terms, the EIP movement encourages and expects practitioners to make practice decisions – especially about the interventions they provide – in light of the best scientific evidence available. In other words, practitioners might be expected to provide interventions whose effectiveness has been most supported by rigorous research and perhaps to eschew interventions that lack such support – even if it means dropping favored interventions with which they have the most experience and skills.
The preceding paragraph used the words in light of the best scientific evidence, instead of implying that the decisions had to be dictated by that evidence. That distinction is noteworthy because some mistakenly view EIP in an overly simplistic cookbook fashion that seems to disregard practitioner expertise and practitioner understanding of client values and preferences. For example, the forerunner to EIP, EBP was commonly misconstrued to be a cost-cutting tool used by third-party payers that uses a rigid decision-tree approach to making intervention choices irrespective of practitioner judgment. Perhaps you have encountered that view in your own practice (or in your own healthcare) when dealing with managed care companies that have rigid rules about what interventions must be employed as well as the maximum number of sessions that will be reimbursed. If so, you might fervently resent the EBP concept, and who could blame you! Many practitioners share that resentment.
Managed care companies that interpret EBP in such overly simplistic terms can pressure you to do things that your professional expertise leads you to believe are not in your clients' best interests. Moreover, in a seeming disregard for the scientific evidence about the importance of relationship factors and other common factors that influence positive outcomes, managed care companies can foster self-doubt about your own practice effectiveness when you do not mechanically provide the interventions on their list of what they might call “evidence-based practices.” Such doubt can hinder your belief in what you are doing and in turn hinder the more generic relationship factors that can influence client progress as much as the interventions you employ. Another problem with the list approach is its potential to stifle innovations in practice. Limiting interventions to an approved list means that novel practices are less likely to be developed and tested in the field. As you read on, you will find that EIP is a much more expansive and nuanced process than simply choosing an intervention from a list of anointed programs and services.

1.2 Defining EIP

The foregoing, overly simplistic view of EBP probably emanated from t...

Table of contents