Practitioner's Guide to Using Research for Evidence-Informed Practice
eBook - ePub

Practitioner's Guide to Using Research for Evidence-Informed Practice

Allen Rubin, Jennifer Bellamy

Buch teilen
  1. English
  2. ePUB (handyfreundlich)
  3. Über iOS und Android verfügbar
eBook - ePub

Practitioner's Guide to Using Research for Evidence-Informed Practice

Allen Rubin, Jennifer Bellamy

Angaben zum Buch
Buchvorschau
Inhaltsverzeichnis
Quellenangaben

Über dieses Buch

The latest edition of an essential text to help students and practitioners distinguish between research studies that should and should not influence practice decisions Now in its third edition, Practitioner's Guide to Using Research for Evidence-Informed Practice delivers an essential and practical guide to integrating research appraisal into evidence-informed practice. The book walks you through the skills, knowledge, and strategies you can use to identify significant strengths and limitations in research. The ability to appraise the veracity and validity of research will improve your service provision and practice decisions. By teaching you to be a critical consumer of modern research, this book helps you avoid treatments based on fatally flawed research and methodologies. Practitioner's Guide to Using Research for Evidence-Informed Practice, Third Edition offers:

  • An extensive introduction to evidence-informed practice, including explorations of unethical research and discussions of social justice in the context of evidence-informed practice.
  • Explanations of how to appraise studies on intervention efficacy, including the criteria for inferring effectiveness and critically examining experiments.
  • Discussions of how to critically appraise studies for alternative evidence-informed practice questions, including nonexperimental quantitative studies and qualitative studies.

A comprehensive and authoritative blueprint for critically assessing research studies, interventions, programs, policies, and assessment tools, Practitioner's Guide to Using Research for Evidence-Informed Practice belongs in the bookshelves of students and practitioners of the social sciences.

Häufig gestellte Fragen

Wie kann ich mein Abo kündigen?
Gehe einfach zum Kontobereich in den Einstellungen und klicke auf „Abo kündigen“ – ganz einfach. Nachdem du gekündigt hast, bleibt deine Mitgliedschaft für den verbleibenden Abozeitraum, den du bereits bezahlt hast, aktiv. Mehr Informationen hier.
(Wie) Kann ich Bücher herunterladen?
Derzeit stehen all unsere auf Mobilgeräte reagierenden ePub-Bücher zum Download über die App zur Verfügung. Die meisten unserer PDFs stehen ebenfalls zum Download bereit; wir arbeiten daran, auch die übrigen PDFs zum Download anzubieten, bei denen dies aktuell noch nicht möglich ist. Weitere Informationen hier.
Welcher Unterschied besteht bei den Preisen zwischen den Aboplänen?
Mit beiden Aboplänen erhältst du vollen Zugang zur Bibliothek und allen Funktionen von Perlego. Die einzigen Unterschiede bestehen im Preis und dem Abozeitraum: Mit dem Jahresabo sparst du auf 12 Monate gerechnet im Vergleich zum Monatsabo rund 30 %.
Was ist Perlego?
Wir sind ein Online-Abodienst für Lehrbücher, bei dem du für weniger als den Preis eines einzelnen Buches pro Monat Zugang zu einer ganzen Online-Bibliothek erhältst. Mit über 1 Million Büchern zu über 1.000 verschiedenen Themen haben wir bestimmt alles, was du brauchst! Weitere Informationen hier.
Unterstützt Perlego Text-zu-Sprache?
Achte auf das Symbol zum Vorlesen in deinem nächsten Buch, um zu sehen, ob du es dir auch anhören kannst. Bei diesem Tool wird dir Text laut vorgelesen, wobei der Text beim Vorlesen auch grafisch hervorgehoben wird. Du kannst das Vorlesen jederzeit anhalten, beschleunigen und verlangsamen. Weitere Informationen hier.
Ist Practitioner's Guide to Using Research for Evidence-Informed Practice als Online-PDF/ePub verfügbar?
Ja, du hast Zugang zu Practitioner's Guide to Using Research for Evidence-Informed Practice von Allen Rubin, Jennifer Bellamy im PDF- und/oder ePub-Format sowie zu anderen beliebten Büchern aus Psychology & Clinical Psychology. Aus unserem Katalog stehen dir über 1 Million Bücher zur Verfügung.

Information

Verlag
Wiley
Jahr
2022
ISBN
9781119858584

PART 1
OVERVIEW OF EVIDENCE-INFORMED PRACTICE

1
Introduction to Evidence-Informed Practice (EIP)

  1. 1.1 Emergence of EIP
  2. 1.2 Defining EIP
  3. 1.3 Types of EIP Questions
    1. 1.3.1 What Factors Best Predict Desirable or Undesirable Outcomes?
    2. 1.3.2 What Can I Learn about Clients, Service Delivery, and Targets of Intervention from the Experiences of Others?
    3. 1.3.3 What Assessment Tool Should Be Used?
    4. 1.3.4 Which Intervention, Program, or Policy Has the Best Effects?
    5. 1.3.5 What Are the Costs of Interventions, Policies, and Tools?
    6. 1.3.6 What about Potential Harmful Effects?
  4. 1.4 EIP Practice Regarding Policy and Social Justice
  5. 1.5 EIP and Black Lives Matter
  6. 1.6 Developing an EIP Practice Process Outlook
    1. 1.6.1 Critical Thinking
  7. 1.7 EIP as a Client-Centered, Compassionate Means, Not an End unto Itself
  8. 1.8 EIP and Professional Ethics
    1. 1.8.1 What about Unethical Research?
    2. Key Chapter Concepts
    3. Review Exercises
    4. Additional Readings
You've started reading the third edition of a book about using scientific research evidence to inform your practice decisions that is being written at a time when the COVID-19 pandemic is on the verge of being conquered by vaccines that were developed and evaluated by means of research. So, despite the ambivalence – perhaps even disdain – that some might have about studying research or using it to inform their practice decisions, we must ask: Especially during this post-COVID era, why would any ethical, caring helping professional choose NOT to be informed by research?
If you approach this topic with an open mind, and if you actually look for research evidence that can enhance your practice, you'll find many scientific studies that can help you to become more effective in your practice and to avoid doing harm. Seeking those studies and critically appraising them are part of what is called evidence-informed practice(EIP).
The term evidence-informed practice was more commonly called evidence-based practice when it became fashionable near the end of the last century. The main ideas behind it, however, are really quite old. As early as 1917, for example, in her classic text on social casework, Mary Richmond discussed the use of research-generated facts to guide the provision of direct clinical services as well as social reform efforts.
Also quite old is the skepticism about the notion that your practice experience and expertise – that is, your practice wisdom – are by themselves a sufficient foundation for effective practice. That skepticism does not imply that your practice experience and expertise are irrelevant and unnecessary – just that they alone are not enough.
Perhaps you don't share that skepticism. In fact, it's understandable if you even resent it. Despite the existence of research studies showing that some intervention approaches are ineffective and perhaps harmful, students learning about clinical practice have long been taught that to be an effective practitioner, they must believe in their own effectiveness as well as the effectiveness of the interventions they employed. Chances are that you have learned this, too, either in your training or through your own practice experience. It stands to reason that clients will react differently depending on whether they are being served by practitioners who are skeptical about the effectiveness of the interventions they provide versus practitioners who believe in the effectiveness of the interventions and are enthusiastic about them.
But it's hard to maintain optimism about your effectiveness if influential sources – like research-oriented scholars or managed care companies – express skepticism about the services you provide. Such skepticism was catalyzed by a notorious research study by Eysenck (1952), which concluded that psychotherapy was not effective (at least not in those days). Although various critiques of Eysenck's analysis later emerged that supported the effectiveness of psychotherapy, maintaining optimism was not easy in the face of various subsequent research reviews that shared Eysenck's conclusions about different forms of human services (Fischer, 1973; Mullen & Dumpson, 1972). Those reviews, in part, helped usher in what was then called an age of accountability – a precursor of the current EIP era.
The main idea behind this so-called age was the need to evaluate the effectiveness of all human services. It was believed that doing so would help the public learn “what bang it was getting for its buck” and, in turn, lead to discontinued funding for ineffective programs and continued funding for effective ones. Thus, that era was also known as the program evaluation movement. It eventually became apparent, however, that many of the ensuing evaluations lacked credibility due to serious flaws in their research designs and methods – flaws that often stemmed from biases connected to the vested interests of program stakeholders. Nevertheless, many scientifically rigorous evaluations were conducted, and many had encouraging results supporting the effectiveness of certain types of interventions.
In addition to studies supporting the effectiveness of particular intervention modalities, perhaps most encouraging to clinicians were studies that found that one of the most important factors influencing service effectiveness is the quality of the practitioner-client relationship. Some studies even concluded that the quality of practitioners' clinical relationship skills has more influence on treatment outcome than the choices practitioners make about what particular interventions to employ. Although that conclusion continues to be debated, some studies show that practitioner effectiveness is influenced by both the type of intervention employed and certain common relationship factors (Cuijpers et al., 2019; Nathan, 2004).

1.1 Emergence of EIP

The accumulation of scientifically rigorous studies showing that some interventions appear to be more effective than others helped spawn the EIP movement. In simple terms, the EIP movement encourages and expects practitioners to make practice decisions – especially about the interventions they provide – in light of the best scientific evidence available. In other words, practitioners might be expected to provide interventions whose effectiveness has been most supported by rigorous research and perhaps to eschew interventions that lack such support – even if it means dropping favored interventions with which they have the most experience and skills.
The preceding paragraph used the words in light of the best scientific evidence, instead of implying that the decisions had to be dictated by that evidence. That distinction is noteworthy because some mistakenly view EIP in an overly simplistic cookbook fashion that seems to disregard practitioner expertise and practitioner understanding of client values and preferences. For example, the forerunner to EIP, EBP was commonly misconstrued to be a cost-cutting tool used by third-party payers that uses a rigid decision-tree approach to making intervention choices irrespective of practitioner judgment. Perhaps you have encountered that view in your own practice (or in your own healthcare) when dealing with managed care companies that have rigid rules about what interventions must be employed as well as the maximum number of sessions that will be reimbursed. If so, you might fervently resent the EBP concept, and who could blame you! Many practitioners share that resentment.
Managed care companies that interpret EBP in such overly simplistic terms can pressure you to do things that your professional expertise leads you to believe are not in your clients' best interests. Moreover, in a seeming disregard for the scientific evidence about the importance of relationship factors and other common factors that influence positive outcomes, managed care companies can foster self-doubt about your own practice effectiveness when you do not mechanically provide the interventions on their list of what they might call “evidence-based practices.” Such doubt can hinder your belief in what you are doing and in turn hinder the more generic relationship factors that can influence client progress as much as the interventions you employ. Another problem with the list approach is its potential to stifle innovations in practice. Limiting interventions to an approved list means that novel practices are less likely to be developed and tested in the field. As you read on, you will find that EIP is a much more expansive and nuanced process than simply choosing an intervention from a list of anointed programs and services.

1.2 Defining EIP

The foregoing, overly simplistic view of EBP probably emanated from t...

Inhaltsverzeichnis