Practitioner's Guide to Using Research for Evidence-Informed Practice
eBook - ePub

Practitioner's Guide to Using Research for Evidence-Informed Practice

Allen Rubin, Jennifer Bellamy

Compartir libro
  1. English
  2. ePUB (apto para móviles)
  3. Disponible en iOS y Android
eBook - ePub

Practitioner's Guide to Using Research for Evidence-Informed Practice

Allen Rubin, Jennifer Bellamy

Detalles del libro
Vista previa del libro
Índice
Citas

Información del libro

The latest edition of an essential text to help students and practitioners distinguish between research studies that should and should not influence practice decisions Now in its third edition, Practitioner's Guide to Using Research for Evidence-Informed Practice delivers an essential and practical guide to integrating research appraisal into evidence-informed practice. The book walks you through the skills, knowledge, and strategies you can use to identify significant strengths and limitations in research. The ability to appraise the veracity and validity of research will improve your service provision and practice decisions. By teaching you to be a critical consumer of modern research, this book helps you avoid treatments based on fatally flawed research and methodologies. Practitioner's Guide to Using Research for Evidence-Informed Practice, Third Edition offers:

  • An extensive introduction to evidence-informed practice, including explorations of unethical research and discussions of social justice in the context of evidence-informed practice.
  • Explanations of how to appraise studies on intervention efficacy, including the criteria for inferring effectiveness and critically examining experiments.
  • Discussions of how to critically appraise studies for alternative evidence-informed practice questions, including nonexperimental quantitative studies and qualitative studies.

A comprehensive and authoritative blueprint for critically assessing research studies, interventions, programs, policies, and assessment tools, Practitioner's Guide to Using Research for Evidence-Informed Practice belongs in the bookshelves of students and practitioners of the social sciences.

Preguntas frecuentes

¿Cómo cancelo mi suscripción?
Simplemente, dirígete a la sección ajustes de la cuenta y haz clic en «Cancelar suscripción». Así de sencillo. Después de cancelar tu suscripción, esta permanecerá activa el tiempo restante que hayas pagado. Obtén más información aquí.
¿Cómo descargo los libros?
Por el momento, todos nuestros libros ePub adaptables a dispositivos móviles se pueden descargar a través de la aplicación. La mayor parte de nuestros PDF también se puede descargar y ya estamos trabajando para que el resto también sea descargable. Obtén más información aquí.
¿En qué se diferencian los planes de precios?
Ambos planes te permiten acceder por completo a la biblioteca y a todas las funciones de Perlego. Las únicas diferencias son el precio y el período de suscripción: con el plan anual ahorrarás en torno a un 30 % en comparación con 12 meses de un plan mensual.
¿Qué es Perlego?
Somos un servicio de suscripción de libros de texto en línea que te permite acceder a toda una biblioteca en línea por menos de lo que cuesta un libro al mes. Con más de un millón de libros sobre más de 1000 categorías, ¡tenemos todo lo que necesitas! Obtén más información aquí.
¿Perlego ofrece la función de texto a voz?
Busca el símbolo de lectura en voz alta en tu próximo libro para ver si puedes escucharlo. La herramienta de lectura en voz alta lee el texto en voz alta por ti, resaltando el texto a medida que se lee. Puedes pausarla, acelerarla y ralentizarla. Obtén más información aquí.
¿Es Practitioner's Guide to Using Research for Evidence-Informed Practice un PDF/ePUB en línea?
Sí, puedes acceder a Practitioner's Guide to Using Research for Evidence-Informed Practice de Allen Rubin, Jennifer Bellamy en formato PDF o ePUB, así como a otros libros populares de Psychology y Clinical Psychology. Tenemos más de un millón de libros disponibles en nuestro catálogo para que explores.

Información

Editorial
Wiley
Año
2022
ISBN
9781119858584
Edición
3
Categoría
Psychology

PART 1
OVERVIEW OF EVIDENCE-INFORMED PRACTICE

1
Introduction to Evidence-Informed Practice (EIP)

  1. 1.1 Emergence of EIP
  2. 1.2 Defining EIP
  3. 1.3 Types of EIP Questions
    1. 1.3.1 What Factors Best Predict Desirable or Undesirable Outcomes?
    2. 1.3.2 What Can I Learn about Clients, Service Delivery, and Targets of Intervention from the Experiences of Others?
    3. 1.3.3 What Assessment Tool Should Be Used?
    4. 1.3.4 Which Intervention, Program, or Policy Has the Best Effects?
    5. 1.3.5 What Are the Costs of Interventions, Policies, and Tools?
    6. 1.3.6 What about Potential Harmful Effects?
  4. 1.4 EIP Practice Regarding Policy and Social Justice
  5. 1.5 EIP and Black Lives Matter
  6. 1.6 Developing an EIP Practice Process Outlook
    1. 1.6.1 Critical Thinking
  7. 1.7 EIP as a Client-Centered, Compassionate Means, Not an End unto Itself
  8. 1.8 EIP and Professional Ethics
    1. 1.8.1 What about Unethical Research?
    2. Key Chapter Concepts
    3. Review Exercises
    4. Additional Readings
You've started reading the third edition of a book about using scientific research evidence to inform your practice decisions that is being written at a time when the COVID-19 pandemic is on the verge of being conquered by vaccines that were developed and evaluated by means of research. So, despite the ambivalence – perhaps even disdain – that some might have about studying research or using it to inform their practice decisions, we must ask: Especially during this post-COVID era, why would any ethical, caring helping professional choose NOT to be informed by research?
If you approach this topic with an open mind, and if you actually look for research evidence that can enhance your practice, you'll find many scientific studies that can help you to become more effective in your practice and to avoid doing harm. Seeking those studies and critically appraising them are part of what is called evidence-informed practice(EIP).
The term evidence-informed practice was more commonly called evidence-based practice when it became fashionable near the end of the last century. The main ideas behind it, however, are really quite old. As early as 1917, for example, in her classic text on social casework, Mary Richmond discussed the use of research-generated facts to guide the provision of direct clinical services as well as social reform efforts.
Also quite old is the skepticism about the notion that your practice experience and expertise – that is, your practice wisdom – are by themselves a sufficient foundation for effective practice. That skepticism does not imply that your practice experience and expertise are irrelevant and unnecessary – just that they alone are not enough.
Perhaps you don't share that skepticism. In fact, it's understandable if you even resent it. Despite the existence of research studies showing that some intervention approaches are ineffective and perhaps harmful, students learning about clinical practice have long been taught that to be an effective practitioner, they must believe in their own effectiveness as well as the effectiveness of the interventions they employed. Chances are that you have learned this, too, either in your training or through your own practice experience. It stands to reason that clients will react differently depending on whether they are being served by practitioners who are skeptical about the effectiveness of the interventions they provide versus practitioners who believe in the effectiveness of the interventions and are enthusiastic about them.
But it's hard to maintain optimism about your effectiveness if influential sources – like research-oriented scholars or managed care companies – express skepticism about the services you provide. Such skepticism was catalyzed by a notorious research study by Eysenck (1952), which concluded that psychotherapy was not effective (at least not in those days). Although various critiques of Eysenck's analysis later emerged that supported the effectiveness of psychotherapy, maintaining optimism was not easy in the face of various subsequent research reviews that shared Eysenck's conclusions about different forms of human services (Fischer, 1973; Mullen & Dumpson, 1972). Those reviews, in part, helped usher in what was then called an age of accountability – a precursor of the current EIP era.
The main idea behind this so-called age was the need to evaluate the effectiveness of all human services. It was believed that doing so would help the public learn “what bang it was getting for its buck” and, in turn, lead to discontinued funding for ineffective programs and continued funding for effective ones. Thus, that era was also known as the program evaluation movement. It eventually became apparent, however, that many of the ensuing evaluations lacked credibility due to serious flaws in their research designs and methods – flaws that often stemmed from biases connected to the vested interests of program stakeholders. Nevertheless, many scientifically rigorous evaluations were conducted, and many had encouraging results supporting the effectiveness of certain types of interventions.
In addition to studies supporting the effectiveness of particular intervention modalities, perhaps most encouraging to clinicians were studies that found that one of the most important factors influencing service effectiveness is the quality of the practitioner-client relationship. Some studies even concluded that the quality of practitioners' clinical relationship skills has more influence on treatment outcome than the choices practitioners make about what particular interventions to employ. Although that conclusion continues to be debated, some studies show that practitioner effectiveness is influenced by both the type of intervention employed and certain common relationship factors (Cuijpers et al., 2019; Nathan, 2004).

1.1 Emergence of EIP

The accumulation of scientifically rigorous studies showing that some interventions appear to be more effective than others helped spawn the EIP movement. In simple terms, the EIP movement encourages and expects practitioners to make practice decisions – especially about the interventions they provide – in light of the best scientific evidence available. In other words, practitioners might be expected to provide interventions whose effectiveness has been most supported by rigorous research and perhaps to eschew interventions that lack such support – even if it means dropping favored interventions with which they have the most experience and skills.
The preceding paragraph used the words in light of the best scientific evidence, instead of implying that the decisions had to be dictated by that evidence. That distinction is noteworthy because some mistakenly view EIP in an overly simplistic cookbook fashion that seems to disregard practitioner expertise and practitioner understanding of client values and preferences. For example, the forerunner to EIP, EBP was commonly misconstrued to be a cost-cutting tool used by third-party payers that uses a rigid decision-tree approach to making intervention choices irrespective of practitioner judgment. Perhaps you have encountered that view in your own practice (or in your own healthcare) when dealing with managed care companies that have rigid rules about what interventions must be employed as well as the maximum number of sessions that will be reimbursed. If so, you might fervently resent the EBP concept, and who could blame you! Many practitioners share that resentment.
Managed care companies that interpret EBP in such overly simplistic terms can pressure you to do things that your professional expertise leads you to believe are not in your clients' best interests. Moreover, in a seeming disregard for the scientific evidence about the importance of relationship factors and other common factors that influence positive outcomes, managed care companies can foster self-doubt about your own practice effectiveness when you do not mechanically provide the interventions on their list of what they might call “evidence-based practices.” Such doubt can hinder your belief in what you are doing and in turn hinder the more generic relationship factors that can influence client progress as much as the interventions you employ. Another problem with the list approach is its potential to stifle innovations in practice. Limiting interventions to an approved list means that novel practices are less likely to be developed and tested in the field. As you read on, you will find that EIP is a much more expansive and nuanced process than simply choosing an intervention from a list of anointed programs and services.

1.2 Defining EIP

The foregoing, overly simplistic view of EBP probably emanated from t...

Índice