Commentary
The Changing Nature of Evidence for EBLIP
Alison
Brettle
Reader in
Evidence Based Practice
School of
Nursing, Midwifery, Social Work and Social Sciences
University of
Salford
Salford,
United Kingdom
Email: a.brettle@salford.ac.uk
Received: 4 Mar.
2016 Accepted:
4 Mar. 2016
2016 Brettle. This is an Open Access article
distributed under the terms of the Creative
Commons‐Attribution‐Noncommercial‐Share Alike License 4.0 International (http://creativecommons.org/licenses/by-nc-sa/4.0/), which permits unrestricted use,
distribution, and reproduction in any medium, provided the original work is
properly attributed, not used for commercial purposes, and, if transformed, the
resulting work is redistributed under the same or similar license to this one.
It has been
an honour and a pleasure to have been involved in the EBLIP journal in so many
roles over the past 10 years. I was a peer reviewer for the first issue, then
Associate Editor Articles, Editor in Chief, and now an Editorial advisor. Time
has flown very quickly. Over recent months fewer editorial duties have enabled
me to be involved in a wider range of EBLIP related projects. This has given me
chance to reflect on the changing nature of evidence over the years.
My initial
experiences of evidence based practice took place before the terms evidence
based libraries, evidence based information practice, or evidence based library
and information practice were coined. In 1993, fresh out of library school, I
worked in a health research unit that specialized in outcome measurement, just
as evidence based medicine was being discussed and taken up within the United
Kingdom. My manager at the time, a strong advocate for librarians who saw them
as essential components in evidence based practice, encouraged me to do
research and apply that evidence to my own information practice. We went on to
develop and move forward a project that examined the feasibility of taking an
evidence based approach to social care (Long et al. 2002a). Here, we found that
it was possible to “be evidence based” but the nature of evidence wasn’t as
clear cut as in medicine. Questions were complex and difficult to focus,
quantitative research methods weren’t always appropriate, the literature was
scattered, and we needed to develop techniques and approaches for searching and
appraising more diverse evidence (Brettle & Long, 2001; Long et al. 2002 a,
b, c, d). To me this always had many parallels with the evidence base in library
and information science (LIS) and the challenges of EBLIP.
Common
definitions of EBLIP (e.g., Booth, 2000) did not stress research evidence over
professional knowledge or user preferences; however research evidence (and in
particular quantitative designs) has seemed to take priority within
professional discourses (Eldredge, 2002) about EBLIP. This has long been
problematic within LIS, as the evidence base doesn’t lend itself to such an
emphasis. The questions that librarians ask may not always lend themselves to
quantitative research designs and if they do there is unlikely to be the
funding to conduct large scale rigorous studies. Throughout all my roles within
the Evidence Based Library and
Information Practice (EBLIP) journal, my ethos has been to publish “best
evidence”. This may be about the “best” type of evidence for a particular
question or it may be the “best” type of evidence that can be collected by a
particular library at a particular time. For example, EBLIP has published systematic reviews (Koufogiannakis & Weibe,
2006), correlational studies (Eng & Stadler, 2015), quantitative analysis
(Newell, 2010), and qualitative studies (Rankin, 2012). Over time as a
profession we can look at our evidence base and seek to improve it, but the
evidence needs to fit the question and the context and, in the meantime, we
need to use the best evidence we can find to help professional decision making.
One of my
early EBLIP related projects was to conduct a systematic review on information
skills training in health libraries (Brettle, 2003). One of the key findings
was a lack of rigorous studies on which to draw conclusions about
effectiveness. Reviews conducted around the same time about clinical librarians
had similar findings (Winning and Beverley, 2003; Cimpl & Wagner, 2003) and
not much changed over the next few years with more systematic reviews noting a
lack of rigorous studies about health libraries and their services (Weightman
& Williamson, 2005; Brettle et al, 2011). More recently, however, I have
begun to sense a change. I have just completed a systematic scoping review of
the evidence for professionally trained and qualified library, information and
knowledge professionals (Brettle & Maden, 2015). Looking across all library
sectors we found evidence of effectiveness, impact, and value for health
librarians, school librarians, academic librarians, and public libraries. The
evidence came from a wide range of study designs, with some sectors favouring
particular approaches. For example the return on investment method was popular
within public libraries; school and academic libraries favoured correlational
or mixed method studies; and health libraries had the largest number of
systematic reviews and randomized controlled trials. These types of evidence
may be problematic for those who follow a traditional medical hierarchy of
evidence, but these studies undoubtedly provide evidence. As well as the wide
variety of study designs, what was particularly heartening was that quite a
number of studies were published either as primary research or as evidence
summaries in the EBLIP journal. There
is still some way to go, however. For many library sectors we found no
evidence, and there are still relatively few systematic reviews (Koufogianakis
& Brettle, 2015). We also noticed much evidence in the grey literature that
is difficult to find and appraise. These issues are a challenge for researchers
– but even more so for practitioners who want to use evidence to help them make
decisions.
I recently
co-edited a book with EBLIP journal
colleague, Denise Koufogiannakis that takes a wider view of the evidence that
librarians use (Koufogiannakis & Brettle, in press). The EBLIP model put
forward in the book emphasizes all types of evidence, whether research
evidence, local evidence or the practitioners’ knowledge (Koufogiannakis,
2013). The second part of the book examines evidence based practice and the
evidence base in different library sectors. EBLIP has developed in different
ways across sectors, and as found in our scoping review (Brettle & Maden,
2015), different sectors favour different types of evidence and study designs.
For example in school libraries, action research has been a key feature, and in
special and public libraries the concept of EBLIP is not so well known, but the
need to use evidence to demonstrate value and impact to stakeholders is key.
In the future
I’m looking forward to seeing the evidence base in LIS develop further and
seeing librarians learn from each other in finding new ways of approaching
evidence based practice and using evidence in their practice. I’m sure that the
EBLIP journal will play a key role in
this for the next 10 years and beyond.
Happy 10th birthday EBLIP
journal!
References
Booth, A. (2000). Librarian heal
thyself: Evidence based librarianship, useful, practical, desirable? Paper
presented at the 8th International Congress on Medical Librarianship, 2nd-5th
July 2000, London, UK
Brettle, A. (2003). Information
skills training: A systematic review of the literature. Health Information
& Libraries Journal, 20(Suppl 1),
3-9.
Brettle, A., & Maden, M.
(2015). What evidence is there to support
the employment of professionally trained library, information and knowledge
workers? A systematic scoping review of the evidence. Salford, University of Salford.
Brettle, A., Maden-Jenkins, M.,
Anderson, L., McNally, R., Pratchett, T., Tancock, J., Thornton, D., &
Webb, A. (2011). Evaluating clinical librarian services: A systematic review. Health Information & Libraries Journal,
28(1), 3-22. http://dx.doi.org/10.1111/j.1471-1842.2010.00925.x
Brettle, A. J., & Long, A. F.
(2001). Comparison of bibliographic databases for information on the
rehabilitation of people with severe mental illness. Bulletin of the Medical
Library Association, 89(4),
353-362.
Eldredge, J. (2002).
Evidence-based librarianship: Levels of evidence. Hypothesis, 16(3), 10-13.
Dolman, H., & Boyte-Hawryluk,
S. (2013). Impact of the reading buddies program on reading level and attitude
towards reading. Evidence Based Library & Information Practice, 8(1),
35-46. http://dx.doi.org/10.18438/B8N89T
Eng, S., & Stadler, D. (2015).
Linking library to student retention: A
statistical analysis. Evidence Based Library & Information
Practice, 10(3), 50-63. http://dx.doi.org/10.18438/B84P4D
Koufogiannakis, D., & Weibe,
N. (2006). Effective
methods for teaching information literacy skills to undergraduate students: A
systematic review and meta‐analysis. Evidence Based Library &
Information Practice, 1(3), 3-43. http://dx.doi.org/10.18438/B8MS3D
Koufogiannakis, D. (2013). How
academic librarians use evidence in their decision making: Reconsidering the
evidence based practice model (Doctoral dissertation, Aberystwyth University).
Retrieved from
http://cadair.aber.ac.uk/dspace/handle/2160/12963
Koufogiannakis, D., & Brettle,
A. (2015). Systematic reviews in LIS: Identifying evidence and gaps for
practice. 8th International Evidence Based Library and Information Practice
Conference, Brisbane, Australia. Retrieved from https://era.library.ualberta.ca/files/nk322d79d#.Vkz-J1WrS71
Koufogiannakis, D. & Brettle,
A. (in press). Being evidence based in
library and information practice.
London: Facet.
Long, A., Godfrey, M., Randall,
T., Brettle, A., & Grant, M. J. (2002). Feasibility
of undertaking systematic reviews in social care, part III. Report. Nuffield Institute of Health, University of
Leeds, and Health Care Practice R&D Unit, University of Salford, Leeds and
Salford.
Long, A. F., Godfrey, M., Randall,
T., Brettle, A., & Grant, M. J. (2002). HCPRDU Evaluation tool for mixed
methods studies. University of Leeds, Nuffield Institute for Health, Leeds.
Retrieved from http://usir.salford.ac.uk/13070/1/Evaluative_Tool_for_Mixed_Method_Studies.pdf
Long, A. F., Godfrey, M., Randall,
T., Brettle, A., & Grant, M. J. (2002). HCPRDU evaluation tool for qualitative
studies. University of Leeds, Nuffield Institute for Health, Leeds. Retrieved
from http://usir.salford.ac.uk/12970/1/Evaluation_Tool_for_Qualitative_Studies.pdf
Long, A.F., Godfrey, M., Randall,
T., Brettle, A., & Grant, M.J. (2002). HCPRDU evaluation tool for
quantitative studies. University of Leeds, Nuffield Institute for Health,
Leeds. Retrieved from http://usir.salford.ac.uk/12969/1/Evaluation_Tool_for_Quantitative_Research_Studies.pdf
Newell, T. S. (2010). Learning in
simulations: Examining the effectiveness of information literacy instruction
using middle school students' portfolio products. Evidence Based Library
& Information Practice, 5(3), 20-38. http://dx.doi.org/10.18438/B85K7T
Rankin, C. (2012). The potential
of generic social outcomes in promoting the positive impact of the public
library: Evidence from the National Year of Reading in Yorkshire. Evidence
Based Library & Information Practice, 7(1), 7-21. http://dx.doi.org/10.18438/B8VG8C
Winning, M.A., & Beverley, C.A. (2003) Clinical
librarianship: a systematic review of the literature. Health Information & Libraries Journal. 20(Suppl 1), 10-21.
Weightman, A. L., &
Williamson, J. (2005). The value and impact of information provided through
library services for patient care: A systematic review. Health Information
& Libraries Journal, 22(1), 4-25. http://dx.doi.org/10.1111/j.1471-1842.2005.00549.x