Editorial
Learning From
Others About Research Evidence
Alison
Brettle
Editor-in-Chief
Senior
Lecturer, School of Nursing Midwifery and Social Work
University
of Salford
Salford,
United Kingdom
Email:
a.brettle@salford.ac.uk
2012 Brettle.
This is an Open Access article distributed under the terms of the Creative
Commons‐Attribution‐Noncommercial‐Share Alike License 2.5 Canada (http://creativecommons.org/licenses/by-nc-sa/2.5/ca/),
which permits unrestricted use, distribution, and reproduction in any medium,
provided the original work is properly attributed, not used for commercial purposes,
and, if transformed, the resulting work is redistributed under the same or
similar license to this one.
Welcome
to the June issue of EBLIP, our first to be published with an HTML version as
well as PDFs for each article. I hope you enjoy and find the alternative
formats useful. As usual the issue comprises an interesting range of evidence
summaries and articles that I hope you will find useful in applying evidence to
your practice.
When
considering evidence, two recent trips to Edinburgh got me thinking about the
wide range of study designs or methods that are useful for generating evidence,
and also how we can learn about their use from other professions.
The
first trip was as part of the cadre of the LIS DREaM
project (http://lisresearch.org/dream-project/). DREaM
has been set up by the LIS Research Coalition to develop a sustainable LIS
research network in the UK. As part of this, a series of workshops aims to
introduce LIS practitioners to a wider range of research methods, thus
expanding the methods used in LIS research. Indeed, a quick scan of the
contents of this issue show a preponderance of surveys, interviews, and
citation analysis, suggesting that broadening our knowledge of methods may well
be a useful idea. The workshops are highly interactive and, at each session
experts from outside the LIS discipline introduce particular research methods
and outline how they could be used in LIS applications. As a result, I can see
the value and understand when to use research methods such as social network
analysis, horizon scanning, ethnography, discourse analysis, and repertory
grids – as well as knowing that data mining is something I’m likely to avoid!
So far I’ve shared my new knowledge with a PhD student who was considering her
methodology and incorporated my new knowledge of horizon scanning into a bid
for research funding. The next (and more exciting) step is to think of a
situation where I can apply one of these methods to examining an aspect of LIS
practice.
The
second trip was the British Association of Counselling and Psychotherapy
Research Conference, an event which I've attended for the last few years (don’t
ask!). Each time, I've been struck by both the similarities and differences
between counselling and LIS research in the UK.
Counselling research is conducted by a relatively small number of
individuals and, as in LIS, the vast majority of practitioners don’t engage in
writing any research up for publication (Clapton, 2010). Particular types of
research dominate in counselling, but most are highly qualitative in manner,
e.g. using biographical approaches. I can’t immediately see how these could become
widely used in LIS, but I do find it fascinating to hear about different
approaches and the evidence this provides. Like many of the things that LIS
professionals do, counselling and psychotherapy is a complex intervention and
it is not always immediately apparent what has caused an effect. It may well be
that the counsellor is only one of a number of elements that has led to a
positive outcome or a change in effect. This makes it difficult to generate
evidence about the effectiveness of counselling, similar, for example, to
trying to generate evidence regarding the effectiveness of information
literacy.
Due
to political drivers there is an increasing interest (and resistance) to a more
evidence based approach in counselling and psychotherapy. One of the main areas
of resistance towards evidence based practice (EBP) in counselling is that the
medical model or paradigm of EBP and the view that the randomized controlled
trial (RCT) is the method of choice for providing high quality evidence on the
effectiveness of services doesn't fit with the way counsellors provide services
to their clients. Each client is seen as an individual and therapy is provided
according to a client’s particular needs at that time rather than following a
set manual or course. This makes it impossible to assess in a "randomized
controlled" manner, before even beginning to worry about the ethical and
practical implications of conducting an experimental study.
The
unsuitability of the RCT has also been raised regarding generating evidence for
EBLIP (e.g. Banks, 2008); however, “best evidence” doesn’t need to be an RCT.
The definition of EBLIP provided by Booth (2006) mentions best quality evidence
(generated from research, among other elements, but makes no mention of
particular research designs). In addition, both Eldredge
(2004) and Crumley and Koufogiannakis (2002) have
argued for the consideration of a wide type of study designs as evidence within
EBLIP, a viewpoint with which I have long agreed. After all, it is much more
important to choose a design that is suitable to answer the question at hand
and provide good quality evidence, rather than trying to use a "good
quality" design at the expense of finding relevant evidence. Bearing that
in mind, I'm racking my brains to think of how I can use webometrics
and techniques from history to investigate my practice. At the same time, I
urge you to think widely about research evidence and try exploring some
different methodologies and see what evidence they can reveal.
References
Banks,
M.A. (2008). Friendly skepticism about evidence based library and information
practice, Evidence Based Library and Information
Practice, 3(3): 86-90.
Booth, A. (2006). Counting what counts:
performance measurement and evidence-based practice. Performance Measurement
and Metrics, 7(2), 63-74.
Clapton, J. (2010). Library and information
science practitioners writing for publication: motivations, barriers and
supports. Library and Information
Research, 34(106): 7-21.
Crumley,
E., & Koufogiannakis, D. (2002). Developing evidence-based librarianship: practical steps for
implementation. Health Information & Libraries
Journal, 19(2), 61.
Eldredge,
J. D. (2004). Inventory of research methods for librarianship
and informatics. Journal of the Medical Library Association, 92(1),
83-90.
Note:
Videos and material from the DREaM events is
available from http://lisresearch.org/dream-project/dream-workshops/