Article

Academic Librarians’ Conception and Use of Evidence Sources in Practice

Denise Koufogiannakis
Collections and Acquisitions Coordinator
University of Alberta Libraries
Edmonton, Alberta, Canada
Email:
dak@ualberta.ca

Received: 20 Aug. 2012 Accepted: 15 Nov. 2012

 

 

cc-ca_logo_xl 2012 Koufogiannakis. This is an Open Access article distributed under the terms of the Creative CommonsAttributionNoncommercialShare Alike License 2.5 Canada (http://creativecommons.org/licenses/by-nc-sa/2.5/ca/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly attributed, not used for commercial purposes, and, if transformed, the resulting work is redistributed under the same or similar license to this one.

 


Abstract

 

Objective – The objective of this study was to explore and understand how academic librarians use evidence in their professional decision making. The researcher aimed to gain insights on the relevance of the current EBLIP model to practice, and to understand the possible connections between scientific research and tacit knowledge within the practice of LIS.

 

Methods – A grounded theory methodology was used, following the approach of Charmaz (2006). Participants were 19 academic librarians in Canada. Data was gathered via online diaries and semi-structured interviews over a six-month period in 2011.

 

Results – Two broad types of evidence were identified (hard and soft), and are generally used in conjunction with one another. Librarians examine all evidence sources with a critical eye, and try to determine a complete picture before reaching a conclusion. As well, librarians use a variety of proactive and passive approaches to find evidence.

 

Conclusions – These results provide a strong message that no single evidence source is perfect. Consequently, librarians bring different types of evidence together in order to be as informed as possible before making a decision. Using a combination of evidence sources, depending upon the problem, is the way academic librarians approach decision making.


 

 

Introduction

 

Evidence based practice (EBP) is a relatively young movement, which began in medicine (Guyatt, 1991) and has since spread to other fields, including library and information studies (LIS). In LIS, very little research has been undertaken on the evidence based library and information practice (EBLIP) model that was directly adapted from medicine, despite the fact that LIS is a social science discipline. This direct adaptation, without reflection on the differences between LIS and medicine, has been a noted criticism of the current EBLIP model (Given, 2006; Hunsucker, 2007). With roots in evidence based medicine, does the focus on quantitative research evidence apply to librarians whose questions often demand explanations rather than judgments on the effectiveness of interventions? Can the current model address academic librarians’ questions and assist with decision making in a meaningful way? The current model may be alienating some librarians who feel that the forms of evidence they are using are not being recognized as important.

 

This research study examined the foundation of EBLIP by exploring how academic librarians use evidence in their practice. The definition of evidence used within this study was from the Oxford Dictionary – “the available body of facts or information indicating whether a belief or proposition is true or valid” (2010) – while keeping in mind that within EBP, evidence is generally considered to be research. The study sought to examine whether this was the case in LIS practice or whether librarians have a broader interpretation of evidence.

 

The research presented in this paper describes evidence sources used by academic librarians, as well as the reasons these sources are used. It also examines how academic librarians view different sources of evidence, and the differences between what is used in practice and what is conceptually considered to be evidence.

 

Literature Review

 

Evidence Sources in Evidence Based Practice

 

Evidence based library and information practice is strongly modelled on the original evidence based medicine (EBM) process. The most widely cited and accepted definition of EBLIP was adapted from McKibbon, Wilczynski, Hayward, Walker-Dilks, and Haynes’s (1995) definition of EBM, keeping all the same components and basic meaning, but inserting “user” in place of “patient” and “librarian” in place of “clinician”:

 

An approach to information science that promotes the collection, interpretation and integration of valid, important and applicable user-reported, librarian-observed, and research-derived evidence. The best available evidence, moderated by user needs and preferences, is applied to improve the quality of professional judgements. (Booth, 2000)

 

The EBM movement has generally focused on research studies as the primary source of evidence. EBM has produced many tools for practitioners, to assist them with critical appraisal of research evidence and with determining the strength of the research evidence. There has been criticism that evidence based models do not account for other forms of knowledge that are a vital part of professional practice (Brophy, 2009; Clark, 2011; Davies, Nutley, & Walter, 2008).

 

Built into the EBM model is a hierarchy of evidence (Howick et al., 2011; SUNY, 2004) which EBLIP has also mirrored (Eldredge 2000a, 2000b, 2002). In the hierarchy of evidence, research methods such as randomized control trials are at the top of the hierarchy because they are more likely to be free of bias. While the levels of evidence are a well-known aspect of EBLIP, they are not something that the EBLIP community has wholeheartedly accepted. The application of such a hierarchy has been a concern for many within the field (Banks, 2008; Booth, 2010; Crumley & Koufogiannakis, 2002; Given, 2006; Koufogiannakis, 2010).

 

Beyond Research Evidence

 

In the evidence based medicine model, scientific research is the main concept explored in relation to practice. However, there are other evidence sources beyond research that impact professional practice and decision making. In this study, practice theory was used as an alternative lens to view the EBLIP model. Practice theory explores what people actually do in practice, and examines how the active doing of a practice leads to knowledge that is important to that practice.

 

Schatzki’s (1996) book, Social Practices, was the first to wholly focus on the practice concept. In that seminal work, Schatzki outlines the theory of practices and the necessity of action within practice. A key element of practice theory is the concept of knowing in practice. In practice, knowing has two elements that cannot be separated; these are “knowing how” and “knowing that,” phrases first coined by Ryle in 1945. Knowing that relates to the mind, and how to do a particular thing, so that it is explainable. Knowing how relates to doing the thing, or action, even if one does not know how to explain how one has done it (tacit knowledge). Polanyi (1966) was the first to delve into tacit knowledge, explaining it as “we can know more than we can tell” (p. 4).

 

Schön, building upon the work of Polanyi, writes in his influential 1983 work, The Reflective Practitioner: How Professionals Think in Action, that “our knowing is in our action” (p. 49). For Schön the work life of a professional depends on this tacit knowing in action. Schön says: “Even when [the practitioner] makes conscious use of research-based theories and techniques, he is dependent on tacit recognitions, judgements, and skilful performances” (p. 50). The two aspects, research and professional knowledge, must go hand in hand.

 

Looking beyond theory, several professions are beginning to embrace a practice-based evidence approach in addition to an evidence based practice one. In the fields of medicine and nursing, Gabbay and Le May (2011) have done ethnographic research to reveal how clinicians acquire and use their knowledge. They convey the importance of “knowledge-in-practice-in-context” (p. 65), and note that medicine is an art in addition to a science. It requires judgment and decision-making skills in addition to scientific knowledge.

 

Many professional fields have also examined the importance to professional practice of evidence sources other than scientific research (Clark, 2011; Fox, 2003; Rolfe, Jasper, & Freshwater, 2011; Usher & Bryant, 1989), not rejecting research but widening the conception of what is required to make good decisions in practice. For practitioners, learning occurs via doing (Schön, 1983). Within communities of practice (Lave & Wenger, 1991), the focus of this learning is on the social nature of the community. Practitioners learn from others within their community, and likewise contribute to that learning. Communities of practice occur whether one is conscious of them or not. In an unconscious format, practitioners rely on their internal networks to assist with learning tacit dimensions of their work, via conversations with colleagues, interactions in groups, and verification from peers. Duguid (2005) explains that in becoming a practitioner, one needs to “learn to be,” which is part of Ryle’s concept of “knowing how,” embodying the art of practice and tacit dimensions that are not easily made explicit.

 

Communities of practice have the potential to allow for individual practitioners to bring their practice-based knowledge to a conversation within their practicing community. Practice-based knowledge is therefore made more explicit, and learning occurs within the group, ultimately influencing practice decisions. How academic librarians function within their communities of practice, and how these communities affect their knowledge and decision making, are of interest to this study because it is within such communities that tacit knowledge is formed.

 

Aims

 

The aim of this study was to explore and better understand how academic librarians use evidence in their professional decision making. The purpose was to gain insights on the applicability of the current EBLIP model for LIS practitioners, and to understand the possible connections between scientific research and tacit knowledge within the practice of LIS.

The following research questions were posed:

 

·         What forms of evidence do academic librarians use when making professional decisions? Why do they use these types of evidence?

·         How do academic librarians incorporate research into their professional decision making?

 

Methods

 

The study used a grounded theory methodology, following the approach of Charmaz (2006). The methods used to collect data were online diaries (blogs) and semi-structured interviews. Ethical approval was received from both Aberystwyth University, where the researcher was a student, and the University of Alberta, where the researcher is employed as a librarian.

 

The study used a purposeful sample of Canadian academic librarians who had some interest in exploring the use of evidence in relation to their professional decision making. Although the research was targeted at academic librarians, a wide variance was sought, and so an open invitation to participate was sent out on mailing lists that are used by academic librarians, such as the Canadian Association of College and University Libraries mailing list, and the Evidence Based Librarianship Interest Group of the Canadian Library Association. The invitation was also sent out on Twitter.

 

Twenty-one librarians initially agreed to participate in the study. Two librarians later dropped out, due to time constraints, leaving a total of 19 participants. This number was sufficient to reach saturation of the data, and included variance amongst participants including demographics, work types within academic libraries, and knowledge of EBLIP. The 19 participants were geographically dispersed across Canada and were all English-language speakers. All worked in academic positions, identified themselves as academic librarians, and worked in a variety of roles and subject areas. The participants’ number of years of experience as librarians varied widely, ranging from less than two years to more than 30 years. They represented all levels of experience, from new librarians in their first job, to senior librarians nearing retirement. Some librarians had many years of experience but had recently begun new positions, while others had been in the same position for many years. Each participant’s familiarity with evidence based practice was assessed based on an analysis of comments in the diaries and interviews, and it was determined that eight participants were very familiar with EBP, three were moderately familiar, and eight had very little to no familiarity with EBP.

 

The process of data collection occurred over a period of nearly six months, simultaneously in conjunction with data analysis. Data collection occurred in a theoretical manner; as concepts emerged and patterns were discovered, the researcher followed up on those emerging concepts with the later participants. The study aimed for depth and richness of information rather than higher numbers of participants; the data is not meant to be generalized, but will be used to provide insights that may aid in the development of theory regarding evidence based approaches in librarianship.

 

Participants wrote in their online diaries for a period of one month. They were asked to note questions or problems that related to their professional practice and how they resolved those issues (see Appendix A). Participants used WordPress.com online blogging software, which allows for blogs to be kept private. All participants who completed the diary portion of the research agreed to a follow-up interview. The semi-structured interview process (see Appendix B) allowed clarification and deeper analysis of specific aspects that participants may have noted in their diary entries, and allowed participants to look holistically at their experiences and to comment on the overall process.

 

Given the wide geographic distribution of participants across Canada, most interviews were conducted via telephone or Skype. All interviews were taped using a digital recorder. Audio tapes were transcribed by a professional transcriptionist and checked for accuracy by the researcher.

 

Analysis of the diaries began as each was completed, using the constant comparison method to closely analyze the text and discover and group concepts related to the decision-making process of participants. This process of comparing each incident in the data with other incidents, and doing so continually as the data is gathered, allows the researcher to determine analytic similarities and differences (Charmaz, 2006; Corbin & Strauss, 2008). As additional diaries and interviews were completed, the information gained from the earlier data was used to refine concepts and discover new ones, as is the norm within the grounded theory method. Memo-writing was used to keep a reflective record of the approach to the research as well as emergent concepts. An open coding approach was used on a printed copy of the diary and interview transcripts, and later transferred into the NVivo software program, which was used to assist with the management of data analysis. Very specific codes were later grouped into categories, as analysis was refined and a picture of the findings began to emerge. Saturation of the data was reached by the 16th interview, when no new theoretical insights arose from the data and no new categories emerged when coding.

 

Findings

 

The Concept of Evidence

 

While the interviews in this study were semi-structured with a focus on following up on situations that participants had raised in their blog diaries, participants were asked a direct question about what they considered to be evidence. Other than one participant, all who responded to the question about what they considered to be evidence were very open to the possibility of what evidence could be within librarianship. Responses that exemplified this outlook included “there are lots of things that are evidence” (Librarian 10) and “I consider every information source to be evidence” (Librarian 14). Most participants named several sources of evidence, and usually put those in context. For example, they chose different evidence sources depending upon the problem faced.

 

All participants noted research literature, or simply “literature,” as evidence, often qualifying this source in terms such as “obviously” or “of course.” However, there were some caveats put on the inclusion of published literature, due to the participants’ discomfort with the quality and relevance of literature they have found in the past. This is exemplified by Librarian 10 who said “obviously research is another kind of evidence although it is not totally implacable” (Librarian 10, interview).

 

Another concept mentioned very frequently as evidence was “looking at what other libraries do.” This evidence may come from the literature in the form of descriptive articles about an innovative service at a particular library, but may also be found by examining other libraries’ websites or catalogues, speaking with librarians at other institutions, or hearing about other library experiences via a conference presentation or an electronic mailing list. This type of evidence provides ideas and insights relating to a problem that a librarian may be working on. As Librarian 20 noted:

 

I do find that hearing the experience of other librarians, getting some of their ideas – maybe it’s not what you would term hardcore evidence, but I do find that that really just generates ideas, better ways of doing things or more interesting things. (Librarian 20, interview)

 

“What other librarians do” also provides a starting point and guidance when approaching a problem one has not encountered before, or when trying something new. There is a reassurance in knowing how things worked for someone else, particularly peer-sized institutions that have similar populations. Such insights provide the level of detail that inquiring librarians need, as they are able to ask specific questions.

Data, or what is commonly referred to as statistics, was another key area mentioned by participants when asked to discuss what they thought of as evidence. Keeping statistics on reference transactions, circulation of books, usage of electronic journals, interlibrary loan requests, and so on are very common in libraries. Hence, it is not surprising that academic librarians are looking to those sources as evidence to help with their decision making. As Librarian 11 pointed out:

 

I think my gut reaction is that I want numbers of things. I want quantitative information. I want numbers of transactions, numbers of uses, and so on. I think that’s probably a fairly shallow interpretation of evidence, but that’s the kind I like. (Librarian 11, interview)

 

As with the literature, most librarians were also cautious about statistics and often qualified their statements by noting that there were problems with this type of evidence, and that it could not simply be viewed in isolation.

Very often, librarians referred to the need to look at many types of evidence, particularly depending upon the situation. This is exemplified by Librarian 14, who stated:

 

I consider every information source to be evidence. And I guess I mean that in the very broadest category, so it could be someone’s opinion or it could be a report. I feel confident in my ability to judge whether evidence is credible or not. So, I think I would look at everything. I wouldn’t discount anything. (Librarian 14, interview)

 

Regardless of whether they felt certain that some sources really were “evidence” or not, participants did mention experience, opinion, and anecdote. These seem to fall into a grey area, as most people who mentioned them did not feel absolutely comfortable or certain that they were evidence sources. One person was very certain that they were not, and another that they were. But most were unclear about these sources, acknowledging that they were used, but uncertain about whether they could or should be considered evidence.

 

Academic librarians generally have a very wide view of evidence, while at the same time, they are for the most part unsure of what constitutes evidence. They want to consider evidence carefully and are willing to take into account whatever may help them with decision making. They also consciously weigh evidence in an effort to make a good decision with the available evidence. For this, they rely on their own professional judgment and knowledge of what is most important in a particular situation.

 

Evidence Sources Used

 

The evidence sources used by academic librarians were numerous and detailed. In order to best convey this information, the evidence sources were grouped into two overarching types, hard evidence and soft evidence, at a final stage of the coding process in order to make a distinction between the types of evidence that were used or mentioned by participants. There were a total of nine categories of evidence, which are listed in Table 1.

 

“Hard” evidence sources are usually more scientific in nature. Ultimately, there is some written, concrete information tied to this type of evidence. A librarian can point to it and easily share it with colleagues. It is often vetted though an outside body (publisher or institution) and adheres to a set of rules. These sources are generally acknowledged as acceptable sources of evidence, and are what a librarian would normally think of as evidence in LIS.

 

The other type of evidence can be thought of as “soft” or non-scientific evidence. These evidence sources focus on experience and accumulated knowledge, opinion, instinct, and what other libraries or librarians do. This type of evidence focuses on a story, and how things fit in a particular context. Soft evidence provides a real-life connection, insights, new ideas, and inspiration. These types of evidence are more informal and generally not seen as deserving of the label “evidence,” although they are used by academic librarians in their decision making.

 

Table 1

Sources of Evidence Used by Academic Librarians

Evidence Source

Definition

Examples

Hard Evidence

Published literature

Scholarly publications that have been vetted via a publication process

Journal articles (research and non-research), books, databases, conference papers, etc.

Statistics

Data pertaining to the use of a particular product or service

Usage statistics, reference statistics, circulation statistics, etc.

Local research and evaluation

The evaluation and assessment of services

Course evaluations, surveys, focus groups, etc.

Other documents

Non-scholarly publications that provide information about a service, event, or person

Policies, Web pages, blogs, course materials

Facts

Things that the majority of people agree to be true

Cost of a product, date of a publication

Soft Evidence

Input from colleagues

Going to colleagues to ask their advice or feedback, or for information about a program or service that they may know about

Discussions, feedback, brainstorming, conference presentations

Tacit knowledge

Knowledge that is embodied by an individual and difficult to transfer to another person

Experience, intuition, “common sense”

Feedback from users

Individual feedback received from users on products or services

Comments, discussions, email

Anecdotal evidence

“Information obtained from personal accounts, examples, and observations. Usually not considered scientifically valid but may indicate areas for further investigation and research” (Jonas, 2005).

Stories, observation

 

 

Hard Evidence Sources

 

Published Literature

 

An important source of evidence consulted by academic librarians is the published literature. The published literature includes journal articles from both LIS journals as well as non-LIS journals, and can include both research articles and non-research articles, and quantitative and qualitative studies. It also includes books, databases, guidelines, bibliographies, and any other similar source that has been published. Participants noted that the literature provides them with a wider context, background information, and theoretical models. It also reinforces certain principles and reassures them of what they are doing. As the following comment illustrates, the literature reassures that one is on the right track:

 

So, the lit search, I think it was useful, at least in terms of giving me confidence that I wasn’t overlooking anything major. That the stuff I had figured out was about right. (Librarian 1, interview)

 

The literature is rarely consulted in isolation. It is considered as just one piece of evidence in a decision and is often used for background information gathering when one is faced with a new problem. However, the literature does not always offer sufficient answers. Librarians find the literature somewhat useful, but at the same time disappointing. They wish that the quality of the library literature was higher and that it was more relevant to their practice. Sometimes, they do not find anything in the literature, or what they do find is not useful. However, no participant was ready to completely disregard the literature. While participants noted different types of literature and occasionally mentioned types of studies or the lack of good research, they detailed differences between specific types of research literature.

 

Statistics

 

Data in the form of library statistics is a very common source of evidence among academic librarians. Participants frequently mentioned using information such as usage data, circulation statistics, reference statistics, interlibrary loan data, room bookings, and Web usage data. This type of evidence is most common when problems arise relating to collection management, and also reference services. Participants generally felt that such statistics provide an overall picture of the general situation as it pertains to use of a particular collection or service. For example, in comparing journals in a particular field, usage statistics would be looked at in order to determine what journals are being most heavily used by faculty and students. This would be considered very strong evidence when faced with decisions about possible cancellations. As Librarian 8 commented: “I can’t quite think of a way to assess a resource without usage statistics” (Librarian 8, diary). Echoing this, Librarian 5 noted: “From my perspective, I need to be able to support positions for or against purchases, cancellations, etc. I tend to base these on usage stats and acknowledge this” (Librarian 5, diary).

 

However, while participants used this type of evidence in their decision making and were frustrated if it was not easily available, they also pointed out that such information could not be used in isolation since there are limitations to relying on such data. Participants emphasized that data and statistics were only one part of the story, and that context and other forms of evidence were also required before making a final decision.

 

Local Research and Evaluation

 

Academic librarians frequently incorporate evaluation and assessment of services into their work. Many also take on research projects that are connected in some way to the work they do. While empirical research projects may be more scientifically rigorous, this type of work is usually not undertaken as frequently as local evaluations of projects or teaching. Such evaluation is a source that academic librarians find useful in the ongoing improvement of their services. For example, when referring to instruction decisions, Librarian 7 stated, “I find, probably, evaluations are the most – the best evidence that we have” (Librarian 7, interview).

 

Sources in this category that were cited by participants include total market surveys such as LibQUAL, university surveys that include the library, time audits to measure workload, staff surveys to generate feedback on workload, in-house surveys, testing how something works, evaluation of instruction, SWOT analysis, workplace climate surveys, individual research projects, pre- and post- assessment instruction surveys, and Web usability testing.

 

Such tools are useful to academic librarians who want input from the communities they serve, or from the staff that work at an institution. For example, Librarian 8 had looked to the literature and discussed the situation with her colleagues, but still did not feel that she had all the evidence required to make her decision about a reference project. She concluded: “I’m convinced that I need to hear the voices of actual users. So, I’ve planned to undertake 3 focus groups next week” (Librarian 8, diary)

 

Other Documents

 

This category includes non-scholarly sources that participants used, such as job postings, position descriptions, brochures, mandate documents, safety standards, collection policies, websites (particularly those of other libraries), collective agreements, internal procedure documents, blogs, Twitter, and consultants’ reports. These types of documents are not scholarly or research based, but they provide pertinent information that may be useful in making decisions. For example, policy and procedure documents will guide what librarians decide in order to conform to the goals of the overall mission of the institution: “Is the decision consistent with our policies and procedures?(Librarian 2, diary).

 

Overall, this category of evidence is a broad one, ranging from the official publications of a university, to those documents that are “on the fly” as pointers or tidbits of information, from sources such as Twitter. Despite this, all these types of “other documents” are a source that librarians draw upon, and are relevant depending upon the situation.

 

Facts

 

Facts are what the majority of people, if not all, agree to be true. In academic librarianship, some of the things that can be placed in this category include the cost of products, physical condition of materials, citation or publication information, what items are in the catalogue, license terms, the amount of physical space available, and hours of operation. Facts are generally not disputed, although they may be occasionally. Academic librarians use facts in their decision making in order to place certain realities around the decision, or to verify details before making a decision. For example, if a library has a $10,000 budget for a new resource but it costs $15,000, the fact of the budget amount in conjunction with the cost of the project may alone determine the decision (unless one or both are negotiable). Another example would be deciding when to keep or cancel a subscription:

 

Checking the catalogue record confirmed: we have only a couple of issues of either publication – with so few issues, I questioned the usefulness of having them in the collection at all; they are not available electronically, they are not indexed, one of the titles appears to be the continuation of another title – which we do not have. (Librarian 6, diary)

 

Soft Evidence Sources

 

Input from Colleagues

 

Advice, feedback, and information from colleagues about a program or service are very common sources of evidence for academic librarians. Almost all participants mentioned this as part of their decision making, whether they conceptualize it as evidence or not. “Colleagues” were generally considered to be other librarians, but this was not always the case. Getting input from colleagues, both from within and outside their institutions, provides academic librarians with a way to learn from others who have more experience in a particular area. It also provides confirmation of direction and support for the decision. This type of interaction combines the evidence of experience and knowledge with factors relating to the politics of the institution. It gives the librarian a sense of what other librarians do, and becomes a confirming experience. For many, it is also a way to obtain different viewpoints from one’s own, ensuring that the full picture is considered:

 

I never want to sort of leave something with just my opinion. I want to see if I can find a couple of other varying opinions to inform what I’m doing. So, maybe it is evidence that informs me because at that point once there is an absence of anything that’s documented, I still think it’s valuable to then go and talk to peers or experts. (Librarian 4, interview)

 

Ways of gaining such input from colleagues include one-on-one conversations, attending conference presentations, asking someone to critique teaching or writing, networking at group events (including conferences), corresponding via email or phone, and getting informal feedback from a number of people. This is usually undertaken in conjunction with other forms of evidence (hard sources), but this type of input is considered very valuable for providing insights and knowledge that cannot be gained from the more concrete sources of evidence. Hence, combining what is found in the literature, or what statistics demonstrate, with the professional experience of colleagues puts other sources of evidence in context, provides insight, and highlights any potential problems.

 

Tacit Knowledge

 

As explained in the literature review, tacit knowledge refers to what we know but cannot easily explain (Polanyi, 1966). Participants mentioned their own professional knowledge acquired via experience and education, professional judgment, intuition, and reflection as elements they draw upon to guide their decision making.

 

Participants generally combined this type of evidence with other sources in order to strengthen and verify their decision making, but do not discount their knowledge and experience as irrelevant. As Librarian 6 reflected in her diary:

Now I’m finding – as a result of more experience, confidence, knowledge, maturity – how important those initial gut reactions/instincts are and I’ve learned how to trust them and work with them and pay attention to them – however insignificant that may be. I’ve learned to bracket those instincts and look to the evidence – but in a way that is realistic and appropriate to the situation/question/issue. (Librarian 6, diary)

 

The academic librarians in this study used tacit knowledge very heavily in their decision making. This is evident in the number of references to tacit knowledge that arose in both the diaries and interviews. What is interesting is that tacit knowledge reveals itself when participants describe how they made decisions and the sources upon which they draw, but when they are directly asked what they consider to be evidence, tacit forms of knowledge are rarely mentioned. Most librarians combine the tacit knowledge aspects of what they know as individual professionals and use it in conjunction with external evidence in order to make decisions.

 

Feedback from Users

 

Obtaining feedback from library users arose in this study as a minor source of evidence. When it is more rigorous (as part of a study or planned evaluation), it can be placed in the category of local research and evaluation, which usually focuses on users of a service. However, it is included here as the individual feedback that librarians receive on products or services. This type of feedback is used most frequently in collections management, and also teaching and instruction activities. Faculty feedback that is related to collections is most often looked favourably upon as a source that holds a great deal of weight in decision making.

 

Student feedback is also important to academic librarians, particularly as it relates to information literacy instruction, since librarians want to ensure they are helping the students be successful. In addition to formal evaluations, the informal feedback received following an instruction session is a valuable tool for reinforcement or as an indication that something needs to change. It may result in changes being made to a presentation or style of teaching for the following session.

 

Anecdotal Evidence

 

Anecdotal evidence is “information obtained from personal accounts, examples, and observations. Usually not considered scientifically valid but may indicate areas for further investigation and research” (Jonas, 2005). Most academic librarians would not include this in a conceptual discussion of what they consider to be evidence; however, it is a source of evidence that is often drawn upon when making decisions. Librarian 16 mused about the usefulness of anecdotal evidence in relation to a collections and access issue:

 

I guess even anecdotal evidence can be – to look at where it confirms or differs from available evidence and then go from there and try to figure out what’s happened and why; why all the librarians think everybody wants to have circulating current issues of journals and there’s no evidence showing that people are asking for this. (Librarian 16, interview).


Table 2

How Academic Librarians Find Evidence

Method

How

Examples

Pull

Proactive and specific

Literature search in databases; Google (Internet) search; gathering statistics for circulation or journal usage; looking up facts; asking colleagues questions related to their experience or sources of information

Push

Passive, general awareness

Notifications via TOC services; Twitter; RSS feeds; attending conferences and listening to presentations; colleagues passing on information; getting feedback from users; anecdotal evidence (hearing stories)

Create

Proactive and specific

Including evaluation with instruction; doing a research project related to the problem; conducting in-house surveys or focus groups; keeping reference statistics

Reflect

Proactive examination of knowledge and experience

Carefully considering context and what is known about the situation; tacit knowledge (unique for each person)

Serendipitous discovery

Passive, by chance

Coming across an article or some other document or piece of evidence that is related to your decision, even though not directly looking for it (for example, picking up a journal and while flipping through it, finding something relevant); seeing something in the news that points to a source that is relevant

 

 

Anecdotal evidence may be the prompt that sets investigation of a potential problem into motion, and it is often used in group conversations when determining a course of action. This type of evidence is most frequently frowned upon as not being worthy, but in the absence of anything else, it is certainly used. Most often, librarians will look to other sources of evidence to confirm or deny anecdotal evidence; as Librarian 15 points out, “anecdotally I know about things like that. But you know, having some actual evidence would be helpful” (Librarian 15, interview).

 

How Academic Librarians Find Evidence

 

Data from the diaries and interviews was also coded according to how the participants obtained the evidence they used to make a decision. This coding resulted in five categories relating to how academic librarians find evidence when faced with a problem or question related to practice. The examples in Table 2 come directly from the participants’ actions, and the grouping of these into broader methods of information finding was done by the researcher.

 

The first and most obvious method of finding evidence to help with decision making is what is known as pulling the information required from various sources (“pull”) (Cybenko & Brewington, 1999). This is a very proactive way of obtaining information, and allows librarians to be specific about their needs. As Librarian 4 commented: “I searched, I looked, I asked” in her quest to locate evidence. Doing a literature search is a well-known way of pulling evidence on a particular topic. Other ways of using the pull method would be searching Google (Internet), gathering statistics for circulation or journal usage at the point of need, looking up facts, and asking colleagues questions related to their experience. While discussing the management of approval plans with a monograph vendor, Librarian 15 commented:

 

I also use the vendor’s database site so I can see what the effects of adding a particular variable to a search would be. For example if I want to see how many slips would be received annually by our Education selector in the LC section G73 (Geography – study and teaching) I can run a search for that LC class, limiting it to appropriate readership levels and one calendar year. This way I can determine whether or not the slips are appropriate in content and if the number of slips is reasonable. (Librarian 15, diary)

 

A passive way of obtaining evidence is to have it pushed to you (“push”) (Cybenko & Brewington, 1999). Setting up table of contents alerts or RSS, following individuals or organizations on Twitter, attending conferences, and listening to presentations are all ways in which evidence sources are pushed to librarians. Since these sources are not the result of a specific search for information on a topic, much of what the librarian receives and filters through may not be directly relevant to the problem at hand, but often such sources provide an early indication of trends or aspects of practice that are changing, or new innovations. As one participant noted: “I have a lot of notifications coming over my desk so I see what sort of the trends are typically in the field so I feel like there are lots of things to learn” (Librarian 10, interview). Upon learning of new things via this method, an academic librarian may then further move to the pull method for more information.

 

Academic librarians also create their own evidence sources. This is very proactive and is usually in reaction to addressing a specific need. It includes situations where librarians conduct research or evaluation in relation to their work. Some examples are including formal evaluation with instruction, designing a research project related to a problem, and keeping reference statistics so that trends in the use of reference service can be monitored over time. Evidence sources that are created are generally used in-house for local decision making, but may also be published and fed back into the evidence base used by others:

 

We – library administration – are looking for ways to improve productivity, efficiency and engagement within the unit, and are considering adding an additional layer of supervision to the existing structure. It has been challenging getting enough staff members to participate in frank discussion on the topic, and to articulate what they see as the major areas in need of improvement in the area. To help with this, we administered a survey to staff which yielded some helpful qualitative evidence with respect to how staff members view a variety of issues within the area and how they might be improved. Opinions we suspected might be held broadly by staff members ended up not to be, and vice versa, which has helped to crystallize some of the planning initiatives we had in mind. (Librarian 11, diary)

 

Reflection” is another way that academic librarians find evidence, by taking time to carefully consider the problem at hand and draw upon their past experiences and knowledge in relation to the problem. Considering the context of the problem, and what a librarian knows about the circumstances and people involved, is often very important for how to best approach a given situation. Schön (1983) argues that such reflection allows practitioners to better deal with situations that are uncertain or unique. Reflection on what is done, and how, strengthens the soft forms of evidence discussed earlier:

 

I like to reflect, you know, when I’ve gathered the evidence I like to reflect, depending on how complex the situation is. But I’m finding more and more that taking some time to reflect is extremely useful and whether that’s – even if that’s half an hour or overnight, I like to give myself time to think about all the evidence that I’ve collected and let it ruminate, let it kind of come together and it helps me with seeing a direction. It helps me if I miss anything. You know, have I missed anything, or misread anything? Because sometimes I’ll go back again to the evidence and look at it again and then I realize oh, actually this person said this and I took it to mean this, but actually now that I read it again I see that it means this.

This changes things. So I’ve found that to be very useful, that reflection as part of the evidence. (Librarian 6, interview)

 

A final way that academic librarians find evidence is by obtaining it serendipitously. “Serendipitous discovery” happens almost as if by accident, when librarians find something they weren’t expecting to find as a pleasant discovery. Foster and Ford (2003) conclude from their research that “serendipity would appear to be an important component of the complex phenomenon that is information seeking” (p. 337). In the case of academic librarians this may mean coming across an article or some other document or piece of evidence that is related to a decision, even though they were not directly looking for it. Such discovery is passive, although subconsciously one may be looking for things that relate to the problem at hand. Librarian 3 titled one of her blog posts “Serendipity!” and went on to state:

 

I knew that ACRL had Guidelines for Instruction Programs in Academic Libraries but I also knew that they are fairly out of date – 2003. I was just reading the latest issue of College and Research Libraries News (usually they sit for months on my desk before I have get to them but for some reason I opened the February 2011 issue) and I see that they have updated draft guidelines out! I looked at the ACRL site, and they also have a new draft of Characteristics of Best Practices of Programs of Information Literacy! These are going to be very useful as we figure out what to do with our program. (Librarian 3, diary)

 

Discussion

 

Evidence Sources

 

This study showed that there are benefits to both broad types of evidence that were identified. Hard evidence sources are generally more scientifically rigorous; they confirm or add to what librarians may already know based on past experience and professional knowledge. They also increase confidence, and other people place more value in hard sources of evidence. Hard evidence can be used for convincing purposes, and ultimately increases the depth of professional knowledge. Soft evidence sources are also important; knowledge and experience allow librarians to judge situations and make quick decisions when necessary. Soft evidence enables the necessary analysis and reflection on hard evidence sources, and facilitates putting problems into context.

 

It is important to consider whether both types of evidence are equal and whether soft types of evidence should really be considered valid evidence. This study showed that both types of evidence were used and valued by academic librarians. However, it was only the hard evidence sources that were truly thought of as evidence by participants. This makes sense, as many of the soft sources of evidence stem from already-acquired internal knowledge; evidence is viewed as something that is external and gathered as proof to assist with solving problems and making decisions. For evidence based practice, which seeks to apply the best documented evidence, the evidence focus turns to the hard sources of evidence, which need to be gathered and critically evaluated. EBLIP must also remember the role of the soft evidence, however, and note its importance.

 

Evidence sources vary depending on the type of problem. For example, as Agor (1989) and Dane and Pratt (2007) point out, there are situations when expert intuition is useful and best used. These include situations with significant time pressures and high uncertainty, in which a quick judgment needs to be made. In these situations, consulting an experienced practitioner (expert) in the field is best to make the decision, and intuition can be effective. Such scenarios occur in libraries when there is an emergency situation, a problematic patron, or a difficult human resource issue, to name a few examples. Decisions have to be made quickly and the soft sources of evidence very much come into play by helping librarians make good decisions in such circumstances. However, for decisions that are more planned and have time for investigation, the soft evidence offers a basis of knowledge from which to work and assist with the process of decision making. In these cases, the librarian would use the hard evidence sources to develop a more complete picture based on data, facts, and research in order to come to a logical conclusion about the best decision. The evidence sources used would be those that are most appropriate depending on the question. For example, in the case of designing an information literacy service for a university, the group working on the strategy would look to the research literature, seek out articles about what other institutions have done, examine any past information literacy evaluation that had taken place at the institution, consider learning outcomes tied to the curriculum, talk with faculty, and so on. Many sources of evidence would be weighed to enable the team to come to a decision on the best way to provide service in that particular library.

 

This study confirms that in academic librarianship, the forms of evidence are much broader than just research. Both soft and hard evidence sources are used in conjunction, bringing together the science and the art of practice. The art of the craft allows librarians to embrace messy situations, find ways to be creative, and put professional judgments to use in order to find the best solutions to meet the needs of individual users. This is achieved by applying the best of what is found in the research literature together with the best of what practitioners know is likely to help a person. The science allows for certainty and   confirmation, and builds the overall knowledge base.

 

The findings show that research is valued by academic librarians and is used as an evidence source in decision making. However, academic librarians do not automatically assume that research is good or beneficial just because it has been published. They look at research with skepticism and want to ensure that the research is applicable to their own situations. The research literature alone rarely provides specific answers to the questions that practitioners have. It is almost always used in conjunction with other forms of evidence, including soft sources such as professional knowledge and intuition. Librarians also incorporate other evidence sources such as statistics, local research and evaluation, and input from colleagues, in order to look at many variables prior to making a decision.

 

Implications for Evidence Based Library and Information Practice

 

While the definition of EBLIP noted earlier (Booth, 2000) includes professional judgments, it does so only in a way that indicates that application of evidence to those professional judgments will improve them. It does not clearly account for the place of professional knowledge, nor is professional knowledge accounted for in the EBLIP model. LIS professionals must reconsider this exclusion. Based on the findings of this study, it is clear that professional knowledge and evidence sources are used together, and they are important aspects of the decision-making process. If broadly interpreted, the EBLIP definition covers much of what this study has found to be used by librarians in their decision making, but has a specific focus on research. The concept of “evidence” should be broadened to include more than the traditionally recognized research article (Figure 1). EBLIP should include other types of data and recognize local circumstances. Being “moderated by user preferences” is an important part of the definition, but is rarely explored in the EBLIP literature. User preferences are necessarily local and can be found through the evidence sources of usage statistics, feedback, local evaluation, research, and even anecdotal evidence.

 

 

Figure 1

Evidence sources in librarianship

 

 

While the need to produce high-quality research that is applicable to practice remains (and this goal of the EBLIP movement should in no way be discouraged), this study shows that there are other forms of evidence beyond research that are also necessary for librarians to make decisions in their daily practice, regardless of the quality of the research literature. Many professional librarians’ questions require local sources of evidence that cannot be obtained from the literature. For example, if the problem or question relates to reference service, then reference usage statistics should be considered, as should local feedback and potential local service evaluations. The EBLIP model should account for these as legitimate sources of evidence and should provide assistance for librarians in determining the best way to use these sources, similar to critical appraisal tools that have been developed for research articles. The EBLIP movement needs to discuss and debate the topic of what counts as evidence and how librarians can weigh different forms of evidence. In the future, EBLIP could focus on how to do better project evaluations, how to interpret user statistics, the best methods for collecting reference statistics, and so on. EBLIP was built on the EBM model, but in LIS many different forms of evidence are used that also need to be considered.

 

As noted, research found in the literature is often not directly relevant to the situation at hand. Input from colleagues provides confirmation and support from those who know the local situation and the nuances of why things may or may not work within a specific context. Hence, both aspects are important in academic librarians’ decision making. This is in keeping with the literature of practice-based evidence which stresses the importance of soft evidence sources. The same can be seen in other professions. In health care, for example, Gabbay and LeMay (2011) found similar results in their ethnographic study on the acquisition and use of knowledge by health care professionals. They developed the concept of “mindlines” and observed that judgment and “knowledge-in-practice-in-context” (p. 65) are essential. The mindlines concept demonstrates the importance of skills and knowledge beyond what is found in the research literature, and its contribution to decision making.

 

A model of EBLIP could take a holistic view of evidence, including that which is driven by practice as well as research. Proponents of EBLIP should consider how evidence may be used in practice, and tie research and practice together rather than separating them. A first step is to recognize that what practitioners do is of utmost importance. Obviously, without the practitioner, there is no practice, and practitioners are the ones who know what is happening within their contexts. Practitioners use and create evidence through the very action of their practice. The local context of the practitioner is the key, and research cannot just be simply handed over for practitioners to implement. Practitioners can use such research to inform their decisions but need to consider other components. The concepts found in practice theory, focusing on the practitioner and their knowing in practice – both local evidence and professional knowledge – help to provide a more complete picture of decision making within our profession. The importance that participants placed on learning about what other libraries do, and the high emphasis on gaining input from colleagues, show that practitioners are working within communities of practice for enhancement of their own knowledge and for reinforcement before moving ahead with new ideas. A community of practice may exist within the workplace, where local context is very important, or at a broader level amongst colleagues at other institutions. This broader community is built through conference attendance, as well as committee work on issues of shared interest, and references from colleagues.

 

Future Research

 

It would be beneficial for LIS researchers or researcher-practitioners to explore and recommend the best evidence sources based on the type of question. This would not be a hierarchical list, but would serve as a guideline on what sources of evidence librarians should consider consulting for a given type of question. For example, for a collections problem, the research literature should be consulted, but other sources of evidence that would provide good information include usage statistics for e-products, circulation statistics, faculty priorities, tools such as OCLC collection analysis, interlibrary loan and link resolver reports, and the publication patterns of faculty. Researchers could determine the most relevant sources for each area of practice, and in what circumstances they are best used.

 

It would also be very beneficial for practitioners if researchers would develop guidance on how to read the results of different evidence sources. This could include what practitioners need to consider when looking at reference statistics, or what elements librarians should consider when conducting an evaluation of their teaching. Some of this information will be found in existing literature, and a scoping review of what has already been documented would be a good start.

 

Limitations

 

This study is not intended to be generalized to all academic librarians. The purposeful sample allowed for depth and richness of information, and saturation in the data was reached, but not all academic librarians would necessarily fit within these findings. In addition, other academic library systems outside of Canada may operate differently. Academic librarians are generally regarded as academics or faculty in Canada, and at many institutions they can obtain tenure. These factors may create a very different work environment and professional outlook from those working in other library sectors. Doing similar research on other librarian groups would strengthen the key findings and applicability of this study.

 

The data collection methods included diary keeping by the participants for a period of one month. The very act of having to keep the diary was something that was not a normal part of their practice, and thus may have impacted their behaviour. For example, they may have felt pressure to do more and be more methodical in their decision-making processes than normal. It is unlikely that false reporting occurred, however, since the follow-up interviews with participants allowed for in-depth probing of the actual decision-making process, confirming what was previously reported.

 

Conclusion

 

This paper has detailed research findings regarding types of information that academic librarians consider to be evidence, and the evidence sources that they use in practice. It answers the research questions, “What forms of evidence do academic librarians use when making professional decisions? Why do they use these types of evidence?” Two broad types of evidence were identified (hard and soft), which are generally used in conjunction with one another in order to ensure that all possible evidence sources applicable to the problem at hand are considered. Neither type of evidence is sufficient on its own. Librarians look at all evidence sources (hard and soft) with a critical eye, and try to determine a complete picture before reaching a conclusion. Information about how librarians find evidence emerged from the data, showing that both proactive and passive approaches are used.

 

This paper also answers the research question, “How do academic librarians incorporate research into their professional decision making?” It is clear that academic librarians do value research and do look for it to assist with their decision making. However, the published research is insufficient on its own. It may not be directly applicable, and the specifics of the question or problem which librarians are trying to solve take them to sources beyond the research literature. Librarians value research literature, but do not use it in isolation. It is only one part of the overall evidence that a librarian needs to consider.

 

Both hard and soft types of evidence instill confidence but from different perspectives, and taken together have the most strength. These results provide a strong message that no evidence source is perfect. As a result, librarians bring different types of evidence together in order to be as informed as possible before making a decision. Using a combination of evidence sources, depending upon the problem, is the way that academic librarians approach decision making. These results suggest that current practice does not fit with the most commonly used definition of EBLIP or the EBLIP model as noted in the literature. A change within EBLIP does not require a full rejection of the name, but rather a realization that more types of evidence can be included within the concept of evidence, and that doing so brings the EBLIP model closer to one that has truly considered the needs of librarians.

 

Acknowledgement

 

This paper is the first from a doctoral study. Future papers will look at how evidence sources are used in decision making, obstacles and enablers to evidence based decision making, and a fuller consideration of possible changes to the EBLIP model itself.

 

 

References

 

Agor, W. H. (1989). Intuition in organizations: Leading and managing productively. Newbury Park, CA: Sage Publications.

 

Banks, M. (2008). Friendly skepticism about evidence based library and information practice. Evidence Based Library and Information Practice, 3(3), 86-90.

 

Booth, A. (2000, July). Librarian heal thyself: Evidence based librarianship, useful, practical, desirable? 8th International Congress on Medical Librarianship, London, UK.

 

Booth, A. (2010). Upon reflection: Five mirrors of evidence based practice. Health Information and Libraries Journal, 27(3), 253-256.

 

Brophy, P. (2009). Narrative-based practice. Surrey, UK: Ashgate.

 

Charmaz, K. (2006). Constructing grounded theory: A practical guide through qualitative analysis. London: Sage.

 

Clark, C. (2011). Evidence-based practice and professional wisdom. In L. Bondi, D. Carr, C. Clark, & C. Clegg. (Eds.). Towards professional wisdom: Practical deliberation in the people professions. (pp. 45-62). Surrey, UK: Ashgate.

 

Corbin, J., & Strauss, A. (2008). Basics of qualitative research: Techniques and procedures for developing grounded theory (3rd ed.). Thousand Oaks, CA: Sage.

 

Crumley, E., & Koufogiannakis, D. (2002). Developing evidence based librarianship: Practical steps for implementation. Health Information and Libraries Journal, 19(4), 61-70.

 

Cybenko, G., & Brewington, B. (1999). The foundations of information push and pull. In G. Cybenko, D. P. O’Leary, & J. Rissanen (Eds.). The mathematics of information coding, extraction and distribution. (pp. 9-30). New York: Springer-Verlag.

 

Dane, E., & Pratt, M. G. (2007). Exploring intuition and its role in managerial decision making. Academy of Management Review, 32(1), 33-54.

 

Davies, H., Nutley, S., & Walter, I. (2008). Why “knowledge transfer” is misconceived for applied social research. Journal of Health Services Research, 13(3), 188-190.

 

Duguid, P. (2005): “The art of knowing”: Social and tacit dimensions of knowledge and the limits of the community of practice. The Information Society: An International Journal, 21(2), 109-118.

 

Eldredge, J. (2002). Evidence-based librarianship levels of evidence. Hypothesis, 16(3), 10-13.

 

Eldredge, J. D. (2000a). Evidence-based librarianship: An overview. Bulletin of the Medical Library Association, 88(4), 289-302.

 

Eldredge, J. D. (2000b). Evidence-based librarianship: Searching for the needed EBL evidence. Medical Reference Services Quarterly, 19(3), 1-18.

 

Evidence. (2010). In Oxford Dictionaries Pro. Retrieved 6 Dec. 2012 from http://english.oxforddictionaries.com/definition/evidence.

 

Foster, A., & Ford, N. (2003). Serendipity and information seeking: An empirical study. Journal of Documentation, 59(3), 321-340.

 

Fox, N. J. (2003). Practice-based evidence: Towards collaborative and transgressive research. Sociology, 37(1), 81-102.

 

Gabbay, J., & Le May, A. (2011). Practice-based evidence for healthcare: Clinical mindlines. New York: Routledge.

 

Given, L. (2006). Qualitative research in evidence-based practice: A valuable partnership. Library Hi Tech, 24(3), 376-386.

 

Guyatt, G. H. (1991). Evidence-based medicine. ACP Journal Club, 114(Mar.-Apr.), A-16.

 

Howick, J., Chalmers, I., Glasziou, P., Greenhalgh, T., Heneghan, C., Liberati, A., Moschetti, I., Phillips, B. & Thornton, H. (2011). The 2011 Oxford CEBM levels of evidence (introductory document). Retrieved 7 Dec. 2012 from http://www.cebm.net/mod_product/design/files/CEBM-Levels-of-Evidence-Introduction-2.1.pdf

 

Hunsucker, R. L. (2007). The theory and practice of evidence-based information work – One world? Paper presented at EBLIP4: Transforming the Profession: 4th International Conference, Evidence-Based Library & Information Practice, University of North Carolina-Chapel Hill, Durham, NC, USA. Retrieved 3 Mar. 2012 from http://www.eblip4.unc.edu/papers/Hunsucker.pdf

 

Jonas, W. B. (Ed.). (2005). Mosby’s dictionary of complementary and alternative medicine. New York: Elsevier

 

Koufogiannakis, D. (2010). The appropriateness of hierarchies. Evidence Based Library and Information Practice, 5(3), 1-3.

 

Lave, J., & Wenger, E. (1991). Situated learning: Legitimate peripheral participation. Cambridge: Cambridge University Press.

 

McKibbon, K. A., Wilczynski, N., Hayward, R. S., Walker-Dilks, C. J., & Haynes, R. B. (1995). The medical literature as a resource for health care practice. Journal of the American Society for Information Science, 46(10), 737-742.

 

Polanyi, M. (1966). The tacit dimension. Garden City, NY: Doubleday & Company, Inc.

 

Rolfe, G., Jasper, M., & Freshwater, D. (2011). Critical reflection in practice: Generating knowledge for care (2nd ed.). London: Palgrave Macmillan.

 

Ryle, G. (1945). Knowing how and knowing that: The Presidential address. Proceedings of the Aristotelian Society, 46, 1-16.

 

Schatzki, T. (1996). Social practices: A Wittgensteinian approach to human activity and the social. Cambridge: Cambridge University Press.

 

Schön, D. A. (1983). The reflective practitioner: How professionals think in action. U.S.A: Basic Books.

 

SUNY Downstate Medical Center. (2004). The evidence pyramid. Retrieved 20 Jan. 2012 from http://library.downstate.edu/EBM2/2100.htm

 

Usher, R., & Bryant, I. (1989). Adult education as theory, practice and research: The captive triangle. London: Routledge.


 

 

Appendix A

Blog Diary Instructions for Participants

 

Over the course of the month that you keep this diary, please write about any incidents where questions arise relating to your professional practice as a librarian. Questions/problems could vary widely. Please make note of each question, your thoughts about it and how you might approach solving the question.

 

Explain any action you took to answer the question, and what, if anything, you did about it. Some questions may be answered immediately, while others may take days or weeks, or not be answered during the diary-keeping period at all. That is ok. Just detail as much of the process you used in your decision making as possible. How did you come to make the decision you did?

 

At any point in the process, please feel free to reflect on the decisions you made and whether they seem to be working. Remember, there are no right or wrong responses. As a researcher, I am looking to understand the process that academic librarians go through in reaching decisions, and what types of evidence may be part of that decision.

 

Examples of professional questions/problems a librarian may be working on:

o    Today I am deciding which print journals we might be able to safely weed from the collection.

o    I’ve been asked to determine the most appropriate hours of operation for the fall term.

o    We are going to be doing renovations to the building this year and I’m on a team looking into what changes would be best.

o    I am planning a one-hour information literacy session for first year biology students and am trying to determine the best method of delivering the information.

o    I was wondering how Scopus journal coverage compares to that of Biosis. Do we need both?

 

Key elements to include in your diary blog entry:

o    The professional question/problem arising in practice.

o    Things you did in working through the question/problem. What types of evidence did you use, if any? Who or what did you turn to in this process to help you?

o    Any roadblocks you encountered in your problem-solving process, and what you did as a result.

o    The end result/outcome if a conclusion was reached; or, steps you plan to take to reach a conclusion.

o    Reflection on your decision making process. How do you feel about what you did; what would you change? Were your sources of evidence sufficient?

 

Please write in your diary as professional practice questions occur. If no entries are received within the period of one week, you will be prompted with a reminder by the researcher. You may contact the researcher at any point, to either ask questions, or drop out of the study if you wish. There is no obligation on your part to participate, all participation is voluntary, and there are no repercussions for dropping out of the study. The blog you are using is private, so only you and I can access or read the content.

 

Denise Koufogiannakis

dkoufogi@gmail.com; dak@Ualberta.ca

780-432-3427 (Home)

Skype: dkoufogi

 

 

Appendix B

Interviewer’s Guide

 

Guiding questions (to be adapted to each situation and allowed to flow from the context of what the participant feels is important to discuss):

 

o    Thinking back on the diary keeping period, were their any specific incidents that stood out for you?

o    Would you say that the things you recorded in your diary were fairly typical of a normal month for you? Why/why not?

o    In your diary entry, one of the things you discussed was [X]. Can you tell me more about your thoughts on this and what it means to you as an academic librarian?

o    You mentioned a question that arose in your practice (name the specific question/incident). Can you tell me about this in a bit more detail?

o    What were some of the barriers or difficulties you encountered during the diary-keeping period?

o    What types of things do you consider to be ‘evidence’?

 

Can you tell me how important or not research is to you as a practitioner? How do you use research? What do you consider to be good research, and how do you use it?