Libraries and Patron Data…it’s complicated

As I progress through a course on learning analytics, I’ve been digging deeper into the relationship that librarians have with patron data, and I’ve determined that, like so many troubled relationships, it’s complicated. While the core value of patron confidentiality is now considered unquestioned and, indeed, inviolate, it is not as historic as some may believe. It was added to the ALA’s “Library Bill of Rights and Code of Ethics” in 1939, nine years after the initial release, at a time that is associated with the rise of Nazism and notable events of government surveillance (Witt 2017). While over 80 years old, this particular ethic has had some troubling beginnings and has taken some serious knocks from both external forces and from within the profession itself.

I also found the placement of this particular code interesting in the ALA Code of Ethics (current) – following customer service and intellectual freedom. Indeed, there are a number of ways that customer service has trumped privacy throughout the years. Consider circulation practices of yore, such as check-out cards with handwritten signatures of patrons who had borrowed each title, easily accessible to any person perusing the shelves. What would be the implications of borrowers of this title following an explosion in a lab if it appeared suspicious?

But I digress…my point is that customer service is the number one ethical obligation, but it is not meant to be at the expense of the other codes. Indeed, these other codes appear to constrain that #1 – guardrails, if you will. But to what extent should these constraints go?

The Prioritizing Privacy project, led by Lisa Hinchliffe and Kyle Jones, of which I have partaken, provides the opportunity for librarians to explore their own concepts and test their own assumptions about privacy and library services, particularly relating to learning and learning analytics. Taking a critical approach to both privacy and Big Data in academia (particularly higher education), the extended workshop brings together individuals at many points along the spectrum of beliefs and attitudes. Particularly intriguing to me were the discussions that brought out the disadvantages of libraries and librarians standing on the sidelines of this issue, not only for the institution of librarianship, but, most notably, for the students themselves, in the loss of services which could help them succeed in their own goals. But the key to ensuring ethical application of such services is the respect of students’ agency, an aspect that even the most ardent opponents of learning analytics can accept as a path to providing that “highest customer service” while still protecting the “patron’s right to privacy and confidentiality”.

I’m still exploring the privacy-analytics issue and have collected what I consider five foundational or key articles that explore the historical development and current struggles librarianship has with privacy (see References below). Witt, as I indicated earlier, provides a look a the intrigues, influences, and events which brought about the current ethos of patron confidentiality and privacy protection within the profession. Of particular note is Witt’s suggestion that the development of the codes of ethics in the 1930’s was less of a concern about the values of the profession and more about the development of librarianship as a profession itself (Witt, pg. 648). Regardless of the motivations, the codes were developed organically from statements in published literature and responses to surveys of ALA members, reflecting overall opinion of librarians, notably that reference transactions should be considered private and libraries should take steps to protect patrons’ privacy of information that they seek (ibid., 649).

Campbell & Cowan’s “Paradox of Privacy” focuses the privacy and service dilemma through the lens of LGBTQ (note: the authors specifically denote the “Q” as “Questioning”, which figures prominently in their discourse). They argue that the need for inquiry and the need for privacy are necessarily integrated, particularly for such a personal journey of the development of a young person’s identity. The title comes from the authors’ statement that, “open inquiry requires the protection of secrets” (Campbell & Cowan, pg. 496). After setting the context of this paradox, they argue for “the need for privacy” and the paradox of individual’s disclosure dilemma – disclosure can be risky and it can be healing. Indeed, the need for privacy within inquiry is imperative for the development of identity, but libraries have had mixed success in dealing with the competing needs of managing inventory with managing privacy. They describe how self-checkout systems have been shown to result in increases of use of LGBTQ material, contrasted with the increase of libraries collecting (or allowing to be collected by third-party systems) “big data” surveillance. Campbell & Cowan summarize Garret Keizer’s treatment and definition of privacy – “‘a creaturely resistance to being used against one’s will.'” – including providing Garret’s observations about privacy and libraries:

  • Surveillance—monitoring what people are reading and sharing private information about them—becomes a form of using other people.
  • Privacy consists of…individual’s power to modulate the extent of his or her self-revelation in specific circumstances.
  • The library occupies a position of significant though paradoxical importance: its status as a public place makes it an ideal place in which to experience genuine privacy.

My takeaway from this slightly-below-surface-level look at libraries and privacy is that librarianship’s relationship with patron privacy is complicated. We value a patron’s right not only to access quality information on any and all subjects, but their right to keep this information private. However, we also value service and easy access to this information, and methods to improve these services can conflict with the value of privacy. But there are ways to reduce this conflict, mostly by returning control of information back to the patron. Transparency and honesty in how the information will be used, providing patrons access to their stored information, enabling patrons to opt-in and opt-out of these services, and simply letting go of the need to control information and resources are the solutions which I have discovered to have the greatest potential to ensuring trust in libraries is retained.

References

Asher, Andrew. “Risks, Benefits, and User Privacy: Evaluating the Ethics of Library Data.” Chap. 4.2 In Protecting Patron Privacy: A Lita Guide: Rowman & Littlefield Publishers, 2017.

Campbell, D. Grant, and Scott R. Cowan. “The Paradox of Privacy: Revisiting a Core Library Value in an Age of Big Data and Linked Data.” Library Trends 64, no. 3 (2016): 492-511. https://doi.org/10.1353/lib.2016.0006.

Jones, Kyle M. L., and Lisa Janicke Hinchliffe. “New Methods, New Needs: Preparing Academic Library Practitioners to Address Ethical Issues Associated with Learning Analytics.” Paper presented at the The Annual Meeting of the Association for Library and Information Science Education (ALISE), 2020.

Witt, Steve. “The Evolution of Privacy within the American Library Association, 1906-2002.” Library Trends 65, no. 4 (2017): 639-57. https://doi.org/10.1353/lib.2017.0022.

Zimmer, Michael, and Bonnie Tijerina. “Foundations of Privacy in Libraries.” Chap. 2 in Protecting Patron Privacy: A Lita Guide, edited by Bobbi Newman and Bonnie Tijerina: Rowman & Littlefield Publishers, 2017.

Measures & factors of collection evaluation

As I have been developing a model of collection evaluation – that is, examining the features of specific subsets of our libraries’ collections, usually based on subjects – I have been collecting a mental list of aspects on which to assess.  Some of these are obvious and quite traditional: number of titles & volumes (by format and by subject), uses (circulation, e-resource use) and expenditures.  There are also measures of need – number of potential and actual patrons, majors & degree programs, etc.  And then there are gaps in need – ILL requests by program members, lower-than-expected use, etc.

However, it can be quite enlightening to step back and consider what others think are important.  This article is by a librarian relatively new to her field, Christina Wray from Indiana University-Bloomington, and her approach to learning collection development “on the job”.  Three of the “four main challenges” to here job involved “understanding the characteristics of…the current users,…the current collection, and…identifying new trends in the subject area.”  I thought it would be useful to use a practitioner’s approach to check against my list of measures.

Users

  • The department’s “fit” in the “institutional heirarchy”.
    • Currently provide: Description of the department and its hierarchy, relative size of faculty, students and graduates
    • Would be useful: Org chart
  • Degrees and certificates awarded.
    • Currently provide: List of degree programs, trends in degrees awarded, online programs
    • Would be useful: ? (what do you all think?)
  • Online programs
    • Currently provide: List of online programs, relative size of online participation
    • Would be useful: size of truly distant students (versus local students who take online courses)
  • Size of faculty and students
    • Currently provide: Absolute size of faculty, students and graduates, trends in enrollment.
    • Would be useful: ?
  • “Crossover” with other departments
    • Currently provide:  Nothing (hmmmm)
    • Would be useful: Anything…I like this idea because of the emphasis on interdisciplinarity in our newest collection development model.
  • Demographics of department
    • Currently provide: Distribution of faculty by level.
    • Would be useful: Not sure…age? Length of time at university?  Is race or ethnicity important for assessing collection need?  Maybe language…
  • Number of classes
    • Currently provide: Nothing
    • Would be useful: Relative number of classes; ratio of classes taught by differing levels of faculty?
  • Research interests & specialization
    • Currently provide:  List of key topics; a word cloud of research interests
    • Would be useful: Not sure…

While discussing users, Wray mentions defining the collection, notably by call number range.  While this seems a bit oddly placed in the article, her idea is at the heart of the collection evaluation model I’ve been developing.  Our subject-based collections all have fairly specific call number ranges assigned, but the collections are neither exhaustive nor, more importantly, mutually-exclusive.  These “profiles” are compared with the courses offered and the faculty research interests to ensure proper coverage.

Which leads to characteristics of the collection:

  • Defining “subcategories” of the collection
    • Currently provide: List of subjects by call number range
    • Would be useful: visualizations of this “map”
  • Inventory of books & journals
    • Currently provide: Well, given that our profiles are purposefully broad and our library is moderately-sized, such an inventory for most collections would be too large to be worthwhile.
    • Would be useful: Complete list of subject ranges linked to the catalog; journal subjects mapped to the profile; other e-resources mapped to the profile.
  • Use (absolute & relative) of collections
    • Currently provide: Circulation by call number range.
    • Would be useful: database & e-journal usage.
  • Historical coverage of books & journals
    • Currently provide: Distribution of book holdings by publication date.
    • Would be useful: Distribution of journal holdings by coverage dates.

What I found intriguing in Wray’s article was her suggestions on applying the data to answer some fundamental questions regarding the strengths & weaknesses of the collection based largely on the relative distributions of holdings and usage.  She also offers ways to “dig deeper”, including:

  • Comparing the holdings & usage with the faculty research interests & course offerings.
  • Understanding faculty satisfaction and perception of the collection.
  • Applying the data to modifying the collection.
  • Evaluating databases & other resources.

Finally, Wray advocates reviewing usage data once or twice annually.  We are currently developing a “collections dashboard” to provide some basic, actionable metrics, and usage data of specific resources could be one such metric.

While (or perhaps, because) this article is aimed at the subject-specialist librarian new to a field, it provides me, the Collection Assessment Librarian, the issues and factors of concern to the liaisons.

 

(2016). Learning Collection Development and Management on the Job. Collection Management: Vol. 41, No. 2, pp. 107-114. doi: 10.1080/01462679.2016.1164646

Source: Learning Collection Development and Management on the Job – Collection Management – Volume 41, Issue 2

Notes from the OA Symposium, 2016

As I mentioned previously (actually, about 30 minutes ago in real time), it’s been a busy week in Denton.  The University of North Texas hosted two separate meetings on overlapping themes of open access publishing: The Library Publishing Forum and the Open Access Symposium.  I’ve already posted my notes of the former, the meeting of the Library Publishing Coalition.  Now on to OA…

This was actually the 7th annual Open Access Symposium hosted by UNT.  Here is a word cloud from my notes to give you an idea of the topics & ideas covered: OAS2016In addition to the usual suspects of such a meeting (“published” and “journal”), you might have noticed “Communism”.  This is the result of a last-minute addition to the program: a presentation from Alexandra Elkabyan, from Sci-Hub & LibGen, which was Skyped from her home.  Interestingly, this was translated live in part by one of the organizers, Kevin Hawkins (you learn something new about somebody at these things…like, Kevin understands and can translate Russian).  My take on this was that her basic point is valid – that the current market-based environment results in the concentration of information and power, and stifles scientific discoveries.  The only time the audience really reacted, though, was her response to our security concerns of sharing authentication parameters, which she cavalierly dismissed.  It is a protest that on its face seems low-risk, but could eventually raise the stakes.

I have been out of the loop regarding this sub-specialty of librarianship…and it has grown tremendously in the last 5-10 years.  I learned all kinds of new systems, coalitions, foundations, and software.  I became aware of a veritable alphabet soup – OJS, APC, OSF, etc.  The Below are my notes, and when the recordings become available, I’ll add the links.

Open Access Symposium 2016

One key concept brought up at both conferences (but not well represented in my notes, apparently) was transparency.  Indeed, it has been the lack of transparency in the pricing of subscriptions from the “Four Horsemen of the (Scholarly Publishing) Apocalypse” that has forced the solutions discussed at these meetings.

What I found lacking from both conferences were two key issues: assessment and true sustainability.  While the latter was discussed much more than the former, the solutions proposed did not seem to be truly sustainable.  The Author Processing Charges (APC) have become essentially Library processing charges, with libraries contributing to them.  Many start with grants, but most seem to “play on guilt” to encourage participation in consortia.  But the problem of “free riders” (which itself is not a problem until there are too many, and too few providing the support) was only mentioned once or twice in Q&A’s.

An idea I did find intriguing and useful was Johan Rooryck’s regarding the convergence of the roles of publisher & press.  Elsevier did not found the journal, Lingua, as it claimed, but was rather “hired” to provide the services of a press – printing & distribution.  At some point, publisher & press converged and the press-cum-publisher came to own the content.  Unlike music publishers, journal article authors are provided no financial renumeration directly from the publication.  Journals should reclaim the responsibilities of publisher and hire out the services of a press.  Thus, they can retain control over charges, which should be based solely on the production costs.  However, this is an idealistic and perhaps unrealistic prospect.  Like Orwell’s Animal Farm, the proclamation that all journals are equal (essentially in the APC’s) could evolve into the corollary, but some journals are “more equal than others.”  Prestige could figure into the APC’s that journal editorial boards (now the publisher) charges, regardless of the production costs.  This was not well addressed.

Despite the issues above, I do believe that the times are a’changin’ and the market will adjust.  How slowly or how quickly will depend on the motivations of the consumers and  the buyers.  Once these align, change will come.

Notes from the Library Publishing Forum

It’s been a busy week at UNT – host of two meetings: The Library Publishing Forum and the 7th annual Open Access Symposium.  The dates, and themes, are overlapping – openness, transparency, shifting paradigms of ownership of scholarly communication.  The Library Publishing Forum (I’ll use LPF for brevity) included a half-day workshop on Open Educational Resources (OER – word of warning…the field is rife with abbreviations), thus bringing education & pedagogy into the fold.

Here is a word cloud of my notes of both meetings…LPF_and_OA_2016

common words jump out, like “Publish” and “Journal” and “Library” and “University”.  Notice other words… “Communism” (more about that shortly), “Open,” “Sustain,” and “Go” (I like that last one in particular).

Now, looking solely at the LPF, other words appear more distinct: LPF2016Specifically, “Faculty” and “Textbook” and “Student”.  This is the effect of the emphasis on OER’s.  The workshop was scheduled due in no small part to the number of proposals submitted on this topic.  So it reflects true growing interest in the community.

It is also interesting to me the geographic representation of the attendees.  The LPF is the meeting of the Library Publishing Coalition (LPC – I warned you).  The members are still primarily American academic institutions, although there is growing internationalism.

If you are interested, the notes that I took have been exported from OneNote into a PDF – another warning: it is long and somewhat annotated.  Recordings will be available shortly – when they are, I will post the links.

Library Publishing Forum 2016

 

Assessing the evolving library collections

Lorcan Dempsey from OCLC® Research recently revisited their 2014 publication on library collections (Collection Directions: The Evolution of Library Collections and Collecting), emphasizing the “facilitated collection” aspect of their report.  This refers to the idea that the networked environment reduces the need for libraries to maintain external resources (the “outside-in” style of physical collections).  Instead, facilitated collections are “a coordinated mix of local, external and collaborative services are assembled around user needs,” including selected (perhaps, indeed, “curated”) lists of external or freely-available resources, the shift from “just-in-case” to “just-in-time” selection, and the greater reliance and participation in shared collections.

The report’s take on the continuum of “shared” collections is interesting.  This ranges from the “borrowed collection” (what we typically consider resource sharing networks with their expedited deliveries), to “shared print” (which we typically associate with collaborative collections), to “shared digital” (think, HathiTrust), and onto the “evolving scholarly record”, which its increase in sharing of not only the “final outcomes” of research, but also the intermediate products (e.g. data sets, working papers, preliminary reports, etc.).

fullcollectionspectrum

So, as I always ask when considering models of library collection development, how would such collections be assessed?  Rather than focusing on what is owned or even “available”, when assessing any service, it is best to start at the end goal of the service – in this case, to “meet research and learning needs in best way”.  Indeed, as I realized when I prepared for my interview for my current position (Collection Assessment Librarian), collection development is as much a service to library users as reference and instruction.  And the service is focused on the users, not the items in the collection.

Assessing Collections as a Service

Dempsey refers to this evolution to “collections as a service” as the result of the shift from the “‘owned’ collection” to the “‘facilitated’ collection”, which itself impacts the “organization, stewardship and discovery” of the collections.  I found particularly intriguing the reference to ‘collection strategist’ job advertisement, which noted reflects a shift of emphasis to the “allocation of resources and attention”.  While this may have always been inferred as a responsibility of collection development librarians, the overt references suggest an increase in focus.  Collection assessment provides the information needed to make these “strategic” decisions. Thus, the information that is gathered needs to be directly or indirectly to the end goals of the collection.

The end goal mentioned in this report is to “meet research and learning needs in best way”.  The goals of the individual libraries may be worded differently, but often incorporate similar ideals.  For the purposes of this post, I’d like to break down this goal to determine how best to assess a collection.

“Meet…needs”

What does “meet…needs” mean?  What does it entail?  How could it be assessed?  Key terms I consider include “available”, “accessible”, “discoverable”, “findable”, “usable”, “transformable”.  Most of these remind me of the Four User Tasks associated with the Functional Requirements of Bibliographic Records (FRBR):  find, identify, select, and acquire.  As you can see, though, the word “transformable” goes beyond that fourth task of acquiring, suggesting that the needs of researchers and students is to incorporate the information resources into their activities and transform them into new knowledge.  One simple example is the ability to import citations or references into papers and other outputs.  More complex examples are available in the rising field of digital humanities, where whole texts are available for analysis and transformed into network diagrams.  So, to assess a collection regarding this aspect of the goal, here are some key measures (not comprehensive):

  • Accessibility – Percent of collection that meets ADA requirements for accessibility for all users
  • Availability – Percent of collection that is available online versus via local use only
  • Discoverability – Rate of collection added to the main library search systems per year
  • Findability – Rate of requests for materials owned that are mistakenly requested through interlibrary loan.
  • Usability – Percent of resources that meet usability benchmarks (this should include physical resources, not only Web-based)
  • Transformability – Percent of resources that can be easily incorporated into differing formats and communication modes

“research and learning needs”

While the above measures describe the ability of a library to meet the needs in terms of delivery and accessibility, they do not measure how well the collection meets the needs conceptually.  Formats, subjects, perspectives, sources, depth, scope and breadth are all reflected in this aspect of the goal.  This requires an understanding of what, conceptually, the needs are, which requires an understanding of the work being conducted at the institution, as well as the institutions goals and vision.  An emphasis on curriculum support, particularly in the traditional classroom-lecture sense, would require different sources, formats and depth than an emphasis on original research.  There will most likely be differences in this emphasis between subject disciplines (e.g. a history department with a PhD program, versus a Spanish department that offers only conversational language courses).  This is similarly true with scope – that history department may focus on American history (or even Southern United States), while the Literature department supports a world literature program.  Formats are notoriously tribal, with some disciplines continuing to rely on more traditional formats (print, even microform) and others have nearly completely transformed to digital.

“in best way”

OK, here is the most subjective part of assessment – after all, how can you define “best way”?  Best for whom?  Some of the perspectives include those of the students, the teaching faculty, the researchers, the librarians, the library administration, the campus administration, and the community.   Best in what ways?  Ease (see above notes on delivery), subject (see above notes on needs), financially, efficiency are a few of the key terms I can think of.  The priorities of these should be understood before attempting a comprehensive assessment – a financially-strapped institution may require greater emphasis on efficiency or even raw costs.  And even within each of these factors, how would you determine which way is best?  Cost-per-use is a common measure of efficiency, but which is better – Cost-per-session?  Cost-per-search?  Per record viewed?  We use the measure that most closely meets the user’s needs of the resource.  What about financially?  A resource that costs 50% of the materials budget and an inflation rate of 7% may be highly efficient (CPU less than $10), but is unsustainable for a budget that is flat.  Assessing collections relative to this part of the overall goal requires consensus on the perspectives and priorities of the competing aspects of defining “best”.

Final thoughts

Assessing or evaluating library collections which are constantly evolving requires concepts which can be applied more broadly than traditional methods.  Thus, the number of microforms is a measure that is much less relevant today than it was thirty (or even fifteen) years ago.  Conversely, success rates of users finding the resources needed for their purposes (perhaps measured at the time of visit to a library (physical or virtual)) could be applied as long to an “owned collection” or a “facilitated” one.  Finally, measures should be developed that put the object into perspective or context.  This could be relative to benchmarks set by the library or the field, comparisons with peers, or against the population of users.

Open Data for LIS Research

This blog is a venue devoted to improving the utility and effectiveness of libraries, librarianship, and librarians through careful study.  This inevitably involves data, a word that has recently become hot, indeed, dare I say, sexy?  Data Science has seen tremendous growth in jobs and university programs, and there is a substantial rise in the use of the term on the Web (BTW – I found this really cool tool, Google Trends – more about that in another post).  Over the past few weeks, I’ve noticed several interesting posts and projects regarding data that would be particularly useful for LIS.

Despite years of growth in databases and data management, gathering, sharing and using data is still a difficult job (hence the the demand).  Walt Crawford has been gathering and analyzing data on Open Access journals for the last several years.  After releasing his latest book, he decided to also make the data available for others.  Apparently it was not an easy task.  Even just providing a link to the spreadsheet requires the end-user to have Excel (or a program that can import Excel).  And this is data that has already been “cleaned” – gathered from disparate sources, standardized, organized, and, of course, anonymized.

Also coming to my attention recently is the Open Syllabus Project, about which I first read in the NY Times.  This piece also caught the attention of Joseph Esposito, who wrote about it on the SPSS Scholarly Kitchen.  Like Joseph, I was quite excited about this idea – “A new dataset can lead to new insights…” – but primarily from a collection development/assessment viewpoint.  And I was thrilled to see my institution in the list.  Possible uses that have flitted across my brain include list-checking and peer-comparisons.  Interestingly, this data set is dependent upon other openly-available data sets, notably the Harvard Library Open Metadata and JStor (see the FAQ).

Another, more tangentially-related piece that I saw was a post about systematic literature reviews.  This is a method of research that is very common in clinical medicine, and one about which my medical librarian readers likely know much.  Indeed, many of them collaborate with clinicians as they prepare clinical trials or meta-analyses of clinical trials, and conduct their own studies.  I’ve never understood why this method has not been taken up more emphatically by those in our profession.  True, there are a lot of literature reviews, but the most comprehensive listing of LIS Systematic Reviews, while numerous, is a pittance compared with the traditional style.  The systematic nature of the literature review requires extensive knowledge of the organization of literature in the field of interest, which, of course, is one of the foundations of our field.

All of these postings made me wonder about the availability of data for LIS research.  There are already rich sources of data that are open right now (and have been available for a long time), including: PubMed, WorldCat (this is sort of mixed open & not-so-open), Google (too numerous to link), NCES Library Survey, IMLS’ Public Libraries in the United States Survey.  But there could be more.  Much more.

One of the most frustrating aspects of trying to analyze collections and their use is having limited benchmarks and theoretical bases.  Data is still too inaccessible – locked in silos behind expensive firewalls (e.g. Web of Science, WorldCat Collection Evaluation System, etc.), or within proprietary database systems that are extremely difficult to share (most ILS’s), or within library/institutional administrations not willing to to share (are we really so competitive that we need to keep our funding & expenditures secret?), or perhaps a myriad of other reasons.  Joseph noted that the value of the Open Syllabus project was that “it gathers a great deal of information together in one place, where it can be analyzed.”  This is the value of Open Data.  This project has given me hope for the future.

Emerald | Looking Back, Looking Ahead with Jaeger, Bertot and Hines

Source: Emerald | Looking Back, Looking Ahead with Jaeger, Bertot and Hines

Here are some thoughts about the past year and the year to come in librarianship.  I think the question, “What were the biggest changes in 2015?”, is not terribly appropriate for our field.  Like an ship, librarianship (OK, pun intended) does not turn on a dime.  Even the rather radical change to demand-driven acquisitions has taken ten years to fully unfold.  A potential change may get a lot of attention (and “air” time in the professional media), but the actual change may or may never happen, and years from now, if it does.  I think a better question, and the one that the authors appear to have answered, is, “What has garnered the most attention in librarianship in 2015?”

I’m not sure if you can say “continued” anything is a big “change”…while I do not doubt the significance of these issues, libraries have been facing budget issues at least throughout the entire modern era of libraries.  And we have been doubting our relevance since I started library school when the Web was a neonate.  Now, the emphasis on community, while not exactly new, I believe did gain momentum.

I’m intrigued by the “sustainable development” concept for 2016…and not thrilled about the prospect of “leaner” staff models.  How much fat is left to trim?  I agree that more libraries will be opening up their spaces and physical books will likely be moved or removed.  Whether that is good, bad or just different…time will only tell.  If we have “leaner staff models” how can we continue to grow new services (community engagement) and maintain our current services (providing a wide selection of print books in open shelves)?

So, what are your thoughts?

Maintaining relevance in the age of OA

Librarians have been at the center of the Open Access movement, stemming in no small part from the unsustainable rise of journal expenses.  But, as more and more articles articles become freely available (free, as in beer), how will libraries maintain their relevance to our users?

I know, yet another existential, navel-gazing post by a Chicken-Little on the future of libraries.  Actually, no, it is not.  Because I have little doubt in the near- and medium-term future of our profession (and, short of total extinction of mankind due to ecologic or planetary tragedies, our long-term looks good, too).  But the question still remains…what if the majority of articles that our patrons use were freely available on the Web?  How would libraries remain relevant to our (former?) patrons?

What got me thinking about this was Aaron Tay’s posting about his own foray into this issue.   His question was: What percentage of citations made by our researchers is to freely available content?  His initial results were actually quite astonishing: 80% of the citations to articles by his institution’s economics faculty were freely available.  Of course, Aaron posed a number of limitations to this result, post problematic being that the timing of availability was unknown.  The papers he examined were from 2009-2015, but it is not known when these papers were available at the time that the authors gathered them.  That timing problem is a major obstacle to doing citation analysis for this question…but that’s another story.

We are aware of the numerous studies suggesting that academic faculty are using library-specific tools less and less for finding the resources they need.  By “library-specific”, I’m referring to the more traditional tools that librarians have developed or maintained, including the catalog, ejournal lists, and even the newer Discovery tools.  Instead, our faculty are turning more and more to Google Scholar or the overall Web.  Turning aside issues related to efficiency, let us assume that most of the faculty use GS most of the time to find articles (we will deal only with published journal articles for this thought experiment).  Will they find the articles they need to be freely-available?

Aaron cites a number of articles that have attempted to answer the corollary to his question: What percentage of articles on the Web are freely-available?  Their estimates ranged from 20% to 61% (Aaron – you have the foundations of a good systematic review and possibly a meta-analysis…keep going!).  Based on two studies that appeared to him to have more valid methodologies, Aaron estimates about 40%.  What if this rises to a critical mass of, say, 60% or 70%?  Could librarians strategically cut their costly and burdensome journal subscriptions?  Would libraries need to continue to be the purchasing agent (or, “wallet” as Aaron puts it) for the faculty – at least for journal articles?  If so, how would we remain relevant, especially to the disciplines for which scholarly outputs are primarily journal articles?

We could respond to this challenge as we did at the dawn of the Web, developing our own solutions, such as attempting to “catalog” it.  Or we could look at what we can and cannot control, and focus on the former and let the latter go.  We are often described as being middle-men when it comes to the flow of information.  We attempt to meet the needs of our patrons, which are very different institution-to-institution.  But we cannot control either side — neither what publishers do, nor what our patrons do.  For instance, just because a version of the article is freely-available does not mean it is easily accessible.  Indeed, publishers have attempted to build-in friction into their OA models, actively resisting attempts to make the OA versions less accessible or desirable.  And just because we expend significant resources to making our expensive resources available (via Discovery services or the catalog) does not mean our patrons will use them.

So, how do we remain relevant?  When I ask that question to myself, I next ask, why should we remain relevant?  Aside from my own future employment, why is it important for libraries to remain relevant as a middle-man?  After all, many middle-man occupations have gone by the wayside – people now buy many things once only available via salesmen.  My response to myself (and now, you) is that information is too important to be driven solely by market forces.  Information is what our decisions are based on, decisions that affect our lives, our livelihood, our future.  Access to information is a core value of librarianship, and by abandoning this to the market, we as a society risk being manipulated by those who control the information.

Given this, perhaps we need to re-think our role as ‘middle-men’.  Middle-men typically are independent, beholden to neither party – they attempt to meet the needs of both parties – the producers and the customers.  So, I would argue that librarians are more beholden to the people or consumers of information than we are to the producers.  True, we need to ensure that producers of information can be sustained, but it is the information itself that is important to our patrons, our people.  So I would posit that we librarians can and do maintain our relevance not by being middle-men or agents, but rather ombudsmen or advocates for the consumers of information.  Now, this should not mean that we control the flow of information, but that we ensure that quality information is available and accessible to our patrons.

Yet another weeding controversy

Obscure and popular books part of Berkeley library weeding process | Berkeleyside.

This article describes yet another controversy over a library director’s attempt to weed the collection.  Going beyond the he-said-she-said aspects of claims & counter-claims (which are stacking up in the director’s favor, if you ask me), I wanted to look at the methods, the problem of time, and communication.

Most librarians probably view the task of weeding with mixed emotions.  We know that it is necessary, but it leaves a bad taste in our mouths.  And the recent experiences at such institutions as the San Francisco, Urbana Free, and now Berkeley, have added yet another reason to put this task off.

The director claims to be using the CREW manual adopted by the Texas State Library and Archives Commission.  This manual provides general guidelines for reviewing the quality of a collection, as well as specific recommendations for general subject areas (by Dewey Decimal range, no less).  The manual was developed in 1976 primarily for small to medium sized public libraries, and has been updated as recently as 2012 to take into consideration changes to resources and technology.  Examples of the general guidelines for de-selection include:

  • outdated or inaccurate information
  • superseded editions
  • materials in poor condition
  • items not circulated in last 3-5 years
  • duplicate copies
  • titles in subject areas that are less frequently used

These can be summed up in the acronym, MUSTIE:

  • Misleading
  • Ugly
  • Superseded
  • Trivial
  • Irrelevant
  • Elsewhere available

The recommendations for specific call number ranges (very broad) include a “formula” or code that includes the minimum age of a work and the most recent years used to be considered for de-selection.  For instance, for 306 (Culture & Institutions), works older than 5 years old and last used more than 2 years ago should be considered for removal.  Of course, these are recommendations and not hard-and-fast rules, and each library should take into consideration interests of the community.  But this manual provides a starting place for librarians evaluating a collection.

Of the examples of titles that were and were not removed, I noticed these examples that clearly reflect the above recommendations.  There are over 500 libraries which already owned the work, Chinese Technology in the Seventeenth Century, while there were only 4 other libraries which held A Guide to Shrubs for Coastal California, which also is relevant to the community.

Now, another issue of contention is who is actually making the decisions.  Formerly, up to 25 librarians with different areas of expertise were involved.  This makes sense…it ensures that the work is evenly distributed and that those with experience and knowledge of the specific topics make the de-selections.  However, the director states that the librarians claimed to not have the time to attend to their routine weeding tasks; so he assigned the task to just two librarians.  Of course, you could argue that the director provide that time, through elimination or suspension of other responsibilities.  However, librarians are not soldiers…we are often an independent sort.  Being professionals, we take pride in having a certain amount of flexibility in deciding what we do on a daily basis.  But this can be frustrating to management, who often view their job of management as herding cats.  Again, weeding is a distasteful and difficult task…I can easily imagine how difficult it would have been to get all 25 librarians to attend to it when there were other more enjoyable jobs to do.  But this is getting back to the he-said-she-said aspect…Ideally, the director would have enabled all 25 librarians to attend to their weeding responsibilities with the same enthusiasm as their reference, instruction, and other jobs.

Finally, there is the issue of communication.  This is likely the most important part of any weeding project – routine or not.  The community needs to know that weeding is a necessary aspect of managing a library, and that their needs are being addressed in the process.  Too often, we attempt to hide it, or at least not disclose it to our patrons, for fear of just such a reaction.  The director did release the spreadsheets of the works that were removed, but only to the reporter and only after the kerfuffle.  But what is lacking from these sheets is the reasons for the decisions.

What if we do the opposite – market the idea of weeding much like we market our services and programs.  It is a service, after all.  We provide the books that our patrons want, and research has shown that older books, worn-out books, outdated books, etc. are not used, and thus not wanted.  The word itself can provide the advertising campaign theme – tending the garden, removing the weeds, making room for the fresh flowers, the newer, more wanted books.  Provide the community to opportunity for input…perhaps a “second-chance” shelf of books that are not clear-cut decisions.  And finally, work with the parent institution (state or local government, academic institution, whatever) to enable all books to be re-distributed (re-sale, or donations) and not “pulped” or destroyed.  Destroying books is most likely to raise the ire of people.

If we harken back Ranganathan’s 5 Laws, we may be distracted by the 2nd & 3rd laws – every reader his book, and every book its reader.  This may suggest that we need to keep all books just in case there is a reader looking for it.  But such principles must be kept in context of the others, including the 1st (books are for use), the 4th (save the time of the reader), and the 5th (libraries are growing organisms).  Retaining all books requires greater and greater amounts of space, making it more and more difficult to find and access the right books.  And community needs change over time, rendering the collection of 1940 much less relevant to community of 2010.

What I believe could be learned from these experiences:

  • Promote weeding as a service to the community
  • Provide the time and positive incentives for librarians tend to this task.
  • Allow the community input (without acceding to unreasonable demands).
  • Use a structured method and document decisions.
  • Keep the 5 Laws as principles guiding the process.

Blog at WordPress.com.

Up ↑