News Feeds, Microfilm, and the Stories We Tell Ourselves

A little over a week ago, I watched the searing and provocative TED talk by British journalist Carole Cadwalladr, “Facebook’s role in Brexit – and the threat to democracy.” It got me thinking about a few library things, which I thought might make for an interesting blog post. Then thinking about these library things took me down a series of rabbit holes, interconnecting and nuanced and compelling enough to chew up the entirety of the time I’d set aside for my turn in the Bitstreams blog rotation. There is no breezy, concise blog post that could pull them all together so I’m just going to do with it what I can with two of the maybe four or five rabbit holes that I fell into.

Cadwalladr took the stage at a TED conference sponsored by Facebook and Google, and spoke about her investigations into the role of Facebook and Cambridge Analytica in the Brexit vote in 2016. Addressing the big tech leaders present – the “Gods of Silicon Valley: Mark Zuckerberg, Sheryl Sandberg, Larry Page, Sergey Brin and Jack Dorsey” – she levelled a devastating j’accuse – “[W]hat the Brexit vote demonstrates is that liberal democracy is broken. And you broke it. This is not democracy — spreading lies in darkness, paid for with illegal cash, from God knows where. It’s subversion, and you are accessories to it.”

It was a courageous act, and Cadwalladr deserves celebration and recognition for it, even if the place it leaves us is a bleak one. As she would admit later, she felt massive pressure as she spoke. I had a number of reactions to her talk, but there was a line in particular got me thinking about library things. It occurred when she explained to that audience that “this entire referendum took place in darkness, because it took place on Facebook…, because only you see your news feed, and then it vanishes, so it’s impossible to research anything.” It provoked me to think about how we use “news feeds” – in the form of newspapers themselves – in the study of history, and the role that libraries play in preserving them.

Continue reading News Feeds, Microfilm, and the Stories We Tell Ourselves

Is there an app for that? The seemingly endless quest to make discovery easier for users

Contributed by Assessment & User Experience Department Practicum Students Amelia Midgett-Nicholson and Allison Cruse 

Duke University Libraries (DUL) is always searching for new ways to increase access and make discovery easier for users. One area users frequently have trouble with is accessing online articles. Too often we hear from students that they cannot find an article PDF they are looking for, or even worse, that they end up paying to get through a journal paywall. To address this problem, DUL’s Assessment and User Experience (AUX) Department explored three possible tools: LibKey Discovery, Kopernio, and Lean Library. After user testing and internal review, LibKey Discovery emerged as the best available tool for the job.  

LibKey Discovery logo

LibKey Discovery is a suite of user-friendly application programming interfaces (APIs) used to enhance the library’s existing discovery system.  The APIs enable one-click access to PDFs for subscribed and open-source content, one-click access to full journal browsing via the BrowZine application, and access to cover art for thousands of journals.  The tool integrates fully with the existing discovery interface and does not require the use of additional plug-ins.

According to their website, LibKey Discovery has the potential to save users thousands of clicks per day by providing one-click access to millions of articles.  The ability to streamline processes enabling the efficient and effective discovery and retrieval of academic journal content prompted the AUX department to investigate the tool and its capabilities further.  An internal review of the system was preceded by an introduction of the tool to Duke’s subject librarians and followed with a preliminary round of student-based user testing.

Current DUL interface
Current DUL discovery interface
LibKey interface
LibKey discovery interface

Pros

  • One-Click Article and Full Journal Access

Both the AUX staff and the subject librarians who performed an initial review of the LibKey Discovery tools were impressed with the ease of article access and full journal browsing.  Three members of the AUX department independently reviewed LibKey’s features and concluded the system does provide substantial utility in its ability to reduce the number of clicks necessary to access articles and journals.

  • Streamlined Appearance

The tool streamlines the appearance and formatting of all journals, thus removing ambiguity in how to access information from different sources within the catalog.  This is beneficial in helping to direct users to the features they want without having to search for points of access. The AUX department review team all found this helpful.

  • Seamless Integration

LibKey Discovery’s APIs integrate fully into the existing DUL discovery interface without the need for users to download an additional plug-in.  This provides users the benefit of the new system without asking them to go through extra steps or make any changes to their current search processes.  Aside from the new one-click options available within the catalog’s search results page, the LibKey interface is indistinguishable from the current DUL interface helping users to benefit from the added functionality without feeling like they need to learn a new system.  

Cons

  • Cost

LibKey Discovery carries a relatively hefty price tag, so its utility to the end-user must be weighed against its cost.  While internal review and testing has indicated LibKey Discovery has the ability to streamline and optimize the discovery process, it must be determined if those benefits are universal enough to warrant the added annual expenditure.

  • Inconsistency in Options

A potential downside to LibKey Discovery is lack of consistency in one-click options between articles.  While many articles provide the option for easy, one-click access to a PDF, the full text online, and full journal access, these options are not available for all content.  As a result, this may cause confusion around the options that are available for users and may diminish the overall utility of the tool depending on what percentage of the catalog’s content is exempt from the one-click features.

LibKey Discovery User Testing Findings

An initial round of user testing was completed with ten student volunteers in the lobby of Perkins Library in early April.  Half of the users were asked to access an article and browse a full journal in the existing DUL system; the other half were asked to perform the same tasks using the LibKey Discovery interface.

Initial testing indicated that student users had a high level of satisfaction with the LibKey interface; however, they were equally satisfied with the existing access points in the DUL catalog.  The final recommendations from the user testing report suggest the need for additional testing to be completed. Specifically, it was recommended that more targeted testing be completed with graduate-level students and faculty as a majority of the original test’s participants were undergraduate students with limited experience searching for and accessing academic journal issues and articles.  It was concluded that testing with a more experienced user group would likely produce better feedback as to the true value of LibKey Discovery.

LibKey Summary

LibKey Discovery is a promising addition to Duke’s existing discovery system.  It allows for streamlined, one-click article and full journal access without disrupting the look and feel of the current interface or requiring the use of a plug-in.  Initial reviews of the system by library staff have been glowing; however, preliminary user testing with student participants indicated the need for additional testing to determine if LibKey’s cost is sufficiently offset by its utility to the user.

Kopernio logo

Kopernio is a free browser plug-in which enables one-click access to academic journal articles. It searches the web for OA copies, institutional repository copies, and copies available through library subscriptions. The tool is designed to connect users to articles on and off campus by managing their subscription credentials and automatically finding the best version of an article no matter where a user is searching.

Given the potential of this tool to help increase access and make discovery easier for students, the AUX department initiated an internal review process. Four members of the department independently downloaded the Kopernio plug-in, thoroughly tested it in a variety of situations, and shared their general and specific notes about the tool.

Pros

  • OA Content + Library Subscription

By its design, Kopernio has an advantage over other plug-in tools that serve a similar function (i.e. Unpaywall). When users first download Kopernio they are asked to register their subscription credentials. This information is saved in the plug-in so users can automatically discover articles available from OA sources, as well as library subscriptions. This is an advantage over other plug-ins that only harvest from freely available sources.

Screenshot: Kopernio sign-in page
Kopernio sign-in page
  • Branding

Kopernio has highly visible and consistent branding. With bright green coloring, the plug-in stands out on a screen and attracts users to click on it to download articles.

  • One-Click

Kopernio is advertised as a “one-click” service, and it pays off in this respect. Using Kopernio to access articles definitely cuts down on the number of clicks required to get to an article’s PDF. The process to download articles to a computer was instantaneous, and most of the time, downloading to the Kopernio storage cloud was just as fast.

Cons

  • Creates New Pain Points

Kopernio’s most advertised strength is its ability to manage subscription credentials. Unfortunately, this strength is also a major data privacy weakness. Security concerns ultimately led to the decision to disable the feature which allowed users to access DUL subscriptions via Kopernio when off-campus. Without this feature, Kopernio only pulls from OA sources and therefore performs the same function that many other tools currently do.

Similar to data privacy concerns, Kopernio also raises copyright concerns. One of Kopernio’s features is its sharing function. You can email articles to anyone, regardless of their university affiliation or if they have downloaded Kopernio already. We tested sending DUL subscription PDFs to users without Duke email addresses and they were able to view the full-text without logging in. It is unclear if they were viewing an OA copy of the article, or if they were seeing an article only meant for DUL authenticated users.

Screenshot: Sharing articles through Kopernio
Sharing an article through Kopernio

Running the Kopernio plug-in noticeably slowed down browser speed. We tested the browser on several different computers, both on campus and off, and we all noticed slower browser speeds. This slow speed led Kopernio to be occasionally buggy (freezing, error messages etc.).

Screenshot showing a buggy screen from Kopernio
Buggy screen while using Kopernio
  • Many Features Don’t Seem Useful

When articles are saved to Kopernio’s cloud storage, users can add descriptive tags. We found this feature awkward to use. Instead of adding tags as you go along, users have to add a tag globally before they can tag an article. Overall, it seemed like more hassle than it was worth.

Kopernio automatically imports article metadata to generate citations. There were too many problems with this feature to make it useful to users. It did not import metadata for all articles that we tested, and there was no way to manually add metadata yourself. Additionally, the citations were automatically formatted in Elsevier Harvard format and we had to go to our settings to change it to a more common citation style.

Lastly, the cloud storage which at first seemed like an asset, was actually a problem. All articles automatically download to cloud storage (called the “Kopernio Locker”) as soon as you click on the Kopernio button. This wouldn’t be a problem except for the limited storage size of the locker. With only 100MB of storage in the free version of Kopernio, we found that after downloading only 2 articles the locker was already 3% full. To make this limited storage work, we would have to go back to our locker and manually delete articles that we did not need, effectively negating the steps saved by having an automatic process.

Lean Library Logo

Lean Library is a similar tool to Kopernio. It offers users one-click access to subscription and open access content through a browser extension. In Fall 2018, DUL staff were days away from purchasing this tool when Lean Library was acquired by SAGE Publishing. DUL staff had been excited to license a tool that was independent and vendor-neutral and so were disappointed to learn about its acquisition. We have found that industry consolidation in the publishing and library information systems environment has lowered competition and resulted in negative experiences for researchers and staff. Further, we take the privacy of our users very seriously and were concerned that Lean Library’s alignment with SAGE Publishing will compromise user security. Whenever possible, DUL aims to support products and services that are offered independently from those with already dominant market positions. For these reasons, we opted not to pursue Lean Library further.

Conclusion

Of the three tools the AUX Department explored, we believe LibKey Discovery to be the most user-friendly and effective option. If purchased, it should streamline journal browsing and article PDF downloads without adversely affecting the existing functionality of DUL’s discovery interfaces.

Smart People Who Care

It’s that time of year at the university when we’re working on our PEPs (Performance Evaluation and Planning forms) and I’m thinking about how grateful I am to have such smart staff who really care about their work, their colleagues, and the people they serve, as we advance technology across the libraries. In contrast to some corporate environments, the process here really does aim to help us improve, rather than rank us as a setup for “resource actions” (firings). This excellent article, The Feedback Fallacy by Marcus Buckingham and Ashley Goodall, reminds me to emphasize the things people do well, and encourage them to build on their strengths.

Attuned to ethical practices within organizations, I’m also excited about increasing awareness of ethics in the effects of the technologies we produce. Justin Sherman, co-founder of the Ethical Tech initiative here at Duke, did a stimulating talk at the Edge Workshop this month about ethical issues that surround technology, such as search engine bias, and AI tools that judges use to determine sentencing for crimes.  Justin recommends this podcast, with Christopher Lydon on Open Source, called Real Education About Artificial Intelligence.  Library staff are participating in the Kenan Institute for Ethics book club program (KIE), where the spring selection is Algorithms of Oppression: How Search Engines Reinforce Racism by Safiya Umoja Noble.

And, I’m pleased to exercise my hiring mantra, “smart people who care”, which has served me well for over 30 years, as we’re recruiting candidates with I/T and team leadership experience for a new position, Computing Services Supervisor.

Happy Spring!
Laura Cappelletti
Director, Information Technology Services

The Commons Approach

Earlier this month, I was invited to give some remarks on “The Commons Approach” at the LYRASIS Leaders Forum, which was held at the Duke Gardens.  We have a great privilege and opportunity as part of the Duke University Libraries to participate in many different communities and projects, and it is one of the many reasons I love working at Duke.  The following is the talk I gave, which shares some personal and professional reflections of the Commons.


The Commons Approach is something that I have been committed to for almost the entirety of my library career, which is approaching twenty years.  When I start working in libraries at Lehigh University, I came into the community with little comprehensive about the inner workings of libraries.  I had no formal library training, and my technology education, training, and work experience had been developed through experimentation and learning by doing.  Little did I know at the time, I was benefiting from small models of the Commons, or even how to define it.

There are a number of definitions of Commons, but one definition I like best that I found in Wikipedia is “a general term for shared resources in which each stakeholder has an equal interest”.  There are many contexts:

  • natural resources – air, water, soil
  • cultural norms and shared values
  • public lands that no one owns
  • information resources that are created collectively and shared among communities, free for re-use

The librarians and library staff around me at Lehigh brought me into their Commons, and gave me the time and space to learn about the norms, shared values, terminology, language, jargon, and collective priorities of libraries so that I could begin to apply them with my skills and experiences and be an active contributor to the Commons of Libraries.  As I began to become more unified with them in the commons approach, my diverse work experience and skill added to our community.  I recognized that I belonged, that I was now a stakeholder with an equal interest, and that I could create and share among the broader library community.

If we start with the foundation that libraries are naturally driven towards a Commons approach within their own campus or organization, we can examine the variety of models of projects and communities that extend the Commons Approach.

  • Open Source Software Projects (Apache Model)
    • earned trust through contributions over time
    • primarily a complete volunteering of time and effort
    • peer accountability, limited risk of power struggles
    • tends to choose the most open license possible
  • Community Source Projects (Kuali OLE, Fedora)
    • “those who bring the gold set the rules”
    • The more commitment you make in resources, the more privilege and opportunity you receive
    • Dependent on defined governance around both rank of commitment and representation
    • Tends to choose open licensing that offers most protection and control
  • Membership initiatives (OLE [before and after Kuali], APTrust, DPN, SPN)
    • Typically tiered, proportional model focused primarily on financial contributions
    • Convergence around a strategic initiative, project, or outcome
    • Pooled financial resources to develop and sustain a solution
  • Consortial Partnerships (Informal, Formal, Mixed)
    • Location- or context-based partnership to collaborate
    • Defined governance structure
    • Informal or formal
  • National Initiatives (Code4Lib, DPLA, DLF)
    • Annual conferences or meetings
    • Distributed communications commons (listserv, Slack, website)
    • Coordinated around large ideas or contexts and sharing local ideas / projects to build grass roots change
    • Focused on democratizing opportunity for sharing and collaboration
  • Community Projects with Corporate Sponsors (for-profit, not-for-profit)
    • Hybrid or mixed models of community source or membership initiatives
    • Corporate services support to implementers
    • Challenges in governance of priorities between sponsor and community

NOTE: There are more models and nuances to these models.

Benefits and Challenges

Each of these models has benefits and challenges.  One of the issues that I have become particularly interested in, and consistently advocate for, is creating an environment that promotes diversity of participants.  The Community Source model of privileging those who bring the gold, for example, tends to bias larger organizations that have financial and human resource flexibility and requires clear proportional investment tiers that recognize varied sizes of organizations wanting to join the community.  But while contribution levels can be defined at varied tiers, costs are constant and usually fixed, especially staffing costs, which will put pressure on the community to sustain. A smaller, startup community thus needs larger investments to incubate the project no matter how equal they intend to share the ownership across all of the stakeholders. Thus:

  • How do we develop our communities to fully embrace a Commons Approach that gives each stakeholder equal opportunity that also embraces differences of the organizations within the community?
  • How do we setup our communities that empower smaller organizations not only to join but to lead?
  • How we do setup our communities to encourage well-resourced organizations to contribute without automatically assuming leadership or control?

Open Source

My first opportunity to experience the Commons Approach outside of my own library was as a member of the VuFind project.  While Villanova University led and sponsored the project, Andrew Nagy, the founding developer, contributed his hard work to the whole community and invited others to co-develop.  As you gained the trust of the lead developers in what you contributed, you earned more responsibility and opportunity to work on the core code.  If you chose to focus on specific contributions, you became the lead developers of that part of the code.  It was all voluntary and all contributing to a common purpose: to develop an open source faceted discovery platform.  As leaders transitioned to new jobs or new organizations, some stayed on the project, and some were replaced with other community members who had earned their opportunity.

Community Source

While I was working at Lehigh and participating in the Open Library Environment as a membership model, it was simpler for me to feel like I belonged because each member had a single member on the governance group.  I was an equal member, and Lehigh had an equal stake.  We held in Common priorities for our community, for the project, and for the outcome.  We held in Common that each of us represented libraries from different contexts: private, public, large, small, US-based, and International.  We held in Common the priority to grow and attract other libraries of various sizes, contexts, and geographic locations.  It felt idealistic.

The ideal shattered when OLE joined the Kuali Foundation and the model changed from membership to Community Source.  The rules of that model were different, and thus the foundation of their Commons was also different.  While tiers of financial contribution were still in place, it was clear that the more resources your organization brought, the more influence your organization would have.  Vendors were also members of the community, and they put in a different level and category of resources.

Moreover, there were joint governance committees overseeing projects that multiple projects were using at the same time.  Which project’s priorities would be addressed first depended on which project was paying more into that project.  I quickly realized that OLE, which was not paying as much as others, would not be getting its needs addressed.  To be fair, this structure worked for some of the Kuali projects and worked well.  But it not a Commons approach that the Open Library Environment had been committed to, and it was not the best model for that community to be successful.

Consortial Partnerships

Consortial partnerships are critical for local, regional, and national collaboration, and these partnerships are centered on a variety of common strategies, from buying or licensing collections to resource sharing to digital projects.  There are formal and informal consortia, some that are decades old and some that are very new.  As libraries continue to face constrained resources, banding together through these common constraints will be more and more critical to providing our users the level of service we expect to provide – another common thread: excellence.

National Initiatives

National initiatives have been getting a lot of press this year, mostly for difficult reasons.  There are also a variety of contexts for national initiatives, but most of the ones we likely consider are joined by a shared commitment to a topic, theme, or challenge.  Code4Lib began in 2003 as a listserv of library programmers hoping to find community with others in the library, museum, and archives community.  Code4Lib started meeting annually when it became clear it would be beneficial to share projects and ideas in person, hack and design things together, and find new ways to collaborate out in the open.

The Digital Library Federation is a community of practitioners who advance research, learning, social justice, and the public good through the creative design and wise application of digital library technologies.  The Digital Public Library of America was founded to maximize public access to the collections of historical and cultural organizations across the country.

The methods each of these national initiatives are quite different from each other, but they have focus on a Commons Approach that recognizes their collective effort is greater than the sum of their individual results.

Community Projects with Corporate Sponsors

The newest model of the Commons Approach is the hybrid open-source or community-source projects that include corporate sponsorship, hosting, or services. There are example of both for-profit and not-for-profit sponsorships, as well as a not-for-profit who tends to act at times like a for-profit, and the library community is continuing to have mixed reactions.  Some libraries embrace this interest by corporate partners, while others outright reject the notion as a type of Trojan horse.  Some are skeptical of specific corporations, while some have biases towards or against specific partners.  This new paradigm challenges our notion of openness, but it also offers an opportunity to explore different means to the Commons Approach.  The same elements apply – what are norms, values, terminology, language, jargon, and collective priorities that we share together?  What benefits can each stakeholder bring to the community?  Is there a diversity of participation, leadership, and contribution that creates inclusion?  Are there new aspects to having corporate sponsors join that the library community cannot do on its own?

Economies of scale

One of these new aspects is developing new means of economies of scale, which is a good step to sustaining a Commons Approach.  Economies of scale allows greater opportunity for libraries of different sizes and financial resources to work together. Open and community source projects in particular need solid financial planning, but great ideas and leadership are not limited to libraries with larger budgets or staff size.  Continuing to increase the opportunity for diversity of the community will be a great outcome and encourage a broader adoption of the Commons Approach.

Yet project staffing and resources are usually fixed costs that are not kind to the attempts to enable libraries to make variable contributions.  It requires a balance of large and small contributions from the community, and the entry of corporate sponsors has enable some new financial and infrastructure security missing in many projects and initiatives.  Yet, as a community focused on common values, norms, and priorities, it is not disingenuous to use due diligence to ensure all members of the community, library and sponsor alike, are committed to the Commons Approach and not in it for a free ride or an ulterior motive.

New Paradigm as a Disruptive Force

And it is accurate to call this new paradigm a disruptive force to open- and community-source projects.  It is up to the community to decide for itself what the best is for their future: embrace the disruption and adapt for the positive gains; hold true to their origins and continue on their path; or be torn apart by change, ignoring or forgetting their Common Approach foundation in the wake of the disruptive force.

Each of the models above have had some manner of corporate sponsorship examples, so we know there is success to be found regardless of the model.  And there are still many examples that remain strong in their original framework.  What I find encouraging, even in the difficult situations of the past year for many organizations, is that we are challenging our notions of how to develop these communities so that we can develop greater sustainability, greater participation from a more diverse and representative community, and achieve broader success of the Commons together – for our users, the most important connecting element of all.

My Family Story through the Duke Digital Collections Program

Hello! This is my first blog as the new Digital Production Service Manager, and I’d like to take this opportunity to take you, the reader, through my journey of discovering the treasures that the Duke Digital Collections program offers. To personalize this task, I  explored the materials related to my family’s journey to the United States. First, I should contextualize. After migrating from south China in the mid-1800s, my family fled Vietnam in the late 1970s and we left with the bare necessities – mainly food, clothes, and essential documents. All I have now are a few family pictures from that era and vividly told stories from my parents to help me connect the dots of my family’s history.

When I started delving into Duke’s Digital Collections, it was heartening to find materials of China, Vietnam, and even anti-war materials in the U.S. The following are some materials and collections that I’d like to highlight.

The Sidney D. Gamble Photographs offer over 5,000 photographs of China in the early 20th century. Images of everyday life in China and landscapes are available in this collection.The above image from the Gamble collection, is that of a junk, or houseboat, photographed in the early 1900s. When my family fled Vietnam, fifty people crammed into a similar vessel and sailed in the dead of night along the Gulf of Tonkin. My parents spoke of how they were guided by the moonlight and how fearful they were of the junk catching fire from cooking rice.

The African American Soldier’s Vietnam War photograph album collection offers these gorgeous images of Vietnam. This is the country that was home for multiple generations for my family, and up until the war, it was a good life. I am astounded and grateful that these postcards were collected by an American soldier in the middle of war. Considering that I grew up in Los Angeles, California, I have no sense of the world that my parents inhabited, and these images help me appreciate their stories even more. On the other side of the planet, there were efforts to stop the war and it was intriguing to see a variety of digital collections depicting these perspectives through art and documentary photography. The image below is that of a poster from the Italian Cultural Posters collection depicting Uncle Sam and the Viet Cong.

In addition to capturing street scenes in London, the Ronald Reis Collection, includes images of Vietnam during the war and anti-war effort in the United States. The image below is that of a demonstration in Bryant Park in New York City. I recognize that the conflict was fought on multiple fronts and am grateful for these demonstrations, as they ultimately led to the end of the war.Lastly, the James Karales Photos collection depicts Vietnam during the war. The image below, titled “Soldiers leaving on helicopter” is one that reminds me of my uncle who left with the American soldiers and started a new life in the United States. In 1980, thanks to the Family Reunification Act, the aid of the American Red Cross, and my uncle’s sponsorship, we started a new chapter in America.

Perhaps this is typical of the immigrant experience, but it still is important to put into words. Not every community has the resources and the privilege to be remembered, and where there are materials to help piece those stories together, they are absolutely valued and appreciated. Thank you, Duke University Libraries, for making these materials available.

Wrangling Messy Data with Airtable

The Assessment & User Experience department at Duke University Libraries keeps the libraries’ physical and virtual spaces responsive to user needs by constantly gathering feedback. In additional to our biennial user satisfaction survey, we run usability tests, hold focus groups, and host meetings of our student advisory boards, all in an effort to keep a finger on the pulse of the DUL patrons.

These activities can generate a lot of unstructured data! For example, in a typical meeting of our undergraduate advisory board, we might collect feedback from a dozen or more students, generating seven or more pages of notes and covering a range of topics. We review and act upon some of these comments immediately, but others may influence longer-term planning. As library staff, we know how important it is to store information in a way that promotes future access. This year we decided to pilot a new system for storing and describing our unstructured data.

Airtable logoEnter Airtable. If you’re not familiar, Airtable is a cloud-based database solution. Similar to Google Sheets, Airtable lets you enter and share data in your web browser, but it also offers more powerful features for projects that have messy data or interconnected components. There are many Airtable templates to show off the different features, including project trackers, event planners, and even product catalogs.

For our messy data, we built a simple spreadsheet that was general enough to collect data from a variety of sources. We included columns like basic demographics, the feedback provided, the original question or prompt, the date when feedback was provided, and how we collected the feedback. Then we took advantage of Airtable’s special features to create a column for topical tags. One of the column types in Airtable is called “multiple select“, which means you can add multiple tags to a single comment. Other spreadsheets can’t understand a list of tags in a single cell, but Airtable treats each tag separately and allows us to group and filter comments by each individual tag.

a screenshot of an Airtable database with tagged comments

The ability to look at comments across different feedback channels in one central location has enormous potential. Instead of having to hunt through old Word documents or emails, we have a single database that can be searched, sorted, or filtered to explore trends in comments over time. When a question comes up about how patrons feels about a particular service or space, we can compile data much more easily, and we no longer have to rely on our memory of what feedback we’ve received and when.

screenshot of Airtable tutorial instructions

Airtable’s free accounts have a limited number of rows allowed in each database,  but they do offer a discount on paid plans to educational institutions. We’re only just starting to explore the potential of Airtable, but so far we’ve been happy with the ability to collect our messy data in one place and organize comments with tags.

Want to learn more? Take a look at our recent tutorial on using Airtable for coding survey data, originally offered at the Designing For Digital 2019 conference.

Bringing 500 Years of Women’s Work Online

Back in 2015, Lisa Unger Baskin placed her extensive collection of more than 11,000 books, manuscripts, photographs, and artifacts in the Sallie Bingham Center for Women’s History & Culture in the David M. Rubenstein Rare Book & Manuscript Library. In late February of 2019, the Libraries opened the exhibit “Five Hundred Years of Women’s Work: The Lisa Unger Baskin Collection” presenting visitors with a first look at the diversity and depth of the collection, revealing the lives of women both famous and forgotten and recognizing their accomplishments. I was fortunate to work on the online component of the exhibit in which we aimed to offer an alternate way to interact with the materials.

Homepage of the exhibit
Homepage of the exhibit

Most of the online exhibits I have worked on have not had the benefit of a long planning timeframe, which usually means we have to be somewhat conservative in our vision for the end product. However, with this high-profile exhibit, we did have the luxury of a (relatively) generous schedule and as such we were able to put a lot more thought and care into the planning phase. The goal was to present a wide range and number of items in an intuitive and user-friendly manner. We settled on the idea of arranging items by time period (items in the collection span seven centuries!) and highlighting the creators of those items.

We also decided to use Omeka (classic!) for our content management system as we’ve done with most of our other online exhibits. Usually exhibit curators manually enter the item information for their exhibits, which can get somewhat tedious. In this case, we were dealing with more than 250 items, which seemed like a lot of work to enter one at a time. I was familiar with the CSV Import plugin for Omeka, which allows for batch uploading items and mapping metadata fields. It seemed like the perfect solution to our situation. My favorite feature of the plugin is that it also allows for quickly undoing an ingest in case you discover that you’ve made a mistake with mapping fields or the like, which made me less nervous about applying batch uploads to our production Omeka instance that already contained about 1,100 items.

Metadata used for batch upload
Metadata used for batch upload

Working with the curators, we came up with a data model that would nest well within Omeka’s default Dublin-core based approach and expanded that with a few extra non-standard fields that we attached to a new custom item type. We then assembled a small sample set of data in spreadsheet form and I worked on spinning up a local instance of Omeka to test and make sure our approach was actually going to work! After some frustrating moments with MAMP and tracking down strange paths to things like imagemagick (thank you eternally, Stack Overflow!) I was able to get things running well and was convinced the batch uploads via spreadsheet was a good approach.

Now that we had a process in place, I began work on a custom theme to use with the exhibit. I’d previously used Omeka Foundation (a grid-based starter theme using the Zurb Foundation CSS framework) and thought it seemed like a good place to start with this project. The curators had a good idea of the site structure that they wanted to use, so I jumped right in on creating some high-fidelity mockups borrowing look-and-feel cues from the beautiful print catalog that was produced for the exhibit. After a few iterations we arrived at a place where everyone was happy and I started to work on functionality. I also worked on incorporating a more recent version of the Foundation framework as the starter theme was out of date.

Print catalog for the exhibit
Print catalog for the exhibit

The core feature of the site would be the ability to browse all of the items we wanted to feature via the Explore menu, which we broke into seven sections — primarily by time period, but also by context. After looking at some other online exhibit examples that I thought were successful, we decided to use a masonry layout approach (popularized by sites like Pinterest) to display the items. Foundation includes a great masonry plugin that was very easy to implement. Another functionality issue had to do with displaying multi-page items. Out of the box, I think Omeka doesn’t do a great job displaying items that contain multiple images. I’ve found combining them into PDFs works much better, so that’s what we did in this case. I also installed the PDF Embed plugin (based on the PDF.js engine) in order to get a consistent experience across browsers and platforms.

Once we got the theme to a point that everyone was happy with it, I batch imported all of the content and proceeded with a great deal of cross-platform testing to make sure things were working as expected. We also spent some time refining the display of metadata fields and making small tweaks to the content. Overall I’m very pleased with how everything turned out. User traffic has been great so far so it’s exciting to know that so many people have been able to experience the wonderful items in the exhibit. Please check out the website and also come visit in person — on display until June 15, 2019.

Examples of 'Explore' and 'Item' pages
Examples of ‘Explore’ and ‘Item’ pages

It Takes a Village to Curate Your Data: Duke Partners with the Data Curation Network

In early 2017, Duke University Libraries launched a research data curation program designed to help researchers on campus ensure that their data are adequately prepared for both sharing and publication, and long term preservation and re-use. Why the focus on research data? Data generated by scholars in the course of their investigation are increasingly being recognized as outputs similar in importance to the scholarly publications they support. Open data sharing reinforces unfettered intellectual inquiry, fosters transparency, reproducibility and broader analysis, and permits the creation of new data sets when data from multiple sources are combined. For these reasons, a growing number of publishers and funding agencies like PLoS ONE and the National Science Foundation are requiring researchers to make openly available the data underlying the results of their research.

Data curation steps

But data sharing can only be successful if the data have been properly documented and described. And they are only useful in the long term if steps have been taken to mitigate the risks of file format obsolescence and bit rot. To address these concerns, Duke’s data curation workflow will review a researcher’s data for appropriate documentation (such as README files or codebooks), solicit and refine Dublin Core metadata about the dataset, and make sure files are named and arranged in a way that facilitates secondary use. Additionally, the curation team can make suggestions about preferred file formats for long-term re-use and conduct a brief review for personally identifiable information. Once the data package has been reviewed, the curation team can then help researchers make their data available in Duke’s own Research Data Repository, where the data can be licensed and assigned a Digital Object Identifier, ensuring persistent access.

 

“The Data Curation Network (DCN) serves as the “human layer” in the data repository stack and seamlessly connects local data sets to expert data curators via a cross-institutional shared staffing model.”

 

New to Duke’s curation workflow is the ability to rely on the domain expertise of our colleagues at a few other research institutions. While our data curators here at Duke possess a wealth of knowledge about general research data-related best practices, and are especially well-versed in the vagaries of social sciences data, they may not always have the all the information they need to sufficiently assess the state of a dataset from a researcher. As an answer to this problem, the Data Curation Network, an Alfred P. Sloan Foundation-funded endeavor, has established a cross-institutional staffing model that distributes the domain expertise of each of its partner institutions. Should a curator at one institution encounter data of a kind with which they are unfamiliar, submission to the DCN opens up the possibility for enhanced curation from a network partner with the requisite knowledge.

DCN Partner Institutions
DCN Partner Institutions

Duke joins Cornell University, Dryad Digital Repository, Johns Hopkins University, University of Illinois, University of Michigan, University of Minnesota, and Pennsylvania State University in partnering to provide curatorial expertise to the DCN. As of January of this year, the project has moved out of pilot phase into production, and is actively moving data through the network. If a Duke researcher were to submit a dataset our curation team thought would benefit from further examination by a curator with domain knowledge, we will now reach out to the potential depositor to receive clearance to submit the data to the network. We’re very excited about this opportunity to provide this enhancement to our service!

Looking forward, the DCN hopes to expand their offerings to include nation-wide training on specialized data curation and to extend the curation services the network offers beyond the partner institutions to individual end users. Duke looks forward to contributing as the project grows and evolves.

Sustainability Planning for a Better Tomorrow

In March of last year I wrote about efforts of the Resource Discovery Systems and Strategies team (RDSS, previously called the Discovery Strategy Team) to map Duke University Libraries’ discovery system environment in a visual way. As part of this project we created supporting documentation for each system that appeared in a visualization, including identifying functional and technical owners as well as links to supporting documentation. Gathering this information wasn’t as straightforward as it ideally should have been, however. When attempting to identify ownership, for example, we were often asked questions like, “what IS a functional owner, anyway?”, or told “I guess I’m the owner… I don’t know who else it would be”. And for many systems, local documentation was outdated, distributed across platforms, or simply nonexistent.

As a quick glance through the Networked Discovery Systems document will evince, we work with a LOT of different systems here at DUL, supporting a great breadth of processes and workflows. And we’ve been steadily adding to the list of systems we support every year, without necessarily articulating how we will manage the ever-growing list. This has led to situations of benign neglect, confusion as to roles and responsibilities and, in a few cases, we’ve hung onto systems for too long because we hadn’t defined a plan for responsible decommission.

So, to promote the healthier management of our Networked Discovery Systems, the RDSS team developed a set of best practices for sustainability planning. Originally we framed this document as best practices for maintenance planning, but in conversations with other groups in the Libraries, we realized that this didn’t quite capture our intention. While maintenance planning is often considered from a technical standpoint, we wanted to convey that the responsible management of our systems involves stakeholders beyond just those in ITS, to include the perspective and engagement of non-technical staff. So, we landed on the term sustainability, which we hope captures the full lifecycle of a system in our suite of tools, from implementation, through maintenance, to sunsetting, when necessary.

The best practices are fairly short, intended to be a high-level guide rather than overly prescriptive, recognizing that every system has unique needs. Each section of the framework is described, and key terms are defined. Functional and technical ownership are described, including the types of activities that may attend each role, and we acknowledge that ownership responsibilities may be jointly accomplished by groups or teams of stakeholders. We lay out the following suggested framework for developing a sustainability plan, which we define as “a living document that addresses the major components of a system’s life cycle”:

  • Governance:
    • Ownership
    • Stakeholders
    • Users
  • Maintenance:
    • System Updates
    • Training
    • Documentation
  • Review:
    • Assessments
    • Enhancements
    • Sunsetting

Interestingly, and perhaps tellingly, many of the conversations we had about the framework ended up focusing on the last part – sunsetting. How to responsibly decommission or sunset a system in a methodical, process-oriented way is something we haven’t really tackled yet, but we’re not alone in this, and the topic is one that is garnering some attention in project management circles.

So far, the best practices have been used to create a sustainability plan for one of our systems, Dukespace, and the feedback was positive. We hope that these guidelines will facilitate the work we do to sustain our system, and in so doing lead to better communication and understanding throughout the organization. And we didn’t forget to create a sustainability plan for the best practices themselves – the RDSS team has committed to reviewing and updating it at least annually!

Ad-Blocking … and a Mea Culpa

It’s all but impossible to use the internet and not be aware of the sheer quantity of advertising out there. Some estimates suggest Google alone serves out nearly 30 Billion ads per day; other estimates suggest 300-700 ads are shown per person per day. In trying to get more eyeballs on their images, advertisers have resorted to more and more intrusive ad displays — pop-over ads (so you can’t see content until you close the ad), pop-under ads (so even after you’ve left the site, you’ll see one more ad), animated gifs (the motion naturally causes your eye to look at it), auto-playing (and loud) videos. More recently, ads have even been implicated in malware attacks.

So it’s no surprise that ad-blocking technology is now mainstream. All major browsers have multiple ad-blocker plug-ins, and any list of “Best Add-Ons for <browser>” will likely include at least one. By blocking ads, these plugins reduce the annoyance of the ads while also helping protect your privacy by reducing sites’ ability to track you.

As an additional bonus, they can also accelerate your web browsing — by not downloading all that ad content, you’re able to see the “real” content faster. A New York Times article showed that between 33% and 50% of the time to download a page was due to advertising — and in extreme cases, advertising could be 80% or more of the time to download and view a webpage.

In IT-Core Services, we had intended to deploy an ad-blocker to all of the public computers in order to allow our patrons and users to block ads while they were doing work, collecting papers, or doing research on Library computers.  The plugin is called “uBlock Origin” and is one of the leading open-source ad-blockers around.

But … Oops.

We accidentally pushed it to all public AND staff computers several months ago.

Given the very few number of tickets we’ve seen about it, we’re guessing people either didn’t notice, or else welcomed the arrival of uBlock.  We’re now planning on keeping uBlock deployed on all staff and public computers.  We feel that the privacy, performance, and security benefits of uBlock outweigh the desire for a “ad-full” web experience — and you can easily un-block any websites you want to, if you find that the blocker is somehow interfering with that site.

How to Un-Block a Website:

To unblock the websitethat you’re visiting — that is, to show the ads on the page — look for the uBlock logo (a brick-red shield) at the top of the main browser window.  Clicking that logo will pop up a dialog box like this:By clicking on the power-button symbol (circle with the line at the top), you’ll tell uBlock to NOT block ads on that webpage in the future.  You should then reload the page to get the ad-full experience (by default, un-blocking a website does NOT reload or re-display the ads, you must explicitly reload the page).  Note:  if the uBlock logo is grey, or the power-button icon is grey, then the current website is already un-blocked (and the browser is showing you the ads).

How to Un-Block All/Many Websites:

To unblock a lot of websites at once, you have to go to the uBlock “Dashboard” or settings menu.  Again, click on the uBlock shield logo at the top of the browser window, then look for the 3-sliders icon (immediately below the power-button, to the right of the dialog box).  Clicking that will bring up a new virtual webpage with a variety of settings on it:

Click on the “Filter Lists” tab and you’ll see a set of “Filter” checkboxes.  Each checkbox represents a set of websites that are to be blocked (checked) or unblocked (unchecked).  To unblock all websites — to essentially deactivate uBlock Origin altogether — just uncheck all of the Filter sets.  FWIW, most of the filter-sets have Home icons where you can find more info on what that filter-set does (e.g. “Malware Domains” links to a website at malwaredomains.com, which is run and maintained by the company RiskAnalytics).

If you have any questions, please just submit a ticket and someone will get back to you.

(title image is the “Million Dollar Webpage“, I’ll link you to the Wikipedia page rather than the ad-full page!)


UPDATE – 15 Feb 2019 – uBlock and DUL Newsletters:

While a great many websites will work fine with uBlock Origin installed, it turns out the Library’s own newsletter system does NOT!  If you are experiencing problems with the newsletter, go into the uBlock settings (process described above) and go to the “Filter Lists” tab.  One of the filter-sets at the bottom is named “Peter Lowe’s Ad and tracking server list” — this is the one that seems to catch the iContact server used by our newsletter.  If you disable that (the box will be un-checked), then reload the page, you should be back in operation.  Sorry!

Notes from the Duke University Libraries Digital Projects Team