CRRA Update Winter 2016

CRRA Update Winter 2016
(December, January, February)
Please see the PDF for the more visually rich version.

Feature Article: Interview with Michael Skaggs
From the Board
Committee Briefs

Newspapers Project

Tech Corner

Collection Highlights

News from Our Members
News From CRRA

For Your Consideration

Save the Date!

The Jesuit Libraries Provenance Project

Kyle Roberts, Loyola University Chicago

The Jesuit Libraries Provenance Project (JLPP) was launched in March 2014 to create a visual archive of provenance marks from historic Jesuit college, seminary, and university library collections and to foster a participatory community interested in the history of these books.

Nineteenth-century Jesuits never met a book that they didn't like to stamp their name on. This stamp is found on books from Loyola's original library collection (c.1870).
Nineteenth-century Jesuits never met a book that they didn’t like to stamp their name on. This stamp is found on books from Loyola’s original library collection (c.1870).

Founded by students, faculty, and library professionals at Loyola University Chicago, the Provenance Project is an outgrowth of an earlier project [http://blogs.lib.luc.edu/archives/] to reconstruct the holdings listed in Loyola’s original (c.1878) library catalog in an innovative virtual library system. That project, which was the subject of a graduate seminar at Loyola in Fall 2013 and will launch later this year, brought together graduate students in Digital Humanities, History, and Public History to recreate the nineteenth-century library catalog in a twenty-first century open source Integrated Library System (ILS). In the course of researching the approximately 5100 titles listed in the original catalog, students discovered that upwards of 1750 might still be held in the collections of Loyola’s Cudahy Library, the Library Storage Facility, and University Archives and Special Collections. A handful of undergraduate and graduate students formed the Provenance Project the following semester to see how many of these books actually survived. As they pulled books off the shelves and opened them up, they discovered a range of provenance marks – bookplates, inscriptions, stamps, shelf-marks, and other notations – littering the inside covers, flyleaves, and title pages of these books. Students soon realized that if the original library catalog could tell them what books the Jesuits collected, provenance marks could reveal from where the books came.

 

The inside covers of books collected by Jesuit can have bookplates, stamps, and sometimes surprising marginalia. From The Spirit of Popery (n.d.)
The inside covers of books collected by Jesuit can have bookplates, stamps, and sometimes surprising marginalia. From The Spirit of Popery (n.d.)

By utilizing the freely accessible online social media image-sharing platform Flickr, the Provenance Project seeks to create a participatory community of students, bibliographers, academics, private collectors, alumni, and others interested in the origin and history of Jesuit-collected books. A photostream within the Provenance Project Flickr site allows visitors to scroll through all of the pictures that have been uploaded while commenting and tagging functions provide the opportunity to share their own knowledge about specific images. For example, visitors can contribute transcriptions of inscriptions (especially ones written in messy or illegible hands), translations of words and passages in foreign languages, and identifications of former individual and institution owners. Not only does the Flickr site provide a visual index of the rich variety of works held by a late nineteenth-century Jesuit college library, but it also inspires reflection and scholarship on the importance of print to Catholic intellectual, literary, and spiritual life.

The Provenance Project also encourages undergraduate and graduate students to undertake mentored primary-source research on the history of individual books as well as broader themes in Catholic and book history. Their findings are shared with the public in a variety of ways. One of the rooms in the Summer 2014 exhibition, Crossings and Dwellings: Restored Jesuits, Women Religious, American Experience 1814-2014 at the Loyola University Museum of Art (LUMA) featured original library books selected by graduate students and accompanied by interpretative labels they wrote. Student interns regularly contribute original scholarship to the Provenance Project’s website as well as to the June 2015 issue of the Catholic Library World on the “Digital Future of Jesuit Studies.” [Citation: “The Digital Future of Jesuit Studies,” Catholic Library World 85:4 (June 2015): 240-259.] They have also given talks on their research at conferences, such as the annual meeting of the American Catholic Historical Association. The 2014 commemoration of the bicentennial of the restoration of the Society of Jesus has brought renewed scholarly to nineteenth-century Jesuits. The work of Provenance Project interns is actively contributing to that resurgence of interest.

The Flickr photostream for the Jesuit Libraries Provenance Project.
The Flickr photostream for the Jesuit Libraries Provenance Project.

As of February 2016, students have tracked down all of the surviving books from the list of 1750 titles and are in the process of discerning how many of these titles are actual matches for those in the original catalog. (The answer appears to be the vast majority, making for a much higher survival rate than initially expected.) The team recently posted its 5000th image to the Flickr archive and still has many more images to upload over the coming months. Images on Flickr have also been usefully organized into albums either by nature of provenance mark (stamp, bookplate), part of book (illustrations, endpapers, binding), or division of the catalog (Pantology, Theology, Legislation, Philosophy, History, Literature). For those who would like to contribute to the Project, there are still many passages in need of translation and ownership marks in need of identification (helpfully gathered into the albums “Unidentified Inscriptions”, “Unidentified Stamps”, “Unidentified Embossed Stamps”, and “Unidentified Bookplates”).

Please follow the JLPP on Flickr (@JLPProject), Facebook and on Twitter (@JesuitProject). We try to post new books everyday and scholarship on the blog every week or so during the semester, so check back often!

A final note: the Provenance Project is beginning conversations about expanding the site to include provenance images from the collections of other historic Jesuit college, seminary, and university libraries. If you are interested in learning more about participating, or want information about how to start a project for your own institution, don’t hesitate to contact Kyle Roberts.

Interview with Michael Skaggs

Michael SkaggsMichael Skaggs is a doctoral candidate in the Department of History at the University of Notre Dame. He studies religion in the American Midwest, and is particularly interested in how interfaith organizations addressed social problems.

What is your current area of research?

Right now, I’m working on a dissertation chapter on Catholic racial activism in 1960s Cincinnati. Partly those men and women did so because they got involved in the contemporary civil rights movement, but the Second Vatican Council’s call for the laity to be active in society had something to do with it, too. But I think the blend of those two motivations is more complicated than it seems at first.

More generally, my dissertation asks how Catholics in one midwestern place – Cincinnati, Ohio – responded to the Second Vatican Council, and how the presence of a substantial Jewish community inflected that response. This presents us with a fascinating opportunity to understand the real richness of American Catholicism, which I think we miss out on if we overlook places like Cincinnati, which usually don’t seem to be all that important to us.

Graduate students in search of dissertation topics are well-positioned to draw attention to topics and places long untouched by scholars! And while there are many scholars across the career timeline ready to embrace digitization, I think the younger generation has a natural ability to work with these resources – maybe even an impatience to do things “the old way.” This is a transitional moment in academia, though, so there’s a real need for students and future scholars to straddle the line between technologies old and new.

How do you use CRRA’s resources for your research?

I first came to know about CRRA just after I had finished a research project on The Criterion, the Archdiocese of Indianapolis’s official newspaper. I did it the new-old-fashioned way: cranking through what felt like miles of microfilm. Now I work with CRRA’s newspaper digitization project and have been excited at the conversations surrounding getting these sources into a format that we can use quickly and easily.

CRRA has been particularly useful in considering how I might shape my research projects to benefit from digitization in the future. Since most of my sources are not yet digitized, it’s been wonderful to look ahead and consider what might reasonably be digitized in the future and the scholarly community that will arise around those sources. It’s exciting to think about being part of a conversation that more and more people enter as sources open up to easy access from afar.

What is the most exciting / surprising source you’ve been able to get access to for your research?

I have to point to the old-school method of research for this one, too, because Cincinnati doesn’t get the attention it really ought to – a lot of Catholic scholarship has been focused to this point on “more important” places in the American Church. So the biggest and most impressive collections that CRRA catalogs come from elsewhere – an imbalance that CRRA is sure to fix in coming years and as more and more diocesan archives get involved. But I would say the most exciting – or one of the most exciting sources – has been The American Israelite, which was published by and for American Reform Jews. It’s fully digitized but only accessible in certain locations – a prime opportunity for CRRA, since the Israelite reported on things Catholic quite often!

The American Israelite points up the potential of partnerships between the academy and religious institutions. While no small project, that one newspaper presented a relatively straightforward digitization task. And many organizations would only be too happy to let CRRA digitize their materials if the funding is available and it can be done in a reasonable amount of time. Furthermore, CRRA has utilized an excellent strategy of asking scholars themselves what they need access to, as this provides a clear (if not concise) idea of sources that might be targeted for digitization. Most scholars with particular research projects can identify exactly which collections it would be useful to digitize, which makes the process manageable, even if not all that easy. From there, related collections can be identified for future scholars who aren’t working with them just yet, or individual archives can propose collections that really ought to be made digital, and so on. It pretty plainly represents the future and I’m happy to know CRRA is working hard to get ahead of the game.

What do you wish you could have access to but is currently unavailable?

I sound like a broken record whenever I’m asked this in CRRA conversations: fully digitized diocesan newspapers from across the United States. I think that would open up research fields historians have not even begun to consider, especially since having all of that information readily available would really help us uncover the complexity of American Catholicism from place to place. Many dioceses have the entire run of their newspaper preserved, in some cases very well so. A program just for diocesan archives – and especially their newspapers – would be a fantastic way to bring these sources into the mainstream of academic research, especially those smaller or “less important” dioceses that historians haven’t thought of yet.

I also think that archival sources on parish histories are a goldmine yet to be tapped by most scholars. The problem here is accessibility, or even knowing where they are kept: more than once I have run into a parish saying their materials are held at the diocesan archives, while the diocesan archivist says the materials are at the parish! And in many cases people just haven’t saved much. But if we really want to know about American Catholicism, we desperately need access to the sources pertinent to the vast majority of American Catholics: the laity, who connect to the Church first at the parish level. These materials don’t need to be all that in-depth to provide something useful, either – I’d be perfectly happy with a solid set of parish bulletins over a given period of time, for example, for what it would tell us about parish life. Again, this is where CRRA is in a great place to help, through utilizing scholars’ needs and wants to identify, catalog, and digitize collections.

OAI and VuFind: Notes to self in the form of a recipe

The primary purpose of this posting is to document some of my experiences with OAI and VuFind. Specifically it outlines a sort of “recipe” I use to import OAI content into the “Catholic Portal“. The recipe includes a set of “ingredients”, site-specific commands. Towards the end, I ruminate on the use of OAI and Dublin Core for the sharing of metadata.

Philadelphia by Eric Morgan

Recipe

When I learn of a new OAI repository containing metadata destined for the Portal, I use the following recipe to complete the harvesting/indexing process:

  1. Use the OAI protocol directly to browse the remote data repository – This requires a slightly in-depth understanding how OAI-PMH functions, and describing it any additional detail is beyond the scope of this posting. Please consider perusing the OAI specification itself.
  2. Create a list of sets to harvest – This is like making a roux and is used to configure the oai.ini file, next.
  3. Edit/configure harvesting via oai.ini and properties files – The VuFind oai.ini file denotes the repositories to harvest from as well as some pretty cool configuration directives governing the harvesting process. Whomever wrote the harvester for VuFind did a very good job. Kudos!
  4. Harvest a set – The command for this step is in the list of ingredients, below. Again, this is very-well written.
  5. Edit/configure indexing via an XSL file – This is the most difficult part of the process. It requires me to write XSL, which is not too difficult in and of itself, but since each set of OAI content is often different from every other set, the XSL is set specific. Moreover, the metadata of the set is often incomplete, inconsistent, or ambiguous making the indexing process a challenge. In another post, it would behoove me to include a list of XSL routines I seem to use from repository to repository, but again, each repository is different.
  6. Test XSL output for completeness – The command for this step is below.
  7. Go to Step #5 until done – In this case “done” is usually defined as “good enough”.
  8. Index set – Our raison d’être, and the command is given below.
  9. Go to Step #4 for all sets – Each repository may include many sets, which is a cool OAI feature.
  10. Harvest and index all sets – Enhance the Portal.
  11. Go to Step #10 on a regular basis – OAI content is expected to evolve over time.
  12. Go to Step #1 on a less regular basis – Not only does content change, but the way it is described evolves as well. Harvesting and indexing is a never-ending process.

Ingredients

I use the following Linux “ingredients” to help me through the process of harvesting and indexing. I initialize things with a couple of environment variables. I use full path names whenever possible because I don’t know where I will be in the file system, and the VUFIND_HOME environment variable sometimes gets in the way. Ironic.

# configure; first the name of the repository and then a sample metadata file
  NAME=luc
  FILE=1455898167_lucoai_coll25_55.xml

  # (re-)initialize
  rm -rf /usr/local/vufind2/local/harvest/$NAME/*.delete
  rm -rf /usr/local/vufind2/local/harvest/$NAME/*

  # delete; an unfinished homemade Perl script to remove content from Solr
  /usr/local/vufind2/crra/crra-scripts/bin/solr-delete.pl

  # harvest; do the first part of the work
  cd /usr/local/vufind2/harvest/; php harvest_oai.php $NAME

  # test XSL output
  clear; \
  cd /usr/local/vufind2/import; \
  php ./import-xsl.php --test-only \
  /usr/local/vufind2/local/harvest/$NAME/$FILE \
  $NAME.properties

  # index; do the second part of the work
  /usr/local/vufind2/harvest/batch-import-xsl.sh $NAME $NAME.properties

Using the recipe and these ingredients, I am usually able to harvest and index content from a new repository a few hours. Of course, it all depends on the number of sets in the repository, the number of items in each set, as well as the integrity metadata itself.

Ruminations

As I have alluded to in a previous blog posting, the harvesting and indexing of OAI content is not straight-forward. In my particular case, the software is not to blame. No, the software is very well-written. I don’t take advantage of all of the software’s features though, but that is only because I do not desire to introduce any “-isms” into my local implementation. Specifically, I do not desire to mix PHP code with my XSL routines. Doing so seems too much like Fusion cuisine.

The challenge in this process is both the way Dublin Core is used, as well as the data itself. For example, is a PDF document a type of text? Sometimes it is denoted that way. There are dates in the metadata, but the dates are not qualified. Date published? Date created? Date updated? Moreover, the dates are syntactically different: 1995, 1995-01-12, January 1995. My software is stupid and/or I don’t have the time to normalize everything for each and every set. Then there are subjects. Sometimes they are Library of Congress headings. Sometimes they are just keywords. Sometimes there are multiple subjects in the metadata and they are enumerated in one field delimited by various characters. Sometimes these multiple subject “headings” are manifested as multiple dc.subject elements. Authors (creators) present a problem. First name last? Last name first? Complete with birth and death dates? Identifiers? Ack! Sometimes they include unique codes — things akin to URIs. Cool! Sometimes identifiers are URLs, but most of the time, these URLs point to splash pages of content management systems. Rarely do the identifiers point the item actually described by the metadata. And then there out & out errors. For example, description elements containing URLs pointing to image files.

Actually, none of this is new. Diane Hillmann & friends encountered all of these problems on a much grander scale through the National Science Foundation’s desire to create a “digital library”. Diane’s entire blog — Metadata Matters — is a cookbook for resolving these issues, but in my way of boiling everything done to their essentials, the solution is two-fold: 1) mutual agreements on how to manifest metadata, and 2) the writing of more intelligent software on my part.