10 Responses to Freeing Images from Inside Digitized Books and Newspapers

  1. Interesting idea, Trevor. Harvesting the images and providing access to them via a single interface would be a fabulous thing for so many people. Images hidden in text is a problem and, as you note, there is a plethora of them available thanks to massive book scanning efforts. I found it noteworthy that you state that you often examine the texts merely to examine the images. One of my gripes within the library and information science world is the near obsession with text to the detriment of providing access to all other formats. This has begun to shift noticeably in the past decade, thankfully, but it hasn’t yet addressed the issue you bring up here.

    From the way I see it, there are two issues at play here. The first is that many of these images will not be described in much detail (if at all) within the texts. This will mean that identification and description of the images will be problematic. The second issue is that the way that images can be described is variable. What is most important to mention? This might shift from person to person, or even be varied with a single person on different days. Having people with specific knowledge of the text, time period, and etc. helping with the description would certainly help things along, however. The legal and technical perspectives would seem to be easy part.

    If you are looking for folks to work on this with you, I am up for it. My research is all about visual information and image description and so I am very much interested in the topic.

  2. Thank you Trevor,
    I am in; I also collect images on Pinterest for fun and for scholarship, and the pop history site Anywhen is also a cool place to gather and extract interesting materials for academic use. I’d love to learn of other methods by which to do this sort of treasure hunt. It is like finding old letters in your neighbor’s attic – in fact it is – a trove of stuff by amateur historians and ordinary hoarders saved my dissertation! This is a great idea.

  3. It seems like it shouldn’t be horribly difficult to algorithmically determine images from text pages by looking for traits like ‘many parallel rows on bitonal image’ or ‘is colorful.’ If there was ever a catalog record created for a text, you’d also have some rough estimates for metadata/subject tags.

    Here’s an example of a project that algorithmically yanked images from archive.org scans (albeit with 80’s game magazines, so the assumptions don’t quite work with 19th century texts): cosmicrealms.com/blog/2012/12/31/c64-magazine-game-wallpaper-generator/

  4. It strikes me that there is a particular market/demand for this kind of thing: anywhere there are antiques or rare books for sale, there are invariably reams of plates cut out of old books, now sold as “singles” to be framed or something. In other words the image has a value well beyond that of whatever original volume housed it. Here of course we would be talking about a digital representation rather than a physical artifact, but I suspect there could still be tremendous interest in such an archive, particularly if it’s rich in metadata, tags, etc.

  5. Great idea, Trevor – both in terms of exposing the vast array of wonderful archival/digitized images and also to help further the visual research agenda. Developing theme-based collections could help scope the project a bit. As you know, the LIS community has had success utilizing crowd-sourced description for image collections so that might be an option. I could assist with user design requirements in terms of trying to figure out what makes an image “more usable” as visual evidence (I explore many of these topics in my dissertation research as we have discussed elsewhere.) Lev Manovich and his team have created an interesting tool for image aggregation with image plot: lab.softwarestudies.com/p/imageplot.html

  6. Would love to join such a session Trevor after what we discussed last week in our Italian Society for the Study of Photographs www-sisf.eu annual meeting when we pointed out how many newspapers and scholarly journals today are still using images in a passive way and not because images are tellingl us something by themselves, they speak their own language. For instance there’s no trace of any meta-information on pictures published in newspapers, not even the name of the photojournalist … So your idea to crowdsource images is nice, it is what google images or, better Tineye, is doing with the web answering our queries. Nice would be to fix it in a database for further retrievals with a caption and contexts, this is the added scholarly value that we all wait for on the web I suppose. Could I say that this info -also aimed at by Sean Takats above- is even more valuable sometime than accessing images that rarely are “speaking by themselves” -I mean for scholarly purposes- with no context nor story to interpreting them. Anyway just to tell you how much I would like to be there, wotkink a lot myself on how the digital turn has influenced photography and how same photographies have been “moving” in web sites. Maybe we could do something similar for the NCPH Monterey conference in March 2014 ?

  7. This is a great application of a common computer vision problem; challenging. Examples that are similar in flavor include these projects:

    Labeling Face Images in the Wild Link

    Label Me: Link

    Learning from a Visual Folksonomy: Automatically Annotating Images from Flickr Link

    Unlike these projects which are more general, the data from digital books should have more structure and information allowing for generation of good descriptive texts.

  8. I too am very interested in this session. I’m also curious how meta data can be completely applied as any given image can have multiple meanings based on context/audience.

  9. Funny you should mention this, since we have a grant-funded project at the Folger right now to digitize and index 10,000 images from 17th-century English books, and ONLY the images. Scans from high-contrast microfilm of most of the texts are already available through EEBO, with transcriptions becoming available through EEBO-TCP, but the visual language in EEBO is silenced. Of course, ideally, we’d scan a whole book and not just its engraved frontispiece, but 10,000 cover-to-cover shots from books won’t get us a corpus of images any time soon.