The document discusses the International Image Interoperability Framework (IIIF) which aims to make digital images, manuscripts, and other cultural heritage objects more accessible and shareable online. It describes how IIIF defines common standards for presenting digital objects through the use of manifests, canvases, and annotations. Manifests provide metadata to drive viewing experiences across institutions. Canvases define spaces for content, and annotations allow information like descriptions, transcriptions, and related resources to be linked to canvases. This allows digital objects to be recombined and explored through annotation.
4. Common themes
• Book readers/viewers/page turners
• Deep Zoom
• Transcriptions, OCR, full text
• Annotation, comments, content creation related to the object
• Linking to other objects
• Some bespoke format for delivering a sequence of images to a
bespoke viewer
5. Tens of thousands of silos of non-interoperable content
Grain elevators, Caldwell, Idaho, by Lee Russell, 1941.
http://www.loc.gov/pictures/resource/fsac.1a34206/
6. The reinvention of wheels
• Billions of digitised hi-res images of cultural heritage
• Millions of resources – books, maps, manuscripts…
• Thousands of beautiful snowflakes*
• We want to share the best tools and platforms
• We want to mix content and share resources across institutions,
across the world
• We want to bring our own tools to use with other people’s digital
objects
12. • How can I provide enough information for a software client to
present a complex digital object, AND provide a good user
experience?
• Don’t get entangled in descriptive metadata standards… especially in
libraries, museums and archives.
• We’re doing something else.
Regroup…
15. Digital Surrogate
I still have my cultural awareness when looking at a digital surrogate
…but the computer needs assistance in presenting that digital
surrogate to me
It needs to be mediated by metadata that can
drive a viewing experience so my human cultural awareness can take
over again
16. IIIF: Two Core APIs
“get pixels” via a
simple, RESTful,
web service
Just enough metadata to
drive a remote viewing
experience
Image API Presentation API
17. • Descriptive metadata
doesn’t help us get
pixels on screen
• Read the pages
• Look at the brush strokes
• See the film grain
• IIIF Presentation provides
• A model for describing digital representations of objects
• A format for software - viewing tools, annotation clients, web sites - to
consume and render the objects and the statements made about them
• The model defines a shared abstract space for rendering this content
18.
19. Manifests for things
A book. A painting. A film. A
sculpture. An opera. A long-
playing record. A manuscript. A
map. An aural history field
recording. A videocassette of a
public information film. A
laboratory notebook. A diary.
A manifest is what you open in
your player/viewer/annotation
workbench
20. A IIIF manifest
• …is a markup document for presenting digital objects
(not describing them)
• It’s also JSON. It’s also linked data.
• Driven by cultural heritage use cases, but not limited to them
• By publishing a manifest, you publish a list of one or more abstract
spaces to the web, which you can paint content on via annotation.
• Other people can make statements about - annotate - these spaces too.
33. Annotations provide content
• Any web resource
• Images, video, sound, text
• The same mechanism for association with the canvas is used for all
the other things we want to say, or other people want to say
34. Saying things about objects and collections
• Why is this object in the exhibition?
• Here is a blog post about this painting
• There’s no video available here, but this is the screenplay for this
missing 20 minutes of The Magnificent Ambersons
• …and here are some production stills for 49m23s-51m17s
• “Tourteletes in fryture” means “Small fig pies basted with honey”
• Here is the curator’s essay for this collection
• Here is a transcription of this letter
• Here is an article about this correspondence
• These four engravings are all views from this location
37. Another perspective
• This slide deck could be a IIIF manifest
• Each slide could be a canvas
• Text, images, audio and video annotated onto each canvas is the content
• I could put an image of a digitised manuscript page on each slide, send you the
deck representing the whole manuscript, and you could make notes on it
• Unlike .ppt, IIIF is JSON
• Specifically, JSON-LD: it’s also linked data
• It’s a lot easier to do things on the web with it
• Canvases have stable IDs so I can talk about your stuff
• We can do things with what people say about the stuff…
38. How do we find all these annotations? - The Search API
41. I have a vinyl album (LP)
It has two discs, each with 2 sides
I have made digital transfers of the discs; one file per side
I have made digital images of the sleeve, the liner notes and the disc labels.
AV archival item
42. Scenario 12
• I have two cassettes, each with two sides.
• They contain a single oral history interview.
• The interview starts on the first side of the
first cassette and ends on the final side of
the second cassette.
• I have made digital transfers from the
cassettes; one file per side.
• I have a plain text transcript with timing
indicators
43. Annotate time-based media onto the canvas
You can do this today, but you can’t target the time dimension
Without that only the very simplest use case is possible
Leaving aside the difference between presentation metadata and descriptive metadata for a moment, let’s take a look at the long history of digitisation and presenting digitised objects on the web. This has happened thousands of times over. Leading to the creation of many silos of non-interoperable content, un-reusable tools, and stagnating microsites for individual collections.
Video showing Mirador being used to compare celestial maps (Stu Snydman, Stanford)
Video showing comparison of IIIF resources from different institutions (Stu Snydman, Stanford)
OK, we need to have the pages in the right order
And structure to drive navigation within the book
And metadata to describe the pagination, reading direction
And we want deep zoom images
And we need metadata to show all the things that the user needs to see
Like the title and the author and what it’s about
And which of our collections it’s in
Wait, we could start using this to describe our objects more generally
And the material, the curator wants the binding material in the model
Does a painting have an author?
Hang on, this problem has been solved already
Let’s have a look at mappings to cool things like CIDOC-CRM…
What went wrong?
The cards in the library catalogue use a metadata schema to convey information such as author and subject. But it is not a requirement of enjoying a library, gallery or museum that we consult the card catalogue; we can just look at things, and the people who take care of the content arrange it so that we can find it. Usually the scheme used to organise the metadata informs the arrangement of objects in the real world – shelving, rooms, etc.
As humans, we don’t need to consult a metadata standard to understand what the text and images on this book cover or on this gallery label mean. When we have it in front of us, we use our shared cultural understanding of the world. We know the publisher, title and author of le Grand Meaulnes by looking at the “metadata” on the cover; we can pick it up and read it.
Link – UV and Mirador GO AND LOOK AT VIEWERS!
This is a manifest. There are millions of these things out there now, published by well over 100 institutions.
This space doesn’t have any content yet. How do we fill it? Through annotation.
This model for annotation is familiar. We have an image, and we make comments about the image or regions of it.
In IIIF, all content – even the image – is an annotation onto a canvas. The Canvas keeps the content separate from the conceptual model of the page of the book, or the painting, or the movie.
Here, lots of annotations have been made. In this case they are derived from OCR.
Here’s the same canvas with the image removed from the canvas. Just the text is painted on. Why might I do this?
No image. Better image later. Lost image.
I might have more than one image. Multispectral. All stacked on top of each other, aligned on the same canvas.
An example of annotating images onto a canvas.
And for real
What’s happening here?
And in the JSON – it’s not really a magic annotation space, it’s just JSON. But it is.
The content we assemble in our shared canvas space gets into that space through annotation
There were 2 annotations on the canvas as published. I’m going to create another one.
The mechanism IIIF uses to associate content with the shared canvas is a W3C Technical Recommendation – like CSS or HTML.
How do we find all these annotations?
This space doesn’t have any content yet. How do we fill it? Through annotation.
preserve the nation’s sound heritage
physical degradation and as the means of playing them disappear from production
15 years
Need IIIF AV now
Don’t want to wait a few years for IIIF to catch up
Maybe I have the words – but also a transcription
Need to capture and describe the objects the sound is from