Tilting at Windmills of Linked Data
Source: By Internet Archive Book Images [No restrictions], via Wikimedia Commons
“What giants?” Asked Sancho Panza.
“The ones you can see over there,” answered his master, “with the huge arms, some of which are very nearly two leagues long.”
“Now look, your grace,” said Sancho, “what you see over there aren’t giants, but windmills, and what seems to be arms are just their sails, that go around in the wind and turn the millstone.”
“Obviously,” replied Don Quixote, “you don’t know much about adventures.”
Let me start off by saying I am not now, nor have I ever been a member of the Commu–, er, a technologist.
@kshockey04 I often feel like metadata creators and library technologists are talking past each other.
— Erin Leach (@erinaleach) March 22, 2016
The original post that is referenced by the above tweets, and what got me thinking about how we are starting to deteriorate in regards to linked data is Roy Tennant’s Broken Furniture and Blood on the Floor. In it, Roy sets out what he sees as steps back from ‘good’ linked data (one example of what is meant by ‘good’ would be the now super-old-by-Internet-standards 5 Stars of Open Linked Data).
Roy argues that the move to better linked data is a bit more revolutionary than some in the library world realize. Although there are other concepts Roy does not mention (like the Open World Assumption vs. Closed World Assumption), Roy points out that the concept of a ‘record’ needs to be disposed of and replaced with the concept of graphs, which would anticipate the need for services tied to what he calls the ‘Bibliographic Graph’.
Kevin Shockey then wrote a rebuttal to Roy’s post, called Wherefore art thou, linked data?. In it, he claims that records are still important. In fact, he asserts records are even more important in a linked data environment.
As the tweets above show, there’s already a false “us vs. them” dichotomy set up between “metadata creators” and “technologists”. Interestingly, Shockey has accused me of creating a false dichotomy, except in his mind, it’s between records and graphs. What’s strange is that the issues brought up with Shockey’s post have been limited to myself and Roy, who cannot be easily classified as merely “metadata creator” or the dreaded “technologist”. Look, I have been in Metadata Services for almost 2 years now. IT is a separate department. But it would be silly for us to rope ourselves into imaginary tribes of metadata vs. tech. We are not different people, and that mental divide contributes to a lot of the inertia and rot that has occurred in academic libraries. Throughout his career, Roy has highlighted this false dichotomy, which acts against the mission of academic libraries. He has also been one of the people in libraries to discard with an old way of thinking, as in his seminal work MARC Must Die (2002).
So, are we supposed to assume that the graph (or a resource composed of various statements from a graph) is a record? Or that a record is now composed of child records? Neither of those statements are accurate. What I believe Shockey is getting at here is that is how digital objects often look like records. To take an example from my institution, the page you can see for the digital object at http://library.ucsd.edu/dc/object/bb6213187w… looks like a record and we can call it a day, right? But what is the data behind that object? What is the data structure underlying that object?
The answer is that the data behind that object is in RDF. And our data model (which is ever-evolving) is a distinct departure from the MODS-based, key-value pair model that preceded it. The end result is we are definitively moving from records to a graph-based representation of data. Everything we assert in this object is a series of triples and part of a giant triplestore (as opposed to a bunch of key-value pair records held in a database).
To highlight how records and graphs of entities differ, let’s use an example of the Principal Investigator on this object, Newell Booth. In MODS, or any kind of system that relies on record-based metadata schema, imagine we find out there’s a typo on his name on this one object. Since again this hypothetical is situated in a records-based environment, I have to assume that this typo could appear in multiple other records elsewhere in the database: or, it may not! I suppose we would query the database with SQL in order to find that typo in any other records, and update the incorrect values in each record accordingly.
In a linked data environment, however, I know that the typo must be at some URI (in our case, an ARK) for the name/entity ‘Newell Booth’. This is because in linked data, it’s all about “things”, and one of those things is entities. Every “things” gets a URI, so I just need to know this one URI. That happens to be at http://library.ucsd.edu/ark:/20775/bb49503739
(Don’t bother trying to navigate to that, it won’t resolve). I can use a handy internal tool we have to visit that ARK, and edit the value. Once that occurs, the change will disseminate to any object in our system that contains that URI. In fact, the object I linked initially may display the name as a hyperlink, but that text does not actually occur in the object’s data. The system pulls the text via the URI. You could imagine the same thing happening if we had a LoC URI, except I can’t really edit those, can I? Can I just ask LoC for this ‘record’?
We could wave our hands and say, well, this doesn’t matter: that object is still a record. Fine. I disagree with you, but fine. However, you cannot say that sticking with the records-based environment will allow you to design good linked data models and ontologies. Because in the mindset, you will make assumptions that eventually run counter to the ideas and best practices inherent in linked data. This has occurred in BIBFRAME multiple times, as when they had a property <bf:authorizedAccessPoint /> but then to their credit realized, why not just use <rdfs:label />? And it will continue to occur as people insist FRBR belongs in BIBFRAME, when FRBR and RDF models differ in important ways. As Karen Coyle points out:
The main point is that there are so many contradictions within FRBR that we really must see it as a flawed model, something that needs to be fixed before we try to do anything with it. Yet, for much of the library world it is being taken as gospel - and not just gospel, but the model for our future data standards.
And yet, we have <bf:hasExpression />. Is it crazy to suggest this has arisen out of conformance to past thinking, records-based thinking? Accepting that past mentality as gospel, and not incorporating enough new ways of thinking?
So, call it a record if you like. But heed Roy’s overall message regardless, because he’s making very good points.