<?xml version="1.0" encoding="utf-8" ?>

<rss version="2.0" 
   xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
   xmlns:admin="http://webns.net/mvcb/"
   xmlns:dc="http://purl.org/dc/elements/1.1/"
   xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
   xmlns:wfw="http://wellformedweb.org/CommentAPI/"
   xmlns:content="http://purl.org/rss/1.0/modules/content/"
   >
<channel>
    
    <title>Coffee|Code : Dan Scott - Structured data</title>
    <link>https://coffeecode.net:443/</link>
    <description>Caffeinated Librarian Geek</description>
    <dc:language>en</dc:language>
    <generator>Serendipity 1.6.2 - http://www.s9y.org/</generator>
    
    

<item>
    <title>Library and Archives Canada: Planning for a new union catalogue</title>
    <link>https://coffeecode.net:443/archives/302-Library-and-Archives-Canada-Planning-for-a-new-union-catalogue.html</link>
            <category>Coding</category>
            <category>Structured data</category>
    
    <comments>https://coffeecode.net:443/archives/302-Library-and-Archives-Canada-Planning-for-a-new-union-catalogue.html#comments</comments>
    <wfw:comment>https://coffeecode.net:443/wfwcomment.php?cid=302</wfw:comment>

    <slash:comments>1</slash:comments>
    <wfw:commentRss>https://coffeecode.net:443/rss.php?version=2.0&amp;type=comments&amp;cid=302</wfw:commentRss>
    

    <author>dan@coffeecode.net (Dan Scott)</author>
    <content:encoded>
    &lt;p&gt;&lt;strong&gt;Update 2015-03-03&lt;/strong&gt;: Clarified (in the Privacy section) that only NRCan runs Evergreen.&lt;/p&gt;
&lt;p property=&quot;description&quot;&gt;I attended a meeting with Library and Archives Canada today in my role as an &lt;a href=&quot;http://accessola.org&quot;&gt;Ontario Library Association&lt;/a&gt; board member to discuss the plans around a new Canadian union catalogue based on OCLC&#039;s hosted services. Following are some of the thoughts I prepared in advance of the meeting, based on the relatively limited materials to which I had access. (I will update this post once those materials have been shared openly; they include rough implementation timelines, perhaps the most interesting of which being that it the replacement system is not expected to be in production until August 2016.) Let me say at the outset that there were no solid answers on potential costs to participating libraries, other than that LAC is striving to keep the costs as low as possible.&lt;/p&gt;

&lt;h3&gt;Basic question: What form does LAC envision the solution taking?&lt;/h3&gt;
&lt;p&gt;Will it be:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&quot;Library and Archives Canada begins adding records and holdings to WorldCat&quot; as listed for many other countries in http://www.oclc.org/worldcat/catalog/national/timeline.en.html;&lt;/li&gt;
&lt;li&gt;Or a separate, standalone but openly searchable WorldCat Local catalogue that Canadians can use like the &lt;a href=&quot;http://adamnet.worldcat.org&quot;&gt;Dutch&lt;/a&gt; or &lt;a href=&quot;http://fablibraries.worldcat.org/&quot;&gt;United Kingdom&lt;/a&gt; union catalogues (which lack significant functionality that standard WorldCat possesses, like the integrated schema.org discovery markup)?&lt;/li&gt;
&lt;li&gt;Or a separate, standalone but closed catalogue like the &lt;a href=&quot;https://www.oclc.org/nl-NL/ggc.html&quot;&gt;Dutch union catalogue GGC&lt;/a&gt; and the &lt;a href=&quot;http://www.oclc.org/en-UK/unityuk.html&quot;&gt;Combined Regions UnityUK&lt;/a&gt; that require a subscription to access?&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The answer was &quot;yes, we will be adding records and holdings to WorldCat, and yes, you will be able to search a WorldCat Local instance for both LAC-specific and AMICUS as a whole&quot; - but they&#039;re still working out the exact details. Later we determined that it will actually be WorldCat Discovery--essentially a rewrite of WorldCat Local--which assuaged some of my concerns about the current examples we can see of other OCLC-based union catalogues.&lt;/p&gt;

&lt;h3&gt;Privacy of Canadian citizens&lt;/h3&gt;
&lt;p&gt;
The &quot;Canadian office and data centre locations&quot; requirement does not mean that usage data is exempt from Patriot Act concerns. Specifically, OCLC is an American company and thus the USA Patriot Act &quot;allows US authorities to obtain records from any US-linked company operating in Canada&quot; (per a &lt;a href=&quot;https://cippic.ca/en/US_access_to_Canadian_data&quot;&gt;2004 brief submitted to the BC Privacy Commissioner by CIPPIC&lt;/a&gt;). Canadians should not be subject to this invasion of their privacy by the agents of another nation simply to use their own national union catalogue.
&lt;/p&gt;
&lt;p&gt;&lt;em&gt;The response:&lt;/em&gt; The Justice, Agricultural, and NRCan agencies use US-hosted library systems (the latter running the open-source Evergreen, by Equinox). However, one of the other participants from a federal agency reported that they had been trying to update to Sierra from their Millenium instance but have been stalled for two years because whatever policy allowed them to go live with US-hosted Millenium is not being allowed now.&lt;/p&gt;
&lt;p&gt;LAC claimed that, due to NAFTA, they are not allowed to insist that data be held in Canada unless it is for national security reasons. They noted that any usage data collected wouldn&#039;t be the same volume of patron data that would be seen in public libraries. They did point out that Netherlands sends anonymized data to OCLC, but that costs money and impacts response time. Apparently the OCLC web site, they claim not to have had a request under Patriot Act.&lt;/p&gt;
&lt;h3&gt;Privacy of Canadian citizens, part 2&lt;/h3&gt;
&lt;p&gt;&lt;em&gt;I didn&#039;t get the chance to bring this up during the call...&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;
LAC noted in their background that modern systems have links to social media, and apparently want this as part of a new AMICUS. This would also open up potential privacy leaks; see &lt;a href=&quot;http://go-to-hellman.blogspot.ca/2014/12/stop-making-web-surveillance-bugs-by.html&quot;&gt;Eric Hellman on this topic&lt;/a&gt;, for example; it is also an area of interest for the recently launched &lt;a href=&quot;http://www.ala.org/lita/about/igs/public/lit-Pp&quot;&gt;ALA Patron Privacy Technologies Interest Group&lt;/a&gt;.
&lt;/p&gt;

&lt;h3&gt;Open data&lt;/h3&gt;
&lt;p&gt;Opening up access to data is part of the federal government&#039;s stated mission.  
&lt;a href=&quot;http://open.canada.ca/en/content/canadas-action-plan-open-government-2014-16&quot;&gt;Canada&#039;s Action Plan on Open Government 2014-16&lt;/a&gt; says &quot;Open Government Foundation - Open By Default&quot; is a keystone of its plan; &quot;Eligible data and information will be released in standardized, open formats, free of charge, and without restrictions on reuse&quot; under the Open Government Licence - Canada 2.0. I therefore asserted:
&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;A relaunched National Union Catalogue should therefore support open data per the federal initiative from launch.&lt;/li&gt;
&lt;li&gt;The open data should include bibliographic, authority, and holdings records. Guy Berthiaume&#039;s reply to &lt;a href=&quot;http://www.cla.ca/Content/NavigationMenu/CLAatWork/Advocacy/LAC_Response_CLA_letter_OCLC_access_26aug2014.pdf&quot;&gt;CLA&lt;/a&gt; and &lt;a href=&quot;http://capalibrarians.org/wp/wp-content/uploads/2014/09/LAC-response-September-62014.pdf&quot;&gt;CAPAL&lt;/a&gt; that libraries can use the Z39.50 protocol to try to access records from individual library&#039;s Z39.50 servers ignores one of the primary purposes of a union catalogue, which is to avoid that time-consuming search across the various Z39.50 servers of the institutions that contributed their data to the union catalogue in the first place.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;em&gt;The response:&lt;/em&gt; The ACAN requirements document indicated a requirement that the data be made available under an ODC-BY license (matching OCLC&#039;s general WorldCat license); and LAC needs to get the data back to support their federated search tool.&lt;/p&gt;
&lt;p&gt;I asked if they had checked to see if ODC-BY and Open Government License - Canada 2.0 licenses are compatible; they responded that that was something they would need to look into.
Happily, the &lt;a href=&quot;http://clipol.org/licences/69?tab=licence_compatibility&quot;&gt;CLIPol tool&lt;/a&gt; indicates that the ODB-BY 1.0 and Open Government License - Canada 2.0 licenses are mostly compatible.&lt;/p&gt;

&lt;h3&gt;Contemporary features: are we achieving the stated goals?&lt;/h3&gt;
&lt;p&gt;
The backgrounder benefits/objectives section stated: &quot;In the current AMICUS?based context, the NUC has not kept pace with new technological functions, capabilities, and client needs.  Contemporary features such as a user?oriented display and navigation, user customization, links to social media, and linked open data output were not available when AMICUS was implemented in the 1990s.&quot;
&lt;/p&gt;
&lt;h4&gt;Canadian resource visibility&lt;/h4&gt;
&lt;p&gt;
To preserve and promote our unique national culture, we want Canadian library resources to be as visible as possible on the web. This is generally accomplished by publishing a sitemap (a list of the web pages for a given web site, along with when each page was last updated) and allowing search engines like Google, Bing, and Yahoo to crawl those web pages and index their data.
&lt;/p&gt;
&lt;p&gt;
To maximize the visibility of Canadian library resources on the open web, we need our union catalogue to generate a sitemap that points to only the actual records with holdings for Canadian libraries, not just WorldCat.org in general. For example, &lt;a href=&quot;http://adamnet.worldcat.org/robots.txt&quot;&gt;http://adamnet.worldcat.org/robots.txt&lt;/a&gt; simply points to the generic http://www.worldcat.org/libraries/sitemap_index.xml, not a specific sitemap for the Dutch union catalogue.
&lt;/p&gt;
&lt;p&gt;Our union catalogue should publish schema.org metadata to improve the discoverability of our resources in search engines (which initiated the schema.org standard for that purpose). WorldCat includes schema.org metadata, but WorldCat Local instances do not.&lt;/p&gt;
&lt;p&gt;&lt;em&gt;The response:&lt;/em&gt; There was some confusion about schema.org, and they asked if I didn&#039;t think that OCLC&#039;s syndication program was sufficient for enabling web discoverability. I replied in the negative.&lt;/p&gt;
&lt;h4&gt;Standards support (MARC21, RDA, ISO etc.)&lt;/h4&gt;
&lt;p&gt;&lt;em&gt;I didn&#039;t get a chance to raise these questions.&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;What standards, exactly, are meant by this?&lt;/p&gt;
&lt;p&gt;&quot;Technical requirements including volumetrics and W3C compliance&quot; is also very broad and vague. With respect to &quot;W3C compliance&quot;, &lt;a href=&quot;http://www.w3.org/standards/&quot;&gt;W3C Standards&lt;/a&gt; is just the start of many standards.&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Presumably there will be WCAG compliance for accessibility - but to what extent?&lt;/li&gt;
&lt;li&gt;Both the adamnet and fablibraries instances landing pages state that their canonical URL is www.worldcat.org, which effectively hides them from search engines.&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;Mobile support&lt;/h4&gt;
&lt;p&gt;The &lt;a href=&quot;http://www.w3.org/standards/&quot;&gt;W3C Standards&lt;/a&gt; page mentions mobile friendliness as part of its standards.&lt;/p&gt;
&lt;p&gt;
WorldCat.org itself is not mobile friendly. It uses a separate website with different URLs to serve up mobile web pages, and does not automatically detect mobile browsers; the onus is on the user to find the &lt;a href=&quot;http://www.worldcatmobile.org/&quot;&gt;&quot;WorldCat Mobile&quot; page&lt;/a&gt;, and that has been in a &quot;Beta&quot; state since 2009. The &quot;beta&quot; contravenes the stated requirements for the AMICUS replacement service to &lt;em&gt;not&lt;/em&gt; be an alpha or beta, unless you choose to ignore the massive adoption of mobile devices for searching and browsing purposes, and the beta mobile experience lacks functionality compared to the desktop version.
&lt;/p&gt;
&lt;p&gt;
The adamnet and fablibraries WorldCat Local instances don&#039;t advertise the mobile option, which is slightly different than the standard WorldCat Mobile version (for example, it offers record detail pages), but the navigation between desktop and mobile is sub-par. If you have bookmarked a page on the desktop, then open that bookmark on your synchronized browser on a mobile device, you can only get the desktop view.
&lt;/p&gt;

&lt;h4&gt;Linked open data&lt;/h4&gt;
&lt;p&gt;Linked open data around records, holdings, and participating libraries has arguably been a standard since the W3 Library Linked Data working group issued its &lt;a href=&quot;http://www.w3.org/2005/Incubator/lld/XGR-lld-20111025/ &quot;&gt;final report in 2011&lt;/a&gt;.&lt;/p&gt;
&lt;ul&gt; 
&lt;li&gt;Data--including library holdings--should be available both as bulk downloads and as linked open data&lt;/li&gt;
&lt;li&gt;Records need to be linked to libraries and holdings. For humans, that missing link in WorldCat is supplied by a JavaScript lookup based on geographic location info that the human supplies. This prevents other automated services from aggregating the data and creating new services based on it (including entirely Canadian-built and hosted services which would then protect Canadians from USA Patriot Act concerns).&lt;/li&gt;
&lt;li&gt;MARC records should be one of the directly downloadable formats via the web. Currently download options are limited to experimental &amp;amp; incomplete ntriple, turtle, JSON-LD, and RDF-XML formats.&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;Application programming interface (API)&lt;/h4&gt;
&lt;p&gt;&lt;em&gt;I didn&#039;t get the chance to bring this up during the call...&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;OCLC offers the xID API in a very limited fashion to non-members, which is one of the only ways to match ISBN, LCCN, and OCLC numbers. LAC should ensure that Canadian libraries have access to some similarly efficient means of finding matching records without having to become full OCLC Cataloguing members.&lt;/p&gt;

&lt;h4&gt;Updating the NUC&lt;/h4&gt;
&lt;p&gt;&lt;em&gt;I didn&#039;t get the chance to bring this up during the call...&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;In an ideal world, the NUC would adopt the standard web indexing practice of checking sitemaps (for those libraries that produce them) on a regular (daily or weekly basis) and add/replace any new/modified records &amp;amp; holdings from the contributing libraries accordingly, rather than requiring libraries to upload their own records &amp;amp; holdings on an irregular basis.&lt;/p&gt; 
    </content:encoded>

    <pubDate>Mon, 02 Mar 2015 22:46:47 -0500</pubDate>
    <guid isPermaLink="false">https://coffeecode.net:443/archives/302-guid.html</guid>
    <category>coding</category>
<category>structured data</category>

</item>
<item>
    <title>Putting the &quot;Web&quot; back into Semantic Web in Libraries 2014</title>
    <link>https://coffeecode.net:443/archives/296-Putting-the-Web-back-into-Semantic-Web-in-Libraries-2014.html</link>
            <category>Coding</category>
            <category>Evergreen</category>
            <category>Libraries</category>
            <category>Structured data</category>
    
    <comments>https://coffeecode.net:443/archives/296-Putting-the-Web-back-into-Semantic-Web-in-Libraries-2014.html#comments</comments>
    <wfw:comment>https://coffeecode.net:443/wfwcomment.php?cid=296</wfw:comment>

    <slash:comments>0</slash:comments>
    <wfw:commentRss>https://coffeecode.net:443/rss.php?version=2.0&amp;type=comments&amp;cid=296</wfw:commentRss>
    

    <author>dan@coffeecode.net (Dan Scott)</author>
    <content:encoded>
    &lt;p&gt;I was honoured to lead a workshop and speak at this year&#039;s edition of
&lt;a href=&quot;http://swib.org/swib14&quot;&gt;Semantic Web in Bibliotheken (SWIB)&lt;/a&gt; in Bonn, Germany. It was an amazing
experience; there were so many rich projects being described with obvious
dividends for the users of libraries, once again the European library
community fills me with hope for the future success of the semantic web.
&lt;/p&gt;

&lt;p&gt;
The subject of my talk &quot;Cataloguing for the open web with RDFa and schema.org&quot;
(&lt;a href=&quot;https://coffeecode.net/swib14/talk&quot;&gt;slides&lt;/a&gt; and &lt;a
href=&quot;http://www.scivee.tv/node/63282&quot;&gt;video recording&lt;/a&gt; - &lt;em&gt;gulp&lt;/em&gt;)
pivoted while I was preparing materials for the workshop. I was searching
library catalogues around Bonn looking for a catalogue with persistent URIs
that I could use for an example. To my surprise, catalogue after catalogue used
session-based URLs; it took me quite some time before I was able to find ULB,
who had hosted a VuFind front end for their catalogue. Even then, the
&lt;code&gt;robots.txt&lt;/code&gt; restricted crawling by any user agent. This reminded me
rather depressingly of my findings from current &quot;discovery layers&quot;, which
entirely restrict crawling and therefore put libraries into a black hole on the
web.
&lt;/p&gt;

&lt;p&gt;
Thses findings in the wild are so antithetical to the basic principles of
enabling discovery of web resources that, in a conference about the semantic
web, I opted to spend over half of my talk making the argument that libraries
need to pay attention to the old-fashioned web of documents first and foremost.
The basic building blocks that I advocated were, in priority order:
&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Persistent URIs, on which everything else is built&lt;/li&gt;
&lt;li&gt;Sitemaps, to facilitate discovery of your resources&lt;/li&gt;
&lt;li&gt;A robots.txt file to filter portions of your website that should not be
    crawled (for example, search results pages)&lt;/li&gt;
&lt;li&gt;RDFa, microdata, or JSON-LD only after you&#039;ve sorted out the first three&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;
Only after setting that foundation did I feel comfortable launching into my
rationale for RDFa and schema.org as a tool for enabling discovery on the web:
a mapping of the access points that cataloguers create to the world of HTML
and aggregators. The key point for SWIB was that RDFa and schema.org can enable
full RDF expressions in HTML; that is, we can, should, and must go beyond
surfacing structured data to surfacing linked data through
&lt;code&gt;@resource&lt;/code&gt; attributes and 
&lt;a href=&quot;http://schema.org/sameAs&quot;&gt;schema:sameAs&lt;/a&gt; properties.
&lt;/p&gt;


&lt;blockquote&gt;
The Semantic Web is an extension of the current web in which information is
given well-defined meaning, better enabling computers and people to work in
cooperation. &lt;cite&gt;Tim Berners-Lee, Scientific American, 2001&lt;/cite&gt;
&lt;/blockquote&gt;

&lt;p&gt;
I also argued that using RDFa to enrich the document web was, in fact, truer to
Berners-Lee&#039;s 2001 definition of the semantic web, and that we should focus on
enriching the document web so that both humans and machines can benefit before
investing in building an entirely separate and disconnected semantic web.
&lt;/p&gt;

&lt;p&gt;
I was worried that my talk would not be well received; that it would be
considered obvious, or scolding, or just plain off-topic. But to my relief
I received a great deal of positive feedback. And on the next day, both Eric Miller
and Richard Wallis gave talks on a similar, but more refined, theme:
that libraries need to do a much, much better job of enabling their resources
to be found on the web--not by people who already use our catalogues, but by
people who are &lt;em&gt;not&lt;/em&gt; library users today.
&lt;/p&gt;

&lt;p&gt;
There were also some requests for clarification, which I&#039;ll try to address
generally here (for the benefit of anyone who wasn&#039;t able to talk with me, or
who might watch the livestream in the future).
&lt;/p&gt;

&lt;h4&gt;&quot;When you said anything could be described in schema.org, did you mean we should throw out MARC and BIBFRAME and EAD?&quot;&lt;/h4&gt;

&lt;p&gt;&lt;em&gt;tldr:&lt;/em&gt; I intended &lt;strong&gt;and&lt;/strong&gt;, not &lt;strong&gt;instead of&lt;/strong&gt;!&lt;/p&gt;

&lt;p&gt;
The first question I was asked was whether there was anything that I had not
been able to describe in schema.org, to which I answered &quot;No&quot;--especially since
the work that the W3C SchemaBibEx group had done to ensure that some of the core
bibliographic requirements were added to the vocabulary. It was not as
coherent or full a response as I would have liked to have made; I blame the
livestream camera &lt;img src=&quot;https://coffeecode.net:443/templates/default/img/emoticons/smile.png&quot; alt=&quot;:-)&quot; style=&quot;display: inline; vertical-align: bottom;&quot; class=&quot;emoticon&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;
But combined with a part of the presentation where I countered a myth about
schema.org being a very coarse vocabulary by pointing out that it actually
contained 600 classes and over 800 properties, a number of the attendees
interpreted one of the takeaways of my talk as suggesting that libraries should
adopt schema.org as &lt;em&gt;the&lt;/em&gt; descriptive vocabulary, and that MARC,
BIBFRAME, EAD, RAD, RDA, and other approaches for describing library resources
were no longer necessary.
&lt;/p&gt;

&lt;p&gt;
This is not at all what I&#039;m advocating! To expand on my response, you
&lt;em&gt;can&lt;/em&gt; describe anything in schema.org, but you might lose significant
amounts of richness in your description. For example, short stories and poems
would best be described in schema.org as a &lt;a href=&quot;http://schema.org/CreativeWork&quot;&gt;CreativeWork&lt;/a&gt;.
You would have to look at the associated description or keyword properties to
be able to figure out the form of the work.
&lt;/p&gt;

&lt;p&gt;
What I was advocating was that you should map your rich bibliographic
description into corresponding schema.org classes and properties in RDFa at the
time you generate the HTML representation of that resource and its associated
entities. So your poem might be represented as a &lt;a
href=&quot;http://schema.org/CreativeWork&quot;&gt;CreativeWork&lt;/a&gt;, with a
&lt;a href=&quot;http://schema.org/name&quot;&gt;name&lt;/a&gt;,
&lt;a href=&quot;http://schema.org/author&quot;&gt;author&lt;/a&gt;,
&lt;a href=&quot;http://schema.org/description&quot;&gt;description&lt;/a&gt;,
&lt;a href=&quot;http://schema.org/keywords&quot;&gt;keywords&lt;/a&gt;, and
&lt;a href=&quot;http://schema.org/about&quot;&gt;about&lt;/a&gt; values and relationships. Ideally,
the &lt;code&gt;author&lt;/code&gt; will include at least one link (either via 
&lt;a href=&quot;http://schema.org/sameAs&quot;&gt;sameAs&lt;/a&gt;,
&lt;a href=&quot;http://schema.org/url&quot;&gt;url&lt;/a&gt;, or &lt;code&gt;@resource&lt;/code&gt;) to an
entity on the web; and you could do the same with &lt;code&gt;about&lt;/code&gt; if you are
using a controlled vocabulary.
&lt;/p&gt;

&lt;p&gt;
If you take that approach, then you can serve up schema.org descriptions of works
in HTML that most web-oriented clients will understand (such as search engines)
and provide basic access points such as name / author / keywords, while
retaining and maintaining the full richness of the underlying bibliographic
description--and potentially providing access to that, too, as part of the
embedded RDFa, via content negotiation, or &lt;code&gt;&amp;lt;link rel=&quot;&quot;&amp;gt;&lt;/code&gt;,
for clients that can interpret richer formats.
&lt;/p&gt;

&lt;h4&gt;&quot;What makes you think Google will want to surface library holdings in search results?&quot;&lt;/h4&gt;

&lt;p&gt;
There is a perception that Google and other search engines just want to sell
ads, or their own products (such as Google Books). While Google certainly does
want to sell ads and products, they also want to be the most useful tool for
satisfying users&#039; information needs--possibly so they can learn more about those
users and put more effective ads in front of them--but nonetheless, the
motivation is there.
&lt;/p&gt;

&lt;p&gt;
Imagine marking up your resources with the Product / Offer portion of schema.org
you are able to provide search engines with availability information in the
same way that Best Buy, AbeBooks, and other online retailers do (as Evergreen,
Koha, and VuFind already do). That makes it much easier for the search engines
to use everything they may know about their users, such as their current
location, their institutional affiliations, their typical commuting patterns,
their reading and research preferences... to provide a link to a library&#039;s
electronic or print copy of a given resource in a knowledge graph box as one of
the possible ways of satisfying that person&#039;s information needs.
&lt;/p&gt;

&lt;p&gt;
We don&#039;t see it happening with libraries running Evergreen, Koha, and VuFind
yet, realistically because the open source library systems don&#039;t have enough
penetration to make it worth a search engine&#039;s effort to add that to their
set of possible sources. However, if we as an industry make a concerted effort
to implement this as a standard part of crawlable catalogue or discovery record
detail pages, then it wouldn&#039;t surprise me in the least to see such suggestions
start to appear. The best proof that we have that Google, at least, is
interested in supporting discovery of library resources is the continued
investment in Google Scholar.
&lt;/p&gt;

&lt;p&gt;
And as I argued during my talk, even if the search engines never add direct
links to library resources from search results or knowledge graph sidebars,
having a reasonably simple standard like the GoodRelations product / offer
pattern for resource availability enables new web-based approaches for building
appplications. One example could be a fulfillment system that uses sitemaps to
intelligently crawl all of its participating libraries, normalizes the
item request to a work URI, and checks availability by parsing the offers at the
corresponding URIs.
&lt;/p&gt; 
    </content:encoded>

    <pubDate>Thu, 04 Dec 2014 16:15:15 -0500</pubDate>
    <guid isPermaLink="false">https://coffeecode.net:443/archives/296-guid.html</guid>
    <category>coding</category>
<category>evergreen</category>
<category>libraries</category>
<category>structured data</category>

</item>
<item>
    <title>How discovery layers have closed off access to library resources, and other tales of schema.org from LITA Forum 2014</title>
    <link>https://coffeecode.net:443/archives/294-How-discovery-layers-have-closed-off-access-to-library-resources,-and-other-tales-of-schema.org-from-LITA-Forum-2014.html</link>
            <category>Coding</category>
            <category>Evergreen</category>
            <category>Structured data</category>
    
    <comments>https://coffeecode.net:443/archives/294-How-discovery-layers-have-closed-off-access-to-library-resources,-and-other-tales-of-schema.org-from-LITA-Forum-2014.html#comments</comments>
    <wfw:comment>https://coffeecode.net:443/wfwcomment.php?cid=294</wfw:comment>

    <slash:comments>0</slash:comments>
    <wfw:commentRss>https://coffeecode.net:443/rss.php?version=2.0&amp;type=comments&amp;cid=294</wfw:commentRss>
    

    <author>dan@coffeecode.net (Dan Scott)</author>
    <content:encoded>
    &lt;p property=&quot;description&quot;&gt;
At the LITA Forum yesterday, I accused (&lt;a href=&quot;http://stuff.coffeecode.net/2014/lita_forum&quot;&gt;presentation&lt;/a&gt;) most discovery layers of not solving the discoverability problems of libraries, but instead exacerbating them by launching us headlong to a closed, unlinkable world. Coincidentally, Lorcan Dempsey&#039;s opening keynote contained a subtle criticism of discovery layers. I wasn&#039;t that subtle.&lt;/p&gt;
&lt;p&gt;
Here&#039;s why I believe commercial discovery layers are not &quot;of the web&quot;: check out their &lt;a href=&quot;http://robotstxt.org&quot;&gt;&lt;code&gt;robots.txt&lt;/code&gt;&lt;/a&gt; files. If you&#039;re not familiar with robots.txt files, these are what search engines and other well-behaved automated crawlers of web resources use to determine whether they are allowed to visit and index the content of pages on a site. Here&#039;s what the &lt;code&gt;robots.txt&lt;/code&gt; files look like for a few of the best-known discovery layers:
&lt;/p&gt;
&lt;pre&gt;
User-Agent: *
Disallow /
&lt;/pre&gt;
&lt;p&gt;
That effectively says &quot;Go away, machines; your kind isn&#039;t wanted in these parts.&quot; And that, in turn, closes off access to your libraries resources to search engines and other aggregators of content, and is completely counter to the overarching desire to evolve to a linked open data world.
&lt;/p&gt;
&lt;p&gt;
During the question period, Marshall Breeding challenged my assertion as being unfair to what are meant to be merely indexes of library content. I responded that most libraries have replaced their catalogues with discovery layers, closing off open access to what have traditionally been their core resources, and he rather quickly acquiesced that that was indeed a problem.
&lt;/p&gt;
&lt;p&gt;
(By the way, a possible solution might be to simply offer two different URL patterns, something like &lt;code&gt;/library/*&lt;/code&gt; for library-owned resources to which access should be granted, and &lt;code&gt;/licensed/*&lt;/code&gt; for resources to which open access to the metadata is problematic due to licensing issues, and which robots can therefore be restricted from accessing.)
&lt;/p&gt;
&lt;p&gt;
Compared to commercial discovery layers on my very handwavy usability vs. discoverability plot, general search engines rank pretty high on both axes; they&#039;re the ready-at-hand tool in browser address bars. And they grok schema.org, so if we can improve our discoverability by publishing schema.org data, maybe we get a discoverability win for our users.
&lt;/p&gt;
&lt;p&gt;
But even if we don&#039;t (SEO is a black art at best, and maybe the general search engines won&#039;t find the right mix of signals that makes them decide to boost the relevancy of our resources for specific users in specific locations at specific times) we get access to that structured data across systems in an extremely reusable way. With sitemaps, we can build our own specialized search engines (Solr or ElasticSearch or Google Custom Search Engine or whatever) that represent specific use cases. Our more sophisticated users can piece together data to, for example, build dynamic lists of collections, using a common, well-documented vocabulary and tools rather than having to dip into the arcane world of library standards (Z39.50 and MARC21).
&lt;/p&gt;
&lt;p&gt;
So why not iterate our way towards the linked open data future by building on what we already have now?
As &lt;a href=&quot;http://kcoyle.blogspot.ca/2014/10/schemaorg-where-it-works.html&quot;&gt;Karen Coyle wrote&lt;/a&gt; in a much more elegant fashion, the transition looks roughly like:
&lt;/p&gt;
&lt;ul&gt;
    &lt;li&gt;Stored data -&gt; transform/template -&gt; human readable HTML page&lt;/li&gt;
    &lt;li&gt;Stored data -&gt; transform/template (tweaked) -&gt; machine &amp;amp; human readable HTML page&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;
That is, by simply tweaking the same mechanism you already use to generate a human readable HTML page from the data you have stored in a database or flat files or what have you, you can embed machine readable structured data as well.
&lt;/p&gt;
&lt;p&gt;
That is, in fact, exactly the approach I took with Evergreen, VuFind, and Koha. And they now expose structured data and generate sitemaps out of the box using the same old MARC21 data. Evergreen even exposes information about libraries (locations, contact information, hours of operation) so that you can connect its holdings to specific locations.
&lt;/p&gt;
&lt;p&gt;
And what about all of our resources outside of the catalogue? Research guides, fonds descriptions, institutional repositories, publications... I&#039;ve been lucky enough to be working with Camilla McKay and Karen Coyle on applying the same process to the Bryn Mawr Classical Review. At this stage, we&#039;re exposing basic entities (&lt;a href=&quot;http://schema.org&quot;&gt;Reviews&lt;/a&gt; and &lt;a href=&quot;http://schema.org/Person&quot;&gt;People&lt;/a&gt;) largely as literals, but we&#039;re laying the groundwork for future iterations where we link them up to external entities. And all of this is built on a Tcl + SGML infrastructure.
&lt;/p&gt;
&lt;p&gt;
So why schema.org? It has the advantage of being a de-facto generalized vocabulary that can be understood and parsed across many different domains, from car dealerships to streaming audio services to libraries, and it can be relatively simply embedded into existing HTML as long as you can modify the templating layer of your system.
&lt;/p&gt;
&lt;p&gt;
And schema.org offers much more than just static structured data; schema.org Actions are surfacing in applications like Gmail as a way of providing directly actionable links--and there&#039;s no reason we shouldn&#039;t embrace that approach to expose &quot;SearchAction&quot;, &quot;ReadAction&quot;, &quot;WatchAction&quot;, &quot;ListenAction&quot;, &quot;ViewAction&quot;--and &quot;OrderAction&quot; (Request), &quot;BorrowAction&quot; (Borrow or Renew), &quot;Place on Reserve&quot;, and other common actions as a standardized API that exists well beyond libraries (see Hydra for a developing approach to this problem).
&lt;/p&gt;
&lt;p&gt;
I want to thank Richard Wallis for inviting me to co-present with him; it was a great experience, and I really enjoy meeting and sharing with others who are putting linked data theory into practice.
&lt;/p&gt; 
    </content:encoded>

    <pubDate>Sat, 08 Nov 2014 11:41:30 -0500</pubDate>
    <guid isPermaLink="false">https://coffeecode.net:443/archives/294-guid.html</guid>
    <category>coding</category>
<category>evergreen</category>
<category>structured data</category>

</item>
<item>
    <title>DCMI 2014: schema.org holdings in open source library systems</title>
    <link>https://coffeecode.net:443/archives/293-DCMI-2014-schema.org-holdings-in-open-source-library-systems.html</link>
            <category>Coding</category>
            <category>Evergreen</category>
            <category>Libraries</category>
            <category>Structured data</category>
    
    <comments>https://coffeecode.net:443/archives/293-DCMI-2014-schema.org-holdings-in-open-source-library-systems.html#comments</comments>
    <wfw:comment>https://coffeecode.net:443/wfwcomment.php?cid=293</wfw:comment>

    <slash:comments>0</slash:comments>
    <wfw:commentRss>https://coffeecode.net:443/rss.php?version=2.0&amp;type=comments&amp;cid=293</wfw:commentRss>
    

    <author>dan@coffeecode.net (Dan Scott)</author>
    <content:encoded>
    &lt;p&gt;My slides from DCMI 2014: &lt;a
href=&quot;http://stuff.coffeecode.net/2014/dcmi_schemabibex/#/&quot;&gt;schema.org in the
wild: open source libraries++&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;
Last week I was at the &lt;a
href=&quot;http://dcevents.dublincore.org/IntConf/dc-2014&quot;&gt;Dublin Core Metadata
Initiative 2014 conference&lt;/a&gt;, where Richard Wallis, Charles MacCathie Nevile
and I were slated to present on schema.org and the work of the W3C Schema.org
Bibliographic Extension Community Group (#schemabibex). As a first-timer at
DCMI, I wasn&#039;t sure what kind of an audience to expect: there is a
peer-reviewed papers track, and a series of sessions on a truly intimidating
topic (RDF Application Profiles), but on the other hand our own topic was
fairly basic. As it turned out, there was an invigoratingly mixed set of
backgrounds present, and Eric Miller&#039;s opening keynote, which gave an oral
history of the origins of DCMI and a look towards the future challenges for the
organization, reassured me that I wasn&#039;t going to be out of my depth.
&lt;/p&gt;
&lt;p&gt;
Special kudos to Eric for his analogy of the Web to a credit card, which offers
both human-readable and machine-readable data. A nice, clean image!
&lt;/p&gt;
&lt;p&gt;
Richard, Charles and I opted to structure our 1.5 hour session as a series of
short talks followed by a long period of discussion. However, as often happens,
the excitement of speaking in front of a room that drew so many attendees that
we had to jam with more chairs led to that plan breaking down. I cut my own
materials back to illustrating how one of my primary contributions to the
#schemabibex effort--representing library holdings using schema.org&#039;s
GoodRelations-based Product/Offer model--had been implemented in free software
library systems, including Evergreen, Koha, and VuFind. I walked from a basic
bibliographic record (represented as a &lt;a
href=&quot;http://schema.org/Product&quot;&gt;Product&lt;/a&gt;), through to the associated
borrowable items (represented as &lt;a href=&quot;http://schema.org/Offer&quot;&gt;Offers&lt;/a&gt;
with a price of $0.00, call numbers as &lt;a
href=&quot;http://schema.org/sku&quot;&gt;SKUs&lt;/a&gt;, and barcodes as &lt;a
href=&quot;http://schema.org/serialNumber&quot;&gt;serialNumbers&lt;/a&gt;), that were offered by
a specific &lt;a href=&quot;http://schema.org/Library&quot;&gt;Library&lt;/a&gt; with its own set of
operating hours, address, and contact information... all published out of the
box as RDFa in modern Evergreen systems.
&lt;/p&gt;
&lt;p&gt;
I did stray a little to posit that the use case for schema.org is not and should
not be limited to &quot;search engine optimization&quot;, but that this very simple level
of structured data could fairly easily form the basis of an API. In the rather
limited discussion that we were able to hold at the end of the session (and
encroaching on break time), Charles counselled that libraries shouldn&#039;t really
bother with dumbing down their beautiful metadata simply to publish
schema.org... while I countered that the pursuit of publishing beautiful
metadata in the past has generally led librarians to publish no metadata at
all, and that schema.org was a great first step towards building a web of
cultural heritage metadata meant for machine consumption.
&lt;/p&gt;
&lt;p&gt;
I wish I could have stayed longer at DCMI, but it was Thanksgiving in Canada
and there were families to visit and feast with--not to mention children to
help take car of--so I had to depart after just a day and a half. I&#039;m
encouraged by the steps the organization is taking to renew itself, and I hope
to be able to participate again in the future.
&lt;/p&gt; 
    </content:encoded>

    <pubDate>Mon, 13 Oct 2014 21:07:13 -0400</pubDate>
    <guid isPermaLink="false">https://coffeecode.net:443/archives/293-guid.html</guid>
    <category>coding</category>
<category>evergreen</category>
<category>libraries</category>
<category>structured data</category>

</item>
<item>
    <title>My small contribution to schema.org this week</title>
    <link>https://coffeecode.net:443/archives/292-My-small-contribution-to-schema.org-this-week.html</link>
            <category>Coding</category>
            <category>Structured data</category>
    
    <comments>https://coffeecode.net:443/archives/292-My-small-contribution-to-schema.org-this-week.html#comments</comments>
    <wfw:comment>https://coffeecode.net:443/wfwcomment.php?cid=292</wfw:comment>

    <slash:comments>0</slash:comments>
    <wfw:commentRss>https://coffeecode.net:443/rss.php?version=2.0&amp;type=comments&amp;cid=292</wfw:commentRss>
    

    <author>dan@coffeecode.net (Dan Scott)</author>
    <content:encoded>
    &lt;p&gt;
&lt;a href=&quot;http://schema.org/docs/releases.html#v1.91&quot;&gt;Version 1.91&lt;/a&gt; of the http://schema.org vocabulary was released a few days ago, and I once again had a small part to play in it.
&lt;/p&gt;
&lt;p property=&quot;description&quot;&gt;
With the addition of the &lt;a href=&quot;http://schema.org/workExample&quot;&gt;workExample&lt;/a&gt; and &lt;a href=&quot;http://schema.org/exampleOfWork&quot;&gt;exampleOfWork&lt;/a&gt; properties, we  (Richard Wallis, Dan Brickley, and I) realized that examples of these CreativeWork example properties were desperately needed to help clarify their appropriate usage. I had developed one for the &lt;a href=&quot;http://blog.schema.org/2014/09/schemaorg-support-for-bibliographic_2.html&quot;&gt;blog post&lt;/a&gt; that accompanied the launch of those properties, but the question was, where should those examples live in the official schema.org docs? CreativeWork has so many children, and the properties are so broadly applicable, that it could have been added to dozens of type pages.
&lt;/p&gt;
&lt;p&gt;
It turns out that an until-now unused feature of the schema.org infrastructure is that examples &lt;em&gt;can&lt;/em&gt; live on property pages; even Dan Brickley didn&#039;t think this was working. However, a quick test in my sandbox showed that it _was_ in perfect working order, so we could locate the examples on their most relevant documentation pages... Huzzah!
&lt;/p&gt;
&lt;p&gt;
I was then able to put together a nice, juicy example showing relationships between a Tolkien novel (&lt;em&gt;The Fellowship of the Ring&lt;/em&gt;), subsequent editions of that novel published by different companies in different locations at different times, and movies based on that novel. From this librarian&#039;s perspective, it&#039;s pretty cool to be able to do this; it&#039;s a realization of a desire to express relationships that, in most library systems, are hard or impossible to accurately specify. (Should be interesting to try and get this expressed in Evergreen and Koha...)
&lt;/p&gt;
&lt;p&gt;
In an ensuing conversation on public-vocabs about the appropriateness of this approach to work relationships, I was pleased to hear Jeff Young &lt;a href=&quot;http://lists.w3.org/Archives/Public/public-vocabs/2014Sep/0045.html&quot;&gt;say&lt;/a&gt; &quot;+1 for using exampleOfWork / workExample as many times as necessary to move vaguely up or down the bibliographic abstraction layers.&quot;... To me, that&#039;s a solid endorsement of this pragmatic approach to what is inherently messy bibliographic stuff.
&lt;/p&gt;
&lt;p&gt;
Kudos to Richard for having championed these properties in the first place; sometimes we&#039;re a little slow to catch on!
&lt;/p&gt; 
    </content:encoded>

    <pubDate>Sat, 13 Sep 2014 03:27:13 -0400</pubDate>
    <guid isPermaLink="false">https://coffeecode.net:443/archives/292-guid.html</guid>
    <category>coding</category>
<category>structured data</category>

</item>

</channel>
</rss>