background preloader

Linkeddata

Facebook Twitter

Logd.tw.rpi.

Rww

Openapi. Linked Data at the Guardian | Open Platform. The semantic web is given a rough raking by the syntactic web, and it is not impossible to see why when you first get taken down the SPARQL/RDF/Ontology rabbit hole. It is not great fun learning to develop with the semantic web today. (As an aside, using a semi-SQL model as a primary metaphor in SPARQL did not help me personally.

But then, SQL has always seemed like an assembly language designed by Prolog programmers) But the capability to use semantic data to accurately join data is fantastically powerful. More seriously, you decide your music application could benefit from a bit of descriptive text and some mashed up functionality. But in general, ours is a cruel universe.

In the MusicBrainz/Wikipedia case, there is a deeper semantic option. There are two things happening here, two sides of the semantic question. What we did... We took the Last.fm path, and extended the Guardian's Content API to include non-Guardian identifiers. How we did it... (Tip: The MusicBrainz API is XML only. Linked Data Entity Extraction with Zemanta and OpenCalais - benjamin nowack's blog. A comparison of the NER APIs by Zemanta and OpenCalais. I had another look at the Named Entity Extraction APIs by Zemanta and OpenCalais for some product launch demos. My first test from last year concentrated more on the Zemanta API. This time I had a closer look at both services, trying to identify the "better one" for "BlogDB", a semi-automatic blog semantifier.

My main need is a service that receives a cleaned-up plain text version of a blog post and returns normalized tags and reusable entity identifiers. So, the findings in this post are rather technical and just related to the BlogDB requirements. I ignored features which could well be essential for others, such as Zemanta's "related articles and photos" feature, or OpenCalais' entity relations ("X hired Y" etc.). Terms and restrictions of the free API The API terms are pretty similar (the wording is actually almost identical). Calling the API API result processing Extracting normalized tags Extracting entities Summary. Linked Data FAQ. Linked data is the first practical expression of the semantic Web, useful and doable today, and applicable to all forms of data. Sources such as the four principles of linked data in Tim Berners-Lee's Design Issues: Linked Data and the introductory statements on the linked data Wikipedia entry approximate -- but do not completely express -- an accepted or formal or official definition of linked data per se.

Building from these sources and attempting to be more precise, here is the definition of linked data used internally by Structured Dynamics: Linked data is a set of best practices for publishing and deploying instance and class data using the RDF data model, and uses uniform resource identifiers (URIs) to name the data objects. The approach exposes the data for access via the HTTP protocol, while emphasizing data interconnections, interrelationships and context useful to both humans and machine agents. All references to linked data below embrace this definition. 1. 2. No. 3. 4. 5. 6. Share photos on Twitter. Pezholio » Blog Archive » A beginner’s guide to SPARQLing linked data (Part 1) Regular readers of this blog will have seen that, over the past 12 months or so, I’ve been banging on about linked data and SPARQL quite a bit, posting up example queries and the like, but with not much explanation about why I’m doing what I’m doing.

Thanks to the good folks at Talis and their offer of a free linked data store for spending data, I’ve also got a nice little store of my own, and I thought it was high time I went back to basics and passed on what I’ve learnt to other people. First, a bit of background The web, as most of us know it, is a network of documents, with each document having an individual URI to show its location on the web. With linked data, the web becomes an network of things with each thing having an individual URI to tell us more about it. When talking about linked data, we talk about URIs, rather than URLs, as URLs refer the the location of a file, whereas URIs are identifiers for things, which could return any format.

SELECT DISTINCT ? SPARQL queries themselves. The Semantic Link is open. At the end of last month, I wrote about the new Semantic Link podcast that I’m involved with for SemanticWeb.com. We recorded the first episode earlier this week and it’s now online, along with an introduction to the series from myself. Please do have a listen, and let me know what you think whilst I work out how to automate adding these to my sidebar on this site… I look forward to lining up some great conversations – and guests – in the year ahead. We will record and publish regularly each month. If there are particular topics you’d like to see covered, particular guests you’d like us to secure, or other comments you’d like to make, please do let me know… Related A little bit of Cray, over on SemanticWeb.com Image via Wikipedia The second of my monthly columns just went up on semanticweb.com.

In "Cloud computing" Semantic Web Gets Closer to Relational Data. Web 2.0 is so yesterday. Perhaps, that’s an overstatement, but Web 3.0, the semantic web, is rapidly moving beyond the academic laboratory to real-world utilization. Now the W3C RDB2RDF Working Group has published the First Public Working Draft of "A Direct Mapping of Relational Data to RDF.” This work is a major step towards making the astounding amount of data stored in relational databases more easily available to the semantic web. Semantic What? Web 2.0 focused on establishing interactive human-to-human communities on the internet with rich content.

Web 3.0 evolves beyond this and strives to make content consumable for people and machines. Semantic technology uses RDF, Resource Description Framework, a XML-based language to mark up content as structured data that computers can interpret. For example, an article about me might indicate that Josette Rigsby writes for CMSWire. SUBJECT: Josette Rigsby PREDICATE: employed at OBJECT: Why Should You Care? Useful article? Opening Linked Data – webr3.org. Linked Open Data cloud takes shape - 25 Nov 2010. Linked Open Data Around-the-Clock (LATC) is a new project that aims to make it easier for government agencies and web developers across the globe to gain access to information from other organisations. The project is being co-ordinated by the Digital Enterprise Research Institute at the National University of Ireland. The Linked Open Data (LOD) cloud has doubled in size every 10 months since 2007, and is a collection of more than 200 datasets that offer more than 25 billion interlinked facts available from various domains.

Some of these include governmental, media, user-generated and geographical data. The LATC project hopes to make it easier to connect to the LOD cloud by offering support to web developers who want to build applications with linked data, and to data owners, such as government agencies. "As fantastic as the LOD cloud is, there are still problems," said Dr Michael Hausenblas, LATC project co-ordinator. Linked Data | Semantic Web. Linked Data Patterns. Virtuoso Linked Data Deployment In 3 Simple Steps. Well, someone had to … — Linked Data Matters.

ISKO–UK Linked Data Conference, 26/10/10, WRAP repository blog. Writing about web page **Finally getting round to making this live after having to put off the editing for OAW and the start of term! ** This event, hosted by UCL, was one that I had been looking forward to for some time. Whether or not linked data is the 'next big thing' in web technology, and one that has to potential to solve a number of thorny problems for the administrators and maintainers of web resources in the face of increasingly complex demands, is a question that only time will answer.

However as it stands at present linked data has enormous potential as a service and as a tool and I wanted to find out more before I started getting any awkward questions from stakeholders! The sessions on the day were a nice mix of technical and non-technical and my biggest fear of being lost before the end of the keynote was mercifully misplaced. Richard Wallis for Talis Information took us on a ‘Linked Data Journey’ next. Veille Oct. II « What is Data Quality? There are several different opinions and definitions on what data quality is supposed to be. Most of the time, we adapt the “fitness for use by data consumers” definition as defined by Richard Wang and Diane Strong [1] who investigated the subjective quality perception of data quality by data consumers. Although this research work has been a milestone in data quality research, in my opinion data quality is not necessarily always subjective.

Imagine valid combinations of cities and countries or accurate population values. Who defines these quality measures? Surely not an individual. Real-world phenomena (e.g. city/country combinations)Organizational policies (e.g. all TV’s in my data must have a screen size)Legal regulations (e.g. all groceries must have an expiration date)IT-needs (e.g. So reflecting the sources of data quality definitions, I would define data quality as the degree to which data fits to the composed requirements for the task at hand.

. [1] Wang, R. So what can I do with the new Ordnance Survey Linked Data? « John’s Weblog. In a previous post I wrote up some of the features of the new Ordnance Survey Linked Data. In this blog post I want to run through a concrete example of the sort of thing you can build using this linked data. A while ago Talis built their BIS Explorer. The aim of this application was to allow users to “identify centres of excellence at the click of a button” and more can be read about the application here. This data mash-up took different data sources about funded research projects and joined them together using linked data.

In the original application you could, for example, look at funded research projects by European Region in Great Britain. This can be seen here. At the time this demo was created Ordnance Survey was yet to publish its postcode data as linked data, but if they had it would have been very easy to get a more fine grained view of research projects down at the county and district level. The basic data model of the original BIS data was fairly straight forward. Like this: