Recently I’ve been “having a think” on issues ranging from rights expression for datasets to realising the value of linked data, but frankly I’ve felt that something is missing ; even with scientific and government linked datasets going online, a voice inside me wonders whether the stakes are still too low to really shake things up. What kind of data haven’t we been hearing about – what kind of data, if it were published according to linked data principles, would surely lead to a wave of outrageously cool applications, demonstrate the inherent value of the linked data approach, and perhaps even surface some interesting monetisation models? One area that immediately comes to mind is multimedia semantic metadata, especially for video and audio content.
Several recent venues have focused on the general topic of publishing and using semantic multimedia metadata, including the Oct-Dec 2009 IEEE Multimedia Magazine special issue on Multimedia Metadata and Semantic Management, and SAMT2009: The 4th International Conference on Semantic and Digital Media Technologies (3-4 Dec 2009; Graz, Austria). Both of these efforts are populated by members of the Multimedia Metadata Community, an outgrowth of the MPEG-7 and MPEG-21 communities that “brings together experts from research and industry in the area of multimedia meta data interoperability for collaborative working environments.” Finally, since 2008 the W3C has been host to Video in the Web activity; its Media Annotations Working Group is developing an ontology and API to facilitate cross-community sharing and use of multimedia metadata in the Web.
IEEE Multimedia (Oct-Dec 2009): This special issue features six research articles focused on different aspects of the semantic management of multimedia and multimedia metadata ranging from retrieval and processing to consumption and presentation. The six articles have varying degrees of relevance to today’s linked data environment, from obvious to unclear.
- “Managing and Querying Distributed, Multimedia Metadata.” (abstract) This article advocates the use of a centralized metadata résumé – a condensed, automatically-constructed version of the larger metadata set – for locating content on remote servers. The authors demonstrate the advantages of their approach using conventional semweb technologies to represent and query semantic metadata.
- “Semantic MPEG Query Format Validation and Processing.” (abstract) The authors present their semantic validation of MPEG Query Format (MPQF) queries and their implementation of a practical MPQF query engine over an Oracle RDBMS. The article introduces methods for evaluating MPQF semantic-validation rules not expressed by syntactic means within the XML schema. The authors highlight their prototype implementation of an MPQF-capable processing engine using several query types on a set of MPEG-7 based image annotations.
- “Diversifying Image Retrieval with Affinity-Propagation Clustering on Visual Manifolds.” (abstract) The authors describe a post-processing subsystem for retrieval systems that improves retrieval result diversity. Image retrieval systems typically focus on determining the similarity between the retrieval and sample images, where the relevance of the retrieval results is considered but diversity, represented by a variety of subtopics, is neglected. This article presents a method for removing duplicate images from a “top 20” list and replacing them with images representing new subtopics.
- “A Media Value Chain Ontology for MPEG-21.” (abstract) The authors have created a semantic representation of intellectual property derived from MPEG-21 Part 19. Their model defines the minimal set of types of intellectual property, the roles of users interacting with them, and the relevant actions regarding intellectual property law. The article is a helpful guide to the standardization efforts, with its many examples and useful insight into the multimedia value chain.
- “Using Social Networking and Collections to Enable Video Semantics Acquisition.” (abstract) The authors consider media production, acquisition, and metadata gathering, the first elements of the multimedia value chain. Methods from video annotation and social networking are brought together to solve problems associated with gathering metadata that describes user interaction, usage, and opinions of video content. Individual user-interaction metadata is aggregated to provide semantic metadata for a given video. Coolness alert: The authors have successfully implemented their model in Tag!T, a Flex-based Facebook application (not available at the time of this writing).
- “A Web-Based Music Lecture Database Framework.” (abstract) This article describes semantic audio authoring and presentation for Web-published music lectures. The authors propose a dynamic programming-based algorithm for MIDI-to-Wave alignment to explore the temporal relations between MIDI and the corresponding performance recording. The synchronized MIDI and Wave can be attached to many kinds of teaching materials.
SAMT’09: Nearly 15 years ago I had the good fortune to present the early stages of my rights metadata research at EDMEDIA’95 in Graz, and visiting the conference web site this weekend – especially seeing the real-time image of the historic “Urhturm” on the hill high about the city – brought back a flood of fond memories! The topics of the three tutorials offered at SAMT’09 demonstrate that current research has definitely taken a turn toward getting multimedia multimedia into the Web. (Unfortunately, only slides from the first are currently available):
- “Web of Data in the Context of Multimedia (WoDMM).” How multimedia content can be integrated into the Web of Data and how users and developers can consume and benefit from linked data. (slides)
- “MPEG Metadata for Context-Aware Multimedia Applications (MPEG).” Overview of MPEG metadata formats that enable the development and deployment of content- and context-aware multimedia applications.
- “A Semantic Multimedia Web: Create, Annotate, Present and Share your Media (SemMMW).” How multimedia metadata can be represented and attached to the content it describes within the context of established media workflow practices, and how users can benefit from a Web of Data containing more formalized knowledge.
Much more information can be found in the proceedings from the 20th International Workshop of the Multimedia Metadata Community on Semantic Multimedia Database Technologies (SeMuDaTe’09)
Metadata Standards for the Web of Data: Finally, research such as that describe above has led to progress on the standards front. As the IEEE Multimedia guest editors note in their foreword, since 2008 there as been quiet but steady progress within the W3C’s Video in the Web activity, which was chartered to make video a first class citizen of the Web by creating an architectural foundation that by taking full advantage of the Web’s underlying principles will enable people to create, navigate, search, link and distribute video… Of its three working groups, the editors highlight the Media Annotations Working Group as being motivated by progress in RDF and topic maps and appears most aligned with emerging linked data activities.
In their forward, the IEEE Multimedia editors provide a very nice summary of the core problem with multimedia metadata and thus the motivation for the W3C efforts:
Most of the standards are tailored to specific application domains. Examples include European Broadcasting Union P/Meta 2.0 for broadcasting; TV-Anytime and SMPTE Metadata Dictionary for TV; and MPEG-21 for the delivery chain of multimedia and technical aspects (such as EXIF). These standards exhibit a different semantic level of detail in their descriptions (from simple keywords to regulated taxonomies and ontologies). Only some of the standards are general purpose, for instance MPEG-7…
Coolness is on the Horizon: This rather lengthy posting has providedd merely a sampling of works-in-progress that promise not only to put multimedia metadata on the Web but more importantly to establish such metadata as a useful and valuable part of the Web. Combined with such visionary efforts as the revamped, linked data-driven BBC web site, I’m increasingly confident that a generation of outrageously cool linked data applications are indeed around the corner, fueled by datasets that inject video and audio metadata into the global semantic mix. Bring it on!
Very good summary, thank you John! FWIW: seems you have missed our LDOW09 paper ‘How to Apply Linked Data Principles to Multimedia Fragments’ [1], one of the components needed to fully apply LD to multimedia.
Further, in late 2008 I gave an invited talk at DERI ‘Hypermedia is dead – long live linked multimedia!’ [2], which covers some of your points and partly goes beyond; you might find it useful as well.
KUTGW!
Cheers,
Michael
[1] http://sw-app.org/pub/ldow09-im.pdf
[2] http://sw-app.org/presentations/DERI-2008-09/future-o-multimedia-semantics-DERI-2008-09.html
By: woddiscovery on January 20, 2010
at 6:43 am