Posted on November 10th, 2010 1 comment
Please note: postponed due to bad weather and cancellations.
What is ePub and what has it got to do with application profiles? ePub is a standard file format for ebooks. It contains some metadata fields that include some simple mandatory elements and can include more elements as required. Any local metadata solution drawing elements from various sources is effectively an application profile. So ePub itself is a file format, but it contains some simple core metadata.
What we would like to do is investigate how useful, achievable and discoverable it is possible to make repository content available on ebook readers using the ePub format. We are bringing developers together with library and repository professionals to participate in a practical, hands-on hackday designed to find solutions to converting repository content into the ePub format and evaluating its metadata requirements in the context of ebooks.
More information is available here. If you are interested in getting repository content onto ebook readers, and in metadata, this event is for you!
Posted on September 21st, 2010 2 comments
Past and present
Up until the present, a number of application profiles have been developed by various metadata experts, with the support of the JISC, with the intention of addressing the needs of practitioners and service providers (and thus ultimately their users) across the higher education sector in the UK. The most significant of these have been aimed at particular resource types that have an impact across the sector.
- SWAP – Scholarly Works Application Profile
- IAP – Images Application Profile
- GAP – Geospatial Application Profile
- LMAP – Learning Materials Application Profile (scoping study only: also the DC Education AP)
- SDAPSS – Scientific Data Application Profile Scoping Study
- TBMAP – Time-Based Media Application Profile
Problems with this approach
However, it cannot be said that a particular type of resource type, set of resource types, or even general subject domain actually constitutes a real, identified problem space that faces large sections of the information community in the UK higher education sector today. Geospatial resources can be any type of resources that have location metadata attached (e.g. place of creation, location as the subject of the resource). Learning materials can be any type of resource that has been created or re-purposed for educational uses, which can include presentations, academic papers, purpose-made educational resources of many types, images, or indeed almost anything else that could be used in an educational context, to which metadata describing that particular use or re-use has been attached. Images might have all sorts of different types of metadata: for instance, metadata about images of herbs might need very different metadata to images of architecture. The same applies to time-based media: what is the purpose of these recordings and what are they used for? why and how will people search for them? Likewise, the type of science in question, of which there are almost innumerable categories and sub-categories, will to a large extent determine the specific metadata that will be useful for describing scientific data.
Of all of the above, only scholarly works, which might more usefully be called scholarly publications, are an entirely focussed, specific set of resource types with a common purpose. The others are loose and sometimes ill-defined collections of resources or resource types that fit into a particular conceptual category. Only in the case of scholarly publications is there an unspoken problem space: discovery and re-use in repositories and similar systems, usually but not exclusively as Open Access resources. There are other related problem spaces such as keeping accurate information about funders and projects for the purposes of auditing that is required by funding bodies and university authorities. The ability to access these resources with new technologies could be a further area of study, and is one that UKOLN is taking an active interest in. Again, the question must be “what do users want to do with these resources?”
It must not be said that the work in creating the application profiles mentioned above has been wasted. At the same time, the above application profiles constitute general purpose solutions that do not target specific problems affecting identifiable communities of practice across the sector. There is considerable work continuing in Dublin Core Metadata Initiative (DCMI) circles on how metadata modelling should best be carried out, for instance on the Dublin Core Abstract Model (DCAM) and on the overlap between application profiles and linked data, where those application profiles contain relationships that can better enable resource discovery in a linked data world.
These approaches remain useful. However, more immediate, specific problem spaces face particular university services (not all of which are necessarily repositories) in trying to describe resources so that they can be discovered, providing copyright and other licensing information so that they can be re-used, providing funding information so that work can be audited and cases can be constructed for funding new projects, and so on. Some of these resources may be textual, but others are increasingly including images (of many types and for many purposes), music, film, audio recordings, learning objects of many types, and large scale corpora of data. Any metadata solution that is tailored to a particular purpose (and, thus, which is usually de facto an application profile) needs to address specific aspects of the Web services that practitioners and other service providers are seeking to develop for their users, not simply provide general catch-all metadata of relatively generic use.
Key to all this is consultation with those communities: first, to scope the most significant two or three problem spaces that face the largest number of resource providers in serving their users; second, to get those practitioners together with developers to draw up practical, workable recommendations and perhaps demonstrations; third, to provide tangible evidence to the developers of existing software platforms, and to engage with them to help solve such problems in practice. To do this, it is necessary to engage practitioners and deverlopers in practical, hands-on activities that can bring the discussion forward and provide tangible solutions.
Posted on May 19th, 2010 2 comments
Drupal 7 is likely to be released soon, and will include native support for RDFa. The RDF module for Drupal 6 already allows this functionality. Why is this important? Because it makes relationships between resources much easier to describe through Drupal’s user-friendly interface and, in the process, would allow documents to be available as linked data.
In Drupal terminology, a “node” is effectively a metadata record, and various Drupal modules enable the easy customisation of metadata. In effect, you could build a repository on the basis of Drupal, by-passing the need for platform-specific knowledge tied to open source software that has increasingly moved towards the “enterprise solution” space, along with all of the technical tie-in that it usually entails. For the service provider, it is not dissimilar to the tie-in experienced with commercial software, especially in the case of information librarians or other professionals who are not developers, or even developers are not part of that particular open source development team.
Application Profiles are essentially structured metadata comprising elements and (usually) relationships, and are therefore inherently linked data solutions. They vary in complexity according to their particular functional requirements: for instance, in the world of scholarly publications, there is a spectrum between the straightforward, unstructured way that DSpace implements Dublin Core (which should perhaps be called the DSpace Application Profile), the simplified FRBR structure of the Scholarly Works Application Profile (SWAP) and the complex entity-relationship model of CERIF, the standard developed for Current Research Information Systems (CRISs). This latter is a de facto application profile, even if it is not normally referred to as such.
Why should Drupal be any better than the repository platforms that already exist? In many ways, it depends on what you need to do with it, and on the resources at your disposal. But the advantage is that Drupal is a flexible Content Management Framework that is designed to be leveraged for any sort of content, and for new modules to be designed easily for new purposes. After all, what does a repository actually do that other websites cannot? They put metadata records and bitstreams (the actual documents or files) on the Web, and add a few additional services such as OAI-PMH, SWORD and statistics. But repositories are only a particular specialised subset of content management systems. Drupal is accessible to any PHP developer without any initial requirement of particular specialist platform knowledge, which is relatively easy to obtain. The community is large and support is quite easily available, as are modules that can be adapted for local purposes. It is designed to be easy to customise and theme.
Sarah Currier recently talked about the idea of a “fauxpository”. If I remember correctly, she pointed out that it could even be based on WordPress. This is clearly a workable idea, although hardly suitable for production use as a university service. I would maintain that Drupal could easily be suitable for such a use with relatively little work, and could make use of and adapt application profiles in a way that the major open source repository platforms have been slow to do, and are still only just beginning to enable as something of an afterthought. UKOLN are investigating how Drupal can be used to make it possible to make use of the JISC’s Dublin Core Application Profiles (DCAPs), and using Drupal is intended to show how it can work independently of tie-in to any specific platforms.
Posted on March 23rd, 2010 2 comments
EPrints 3.2.0 was released on 10th March 2010. It has some remarkable new features relating to linked data and, consequently, to Dublin Core Application Profiles based on multiple entity domain models such as SWAP, IAP and TBMAP (the GAP does not have a domain model). Here are the key points:
Linked Data Support
- Ability to establish arbitrary relations between objects or provide additional metadata in triple form.
Semantic Web / Linked Data (RDF)
We have made a (difficult) decision to move these features to 3.2.1 (due out soon after 3.2.0) because testing showed it caused a significant slow down.
We’re rewriting it to do the same thing but with much less overhead!
However, as may be seen on the EPrints wiki, the latter section read as follows until 11th March 2010:
Semantic Web Support
- RDF+XML Format
- N3 Format
- URIs for all objects, including non dataobjs. [sic] eg. Authors, Events, Locations.
- BIBO Ontology
- URIs now use content negotiation to decide which export plugin to redirect to, based on mime-types supplied by plugins and the “accept” header.
- Relations between eprints and documents
If this is understood on face value, it appears that there has been significant progress in enabling features that would allow the full implementation of the JISC’s DCAPs based on the simplified FRBR model, although we must wait for some important details until the promised version 3.2.1, which is to be released “soon after 3.2.0″ according to the statement above. Although objects may be described with “arbitrary relations” and “additional metadata” (additional to what?) can be described in triple form, there are not yet URIs for all entities, such as Authors and so on. Presumably, the support for BIBO would be more demanding that the support required for the cut-down version of FRBR as seen, for example, in SWAP.
This is all very promising, especially in the light of the same functionality being promised in DSpace 2.0, which were not yet implemented in the recent release of DSpace 1.6.0. However, all of this must come with the caveat that, until this is tried out in practice, it is not certain which levels of implementation are possible: clearly, the actual metadata fields can easily be adopted by any repository, but what about the relationships between entities, and the relationships with other complex objects? How exactly will these be implemented in practice? For the purposes of linked data, we also have to wait until EPrints 3.2.1 for metadata in the RDF+XML format.
To this end, although UKOLN cannot offer a publicly accessible test repository with user access, we hope wherever possible to implement and test these pieces of repository software for their usability with SWAP, IAP, TBMAP, GAP and DC-Ed in the first instance, since the majority of repositories in the UK HE sector use these platforms. Of course, we would also like to do the same with Fedora at some point in the future. However, if you have evidence of any such implementations, even for test purposes, and if you are happy for us to evaluate these, we would be very happy to hear from you.
Posted on August 6th, 2009 No comments
We’ve recently started trying out various methodologies for testing whether the different bits of application profiles work for the people trying to use those resources. The main thing to remember is that the approach must not be too technical: anybody ought to be able to understand what the metadata terms and the relationships between digital objects on the web are trying to achieve. This is hands-on metadata for real people!
So we’ve been to various meetings lately. The first one is perhaps the least relevant from most people’s point of view, the Metadata Registries Meeting at the Novotel Centre, York, 23-24 July 2009. We were seeking feedback and discussion of our methodology, as well as talking about a few technical possibilities, which was a useful thing to do. You may ask, what are registries? Well they aren’t the subject of this blog, but in brief they are places that allow people to share their metadata schemas, application profiles and so on, as well as to find tools to help them develop, build and maintain them over time. Remember that application profiles are living structures that should change as the metadata needs of your users in dealing with the resources that you provide change over time. We (UKOLN) operate a registry called IEMSR.
So what did we do for the people?! Well, first of all we went to the Institutional Web Managers Workshop 2009 (IWMW), 28-30 July 2009 because we felt that they are people who are focussed on making services work for users. It may have been an advantage in some ways that they weren’t by and large repository-related people and could look on things from a fresh perspective. It’s always good to get a range of different approaches: after all, won’t the users come to a repository, VRE, VLE or other service with a whole range of points of view and things they want to do? You can see a slightly ad hoc and only mildly embarrassing interview with me, Talat Chaudhri of UKOLN, explaining in about 20 seconds of profound unreadiness over coffee what it is that application profiles (should) do. (Why on earth did the kind editor choose that particular first frame to stop the video?!) Not a bad attempt, given the lack of coffee, I hope you may agree.
What we did was to get people to think about resources, and reasons why users would want to be looking at them. We played with post-it notes (also called stick-it notes elsewhere?) that had metadata terms written on them, and arranged them in logical groups that would help a person who was trying to perform focussed searches for resources. Any resource type will do: for instance, we tested it out before the session on “beach life: what you might find on a beach and what you might want to know about it”. It doesn’t even have to be sensible! In real life application profiles, however, you obviously need to think of the whole range of things that people will want to do with your resource. The best way: ask them! Don’t engineer things that people won’t want to use. The extra complexity creates the very real danger of making your structures difficult to search, which will put off the very users that are supposed to be using the service.
This method is called card sorting, and is quite well known. It does have some limitations, but we have already shown its usefulness in focussing attention on what users need to do. One limitation, for instance, is that it’s rather hard not to prejudice the process from the beginning. If you ask the participants to think of the scenarios that users might search for resources first, then participants will come with pre-ordained ideas that will tend to undermine the fresh analysis of user requirements that we are looking for. On the other hand, if you don’t let them know until they have already thought of the terms that they need to describe the resource, on the first try they will tend to organise them in ways that don’t work with the scenarios. Let us remember, though, that this is just the first iteration of a development cycle for metadata solutions. You need to take every new version back to the users and check that it does what they need it to.
A second limitation is that paper prototyping can’t produce the complex cross-links that you’d find in a real database. A third one is that it doesn’t begin to touch the importance of interface design to usability testing of metadata terms and structures. You may (or may not) need a complex data structure. However, your user should only see what they need to see in order to accomplish what they want. Anything else will actually hinder their use of the service, be it a web page or a repository deposit interface. That complexity can be generated behind the scenes by software, so that users are asked understandable, intuitive and above all useful questions that facilitate their end user experience. We’re also working on these areas.
We then went to the Repositories Fringe 2009 in Edinburgh, 30-31 July 2009. (You will see from the above dates that this was a bit of a marathon!) The session was broadcast live on the Web, and I hope that the recording will be made available before long. I will add a link here if/when that happens. Having learnt a little from the above session, we did more of the same. We learnt a lot about how to get user requirements, and even more about how not to do it!
We were asked if we were running a focus group. If people want application profiles like SWAP, IAP, GAP, TBMAP and so on to be implemented, we will certainly have to consult focus groups, but we will tell people when that is what we’re doing. First, however, we are trying to raise discussion about how we can analyse user requirements on an ongoing basis and transmit that hard evidence to developers, so that will have a reason to go to the trouble of incorporating it into their software releases. At present, we can’t show them sufficient evidence that these APs do what they are intended for, which is why repository software developers in particular have been understandably agnostic about APs. But the other thing that is crucial is to engage service providers and users. Why do they want to come? If they don’t get something out of the event that will improve their service or their knowledge, preferably both, they won’t come. This was as much an outreach and training event as a focus group.
We’re hoping that this is a good start towards an iterative, user-driven method for analysing existing APs for various purposes, as well as for designing new APs from scratch. We’re confident that it’s going well at the moment and that we are beginning to get answers. But the task of making your metadata fit the service that you provide is ongoing, because services also change over time. It’s best not to be too prescriptive, as different institutional or web services take different approaches to achieving the same things. We are aiming at a flexible, iterative, toolkit approach that works for as many people as possible, and offers a range of tools to implement relevant parts of an overall solution that work for the services and users concerned.
Lastly, the fact that we are reviewing APs should not be taken as a criticism of the ones that we have, even if deficiencies are found that need to be rectified, or new approaches taken. The work that was done in creating them has laid the groundwork for this new activity, which is aimed precisely at making the results of that work more useful in the community of web services that they were intended for. Change should be welcomed because needs and requirements change, along with our understanding of how best to analyse them.application profiles, user testing application profiles, APs, card sorting, DCAPs, GAP, IAP, IEMSR, interface design, iterative development, IWMW, jiscob, metadata, Metadata Registries Meeting, metadata terms, paper prototyping, repositories, Repositories Fringe, SWAP, TBMAP, usability testing, user requirements, user testing, VLEs, VREs
Posted on May 8th, 2009 1 comment
Workshop: Application Profiles in Practice, 6 May 2009
This was an event in two parts: firstly, an introduction to the user testing methodology being developed by the AP Support project in collaboration with the IEMSR and the IE Demonstrator project; secondly, an iteration of the paper prototyping element of the user testing. On this occasion the audience was comprised largely of experts rather than an especially representative group of typical users – quite understandably, given the nature of the meeting. (While it is very helpful to engage repository managers in user testing, it is more difficult to involve entirely non-specialist users, so there is a need for further work in facilitating this.) The session proved to be a success in raising considerable interest in current developments in application profiles.
It was always the intention to use this particular event as a platform for consulting colleagues in the repositories community about the usefulness of the approach. In this respect, the workshop was highly successful: attendees responded positively to the intention of engaging users in order to analyse and address the strengths and weaknesses of the various application profiles, raising some insightful questions and contributing to an animated debate. Rachel Bruce of JISC commended the workshop in her speech closing the Programme Meeting on the following day.
“Working with the Repositories Community: WRAP Project” (Jenny Delasalle, Warwick University), 6 May 2009
Jenny Delasalle referred to the difficulties faced in pioneering an implementation of SWAP in an institutional repository based on EPrints 3.0. Unlike in its successor EPrints 3.1, versioning was unsupported at the time, which to a great extent hampered the SWAP effort in WRAP at Warwick. She considered that in its present form, SWAP represents too complex a metadata model for adoption by the typical IR. But since there is not necessarily a need to employ all of the SWAP metadata terms (any more than one would necessarily need to employ all of the terms in DC Simple or Qualified DC), it must be presumed that the FRBR structure and the lack of automated means to populate fields with structural metadata represent a significant part of the problem. It would be useful to get a clarification from Jenny on this.
That the feasibility of complex metadata schemas could be radically improved by the use of text mining to autopopulate metadata fields, thus requiring far less input and/or correction from the user, was raised later in the Forum in the discussion “How can text mining support repository tasks?”, convened by James Farnhill of JISC and led principally by Brian Rea of NaCTeM, University of Manchester. This would be of obvious and immediate relevance to the liklihood of SWAP being more widely implemented, whether in its present form or following the recommendations from the user testing effort.
Repositories Roadmap Session (Rachel Heery, external consultant for JISC), 7 May 2009
Rachel Heery gave a summary of her Digital Repositories Roadmap Review, revised from the original version by herself and Andy Powell in 2006. Recommendation 11 referred to SWAP specifically, proposing a cut-down version without the FRBR entity-relationship model and a re-analysis of the sort undertaken in the current user testing programme; Recommendation 12 made an interesting reference to OAI-ORE in the context of SWAP.
Recommendation 11: Explore deployment of a cut down version of SWAP, possibly at the copy level, retaining the cataloguing rules to ensure a consistent approach to linking to full text. Evaluate whether use of SWAP is consistent with a Web architecture approach to repositories.
Recommendation 12: Explore use of OAI-ORE to enable applications to handle complex objects. Demonstrate how OAI-ORE facilitates the re-use of research outputs and research data. Clarify different roles of OAI-ORE and SWAP.
There was considerable discussion of SWAP on Twitter among colleagues at Eduserv, UKOLN and elsewhere on both days of the meeting, focussing on both the structure and implementation of SWAP as it was originally intended, and in response to Rachel Heery’s recommendations. The need to solve the lack of implementation of the Dublin Core Application Profiles appears to have regained significant impetus from the interest in the series of user testing events planned by UKOLN. In particular, new impetus has been given to the SWAP implementation effort, in which expectations had previously subsided. Given Rachel Heery’s review, it is clear that SWAP may need to be considered once more as an ongoing project rather than a past product that failed to gain support, and one that may need substantial revision in future iterations. It is important to keep an open mind about the nature of those revisions, which should be conditioned by the results of the ongoing user testing effort.application profiles, user testing Andy Powell, application profiles, Birmingham, Brian Rea, DC, DC Simple, DCAPs, Digital Repositories Roadmap Review, Dublin Core, Eduserv, EPrints, EPrints 3.0, Eprints 3.1, FRBR, IE Demonstrator, IEMSR, James Farnhill, Jenny Delasalle, JISC, jiscob, NacTeM, OAI-ORE, paper prototyping, Qualified DC, Rachel Bruce, Rachel Heery, Repositories and Preservation Programme Meeting, repository software, SWAP, text mining, Twitter, UKOLN, user testing, WRAP
Posted on April 29th, 2009 3 comments
On Monday 27 July, the first trial of methods for user testing for SWAP were conducted at UKOLN. This was very much an internal “dry run”, the success of which leaves us in a strong position to take every opportunity to repeat the exercise more widely within the repositories community.
In collaboration with the IEMSR and IE Demonstrator projects, which also have an interest in developing and implementing application profiles in repositories, we are very interested in developing methodologies for the evaluation of the Dublin Core Application Profiles (DCAPs) funded by JISC. Of these, SWAP has the best developed online presence and content in institutional repositories, as well as a strong and developed user community focussed on developing Open Access content on the Web.
Our current work is therefore focussed on SWAP in the first instance, but we naturally intend to develop the process of practical user testing for the other DCAPs. We are of course aware that the needs of different resource types and repository communities will differ very widely. This is the reason that we are interested in practical user testing within those communities, to ensure that theoretical approaches to constructing application profiles actually fill the needs and requirements than underpin the development of such content in repositories and other related services on the Web.
In many ways, it is fair to say that SWAP is the “lowest hanging fruit” for this endeavour, but the impact of getting application profiles right for such a large and growing proportion of repository content should not be underestimated. Being largely textual resources, scholarly works are likely to be an area where significant lessons can be learnt for other resource types with more specific constraints and requirements. It is intended that we conduct user testing of several other DCAPs during the summer of 2009, if possible, following initial work on SWAP.
It is perhaps worth remembering that developments in repository technology have come a long way since SWAP was first developed, the example of which the other DCAPs have tended to follow, especially in the matter of using the FRBR structure. It is by no means certain that this is the only way, or the best way, to create relationships in so-called complex objects, which is to say sets of resources that relate to each other as versions. In particular, OAI-ORE is an exciting development that may provide an alternative, although its relevance for this purpose needs to be carefully evaluated and compared to existing approaches and technologies. It will not do to simply adopt the newest, coolest approach without a careful analysis of how the needs of users relate to the functionality that is presently available. If these do not correlate well within the software contexts currently in use in the community, the application profiles will fail accordingly.
It has become obvious that implementation of the DCAPs has been slow. In the case of SWAP, which has been around the longest, that lag has become a profound apathy towards efforts to implement the application profile widely in repositories. It is not even clear how best this should be done, as neither methods nor benefits have been convincingly demonstrated. It would be a great shame if the investment of expertise in improving the metadata vocabulary were wasted because the structure has not been successfully integrated into repositories. This mismatch must be understood and resolved if the situation is to be turned round and the expected gains of SWAP and the other DCAPs are to be realised.
There are a variety of technological approaches to application profiles that will need careful study, once the user testing brings a better understanding of how users need to relate particular types of resources together in repositories. These may include the Description Set Profile approach, traditional XML with XML Schema and OAI-ORE. But more importantly, it must be shown beyond doubt how the DCAPs fit the applications that users need them for, and which changes may be required, before software developers will have the motivation to address those needs by implementing the DCAPs in the major repository platforms.
Posted on March 12th, 2009 No comments
At present we are working on practical evaluation and user testing of SWAP, the first and most fully developed of the DCAPs funded by JISC. The aim of this work is to enable us to report on how well SWAP fits the real needs of repository managers in their day-to-day work. On this basis, we intend to organise further practical events, both for repository managers and for repository platform developers. The hope is to provide an impetus for SWAP to be supported by the major repository platforms in an appropriate form. This feedback should provide a sound basis to continue the development cycle of SWAP and improve it over time.
We are not forgetting, however, the needs of the other resource types for which DCAPs have been developed, and the process is intended to be an iterative one, learning from the experience of SWAP in the first instance and inviting domain specialists for each resource type to help adapt the process for the needs of the section of the repository community involved with that particular resource. We invite those user communities and specialists to engage with us in the same process as outlined for SWAP above.
Web resources have living, changing needs and user communities, so we believe that their application profiles should reflect this. Obviously, there is a balance to be struck between developing and maintaining useful standards that can be relied upon, and meeting these changing requirements. The only way to do this properly is to use inclusive methods, consulting domain specialists and real users as much as possible.