JISC Beginner's Guide to Digital Preservation

…creating a pragmatic guide to digital preservation for those working on JISC projects

Archive for the 'dcc' Category

Registration now open for the International Digital Curation Conference 2013

Posted by Marieke Guy on 19th September 2012

Registration is now open for the 8th International Digital Curation Conference 2013. The conference has the theme ‘Infrastructure,Intelligence,Innovation:driving the Data Science agenda’ and takes place from 14-16 January 2013 at the Mövenpick Hotel, Amsterdam, Netherlands

The programme will open on Monday 14 January 2013 with the Pre-Conference Drinks reception at the NEMO Science Center in Amsterdam. The welcome address will be given by Konstas Glinos who leads the Géant & e-Infrastructures Unit of the Directorate General for Information Society and Media at the European Commission.

The main conference will start on Tuesday 15 January. Speakers will be drawn from a range of disciplines, institutions and organisations and will include:-

  • Hans Pfeiffenberger, Alfred Wegener Institute
  • Anthony Beitz, Monash eResearch Centre
  • Patricia Cruse, University of California Curation Center
  • Kaitlin Thaney, Digital Science
  • Clifford Lynch, Coalition for Networked Information
  • Herbert Van de Sompel, Los Alamos National Laboratory
  • Chris Greer, National Institute of Standards and Technology

There will be an exhibition of posters and demonstrations throughout the conference, a full programme of research and practice papers and an interactive symposium which will pose the question “What is a Data Scientist?”

This is the first time that IDCC has been held in mainland Europe and the DCC are delighted to have support from two major institutions in the Netherlands – Data Archiving and Networked Services (DANS) and Delft University of Technology (TU Delft).

IDCC13 is organised by the Digital Curation Centre UK, in partnership with the Coalition for Networked Information (CNI) and with sponsorship from Microsoft Research.

Posted in dcc, Events | Comments Off

International Digital Curation Conference 2013: Call for Papers

Posted by Marieke Guy on 31st July 2012

The 8th International Digital Curation Conference 2013 (IDCC13) with the theme ‘Infrastructure, Intelligence, Innovation: driving the Data Science agenda’ will take place on 14-16 January 2013 at the Mövenpick Hotel, Amsterdam, Netherlands.

The call for papers is currently open and the IDCC11 Programme Committee that reflect the conference theme. The theme recognises that in recent years there has been an explosion in the amount of data available, whether from tweets to blogs, data from sensors through to “citizen science”, government data, health and genome data and social survey data. Technology allows us to treat as ‘data’ content which would not once have merited the term – recordings of speech or song, video of dance or theatre or animal behaviour – and to treat as quantitative what once could only be qualitative. There are challenges in finding data and making it findable, in the ability to use it effectively, to take and understand data, to process, to analyse and extract value from data , to visualize data and then to communicate the stories behind it.

This process is now being termed data science. It is being used across sectors to describe a wide range of data activities in the commercial, government and academic sectors dealing with information whose primary purpose is often not research-related. Activities are not discipline-specific; in fact data science is being described in some quarters as a new discipline.

The Call for Papers including a list of topics can be found at:- www.dcc.ac.uk/events/idcc13/call-papers.

Posted in dcc, Events | Comments Off

DCC Tools Catalogue

Posted by Marieke Guy on 8th May 2012

The Digital Curation Centre (DCC) has has recently updated their catalogue of tools and services for managing and curating research data.

The catalogue is available from
http://www.dcc.ac.uk/resources/external/tools-services

This is more than a new look; the catalogue has been overhauled to focus on software and services that directly perform curation and management tasks. It splits these resources into five major categories, based on who the intended users are and what stage of the data lifecycle they will be most useful in.

There is a category for Archiving and Preserving Information Packages with sub categories including:

  • Access Platforms – Tools to publish content and metadata to the web.
  • Backup and Storage Management – Tools to coordinate responsible storage and preservation strategies.
  • Creating and Manipulating Metadata – Enriching object descriptions and standardising records.
  • Emulation – Re-creating obsolete software environments to access old formats.
  • File Format ID and Validation – Defining and validating digital files.
  • Metadata Harvest and Exposure – Using OAI-PMH to share records across repositories.
  • Normalisation and Migration – Transferring digital materials into preservation-friendly formats.
  • Persistent ID Assignment – Creating unique identifiers for digital objects.
  • Repository Platforms – Enabling deposit, preservation, and access to digital content.

Sub-categories contain tables for quick comparison of tools against others that perform similar functions, linked to in-depth descriptions of how the resource can help.

This resource will evolve; if you have suggestions of tools to add please send them to info@dcc.ac.uk

Posted in Archiving, dcc, irg, tools | Comments Off

Preserving Moving Pictures and Sound

Posted by Marieke Guy on 1st May 2012

The Digital Preservation Coalition (DPC), Richard Wright and Charles Beagrie Ltd have announced the release of the latest DPC Technology Watch Report Preserving Moving Pictures and Sound, written by Richard Wright, formerly of the BBC.

Moving image and sound content is at great risk. Surveys have shown that 74 per cent of professional collections are small: 5,000 hours or less. Such collections have a huge challenge if their holdings are to be preserved. About 85 per cent of sound and moving image content is still analogue, and in 2005 almost 100 per cent was still on shelves rather than being in files on mass storage.

Surveys have also shown that in universities there is a major problem of material that is scattered, unidentified, undocumented and not under any form of preservation plan. These collection surveys are from Europe and North America because there is no survey of the situation in the UK, in itself a cause for concern. This report is for anyone with responsibility for collections of sound or moving image content and an interest in preservation of that content. New content is born digital, analogue audio and video need digitization to survive and film requires digitization for access. Consequently, digital preservation will be relevant over time to all these areas.

The report concentrates on digitization, encoding, file formats and wrappers, use of compression, obsolescence and what to do about the particular digital preservation problems of sound and moving images.” – Richard Wright

The report discusses issues of moving digital content from carriers (such as CD and DVD, digital videotape, DAT and minidisc) into files. This digital to digital ‘ripping’ of content is an area of digital preservation unique to the audio-visual world, and has unsolved problems of control of errors in the ripping and transfer process. It goes on to consider digital preservation of the content within the files that result from digitization or ripping, and the files that are born digital. While much of this preservation has problems and solutions in common with other content, there is a specific problem of preserving the quality of the digitized signal that is again unique to audio-visual content. Managing quality through cycles of ‘lossy’ encoding, decoding and reformatting is one major digital preservation challenge for audio-visual as are issues of managing embedded metadata.

DPC Technology Watch Reports identify, delineate, monitor and address topics that have major bearing on ensuring our collected digital memory will be available tomorrow. They provide an advanced introduction in order to support those charged with ensuring a robust digital memory and they are of general interest to a wide and international audience with interests in computing, information management, collections management and technology. The reports are commissioned after consultation with members; they are written by experts; and they are thoroughly scrutinised by peers before being released. The reports are informed, current, concise and balanced and they lower the barriers to participation in digital preservation. The reports are a distinctive and lasting contribution to the dissemination of good practice in digital preservation.

Preserving Moving Pictures and Sound is the second Technology Watch Report to be published by the DPC in association with Charles Beagrie Ltd. Neil Beagrie, Director of Consultancy at Charles Beagrie Ltd, was commissioned to act as principal investigator and managing editor of the series in 2011. The managing editor has been further supported by an Editorial Board drawn from DPC members and peer reviewers who have commented on the text prior to release. The Editorial Board comprises William Kilbride (Chair), Neil Beagrie (Series Editor), Janet Delve (University of Portsmouth), Sarah Higgins (Archives and records Association), Tim Keefe (Trinity College Dublin), Andrew McHugh (University of Glasgow) and Dave Thompson (Wellcome Library).

The report is available online.

Posted in dcc | Comments Off

Seasons Greetings!

Posted by Marieke Guy on 22nd December 2011

As I’m now working for the Digital Curation Centre (DCC) I’m going to make use of their Christmas Card to wish you all a Merry Christmas and a Happy New year!

Posted in dcc | Comments Off

DCC Roadshow in Cardiff

Posted by Marieke Guy on 15th December 2011

Snow, sleet, hailstones, rain and sunshine! The Cardiff weather couldn’t make up its mind, but the Digital Curation Centre (DCC) roadshow carried on regardless. Although I have attended various days of the travelling roadshow (Bath and Cambridge) I’ve never actually managed to catch a day one. The opening day is an opportunity to hear an overview of the research data management landscape and is also the day on which local case studies make it onto the agenda, so I was looking forward to it.

Welcome: Janet Peters, Cardiff University

Janet Peters, Director of University Libraries and University Librarian for Cardiff University, opened the day by saying how keen she was to have the roadshow take place locally; feeling it to be very timely given current research data management (RDM) work in Cardiff. Janet explained that her attendance of the Bath roadshow had kick-started Cardiff’s work in this area. Cardiff have recently revitalised their digital preservation group and have been providing guidance and assisting departments with implementing changes to their RDM processes – more on this later. They have also recently rolled out an institutional repository, though it doesn’t cover data sets (at the moment).

The Changing Data Landscape: Liz Lyon, UKOLN

Liz Lyon on The Changing Data Landscape

Liz set the scene for the day by outlining the current data landscape. She began by introducing the new BIS report entitled Innovation and Research Strategy for Growth which expresses the government’s support for open data and introduces the Open Data Institute (ODI). Only last week David Cameron made the suggestion that “every NHS patient should be a “research patient” with their medical details “opened up” to private healthcare firms”. Openness and access to data are two of the biggest challenges of the moment and have stimulated much debate. Liz gave the controversial example of one tobacco companies FOI request to the University of Stirling for information relating to a survey on the smoking habits of teenagers. She explained that proposed amendment to FOI data will allow institutions to ask for exemption to FOI requests when research is ongoing. It’s often the case that researchers don’t want to share data and there have been instances when governments have placed restrictions on data use(e.g. the bring your genes to cal project. Liz shared some examples of more positive cases of when research is shared e.g. Alzheimers research, 1000 Genomes Project, Personal Genome Project, openSNP. She also offered some citizen science examples: BBC nature, project Noah http://www.projectnoah.org/, Galaxy Zoo, Patients Participate, BBC Lab. The Panton Principles are a recent set of guidelines that offer possible approaches: Open knowledge, open data, open content and open service. To some degree the key to all of this is knowing about data licensing and the DCC offer advice in this area.

Liz then moved on to what is often seen as the biggest challenge of all: the sheer volume of data now created e.g. large hydron collider. In the genomics area there are lots of shocking statistics on the growth of data and the implications of this. Another new report phg foundation: Next steps in the sequence gives the implications of this data deluge for the NHS. The text the Forth paradigm highlights data intensive research as being the next step in research. The DCC are working with Microsoft Research Connections to create a community capability model for data intensive research

It is apparent that big data is being lost, but so is small data (like excel spreadsheets) and part of the challenge is working out how scientists can deal with the longtail. What is framed as gold standard data is when you can fully replicate the code and the data, reproducible research is the second best approach. Data storage needs to be scalable, cost-effective, secure, robust and resilient, have a low entry barrier, have ease of use. Liz also also asked us to consider the role of cloud services, giving Dataflow http://www.dataflow.ox.ac.uk/, VIDaaS, BRISSKit, lab notebook as 4 JISC projects to follow in this area.

Liz then talked a little about policy, giving research council examples. The most relevant is the fairly demanding EPSRC expectations that have serious implications for HEIs: Institutions must provide a RDM roadmap by 1st May 2012 and must be compliant with these expectations by 1st May 2015. At the University of Bath, where Liz is based, there is a new project called research360@Bath and they have a particular emphasis on faculty-industry focus. There will also be a new data scientist role based at UKOLN. A full list of funders and their requirements is available from the DCC Web site.

Resources are available and the Incremental project http://www.lib.cam.ac.uk/preservation/incremental/ back in 2010 found that many people felt that institutional policies were needed in the RDM area. Edinburgh have developed an aspirational data management policy. The DCC have pulled together exemplars of data policy information http://www.dcc.ac.uk/resources/policy-and-legal/institutional-data-policies, ANDS also have a page on local policy.

It is also important to consider how you incentivise data management? There is quite a lot of current work on impact, data citation and DOIs. Some example projects: Total Impact http://total-impact.org/ and SageCite.

And what about the cost? Useful resources include the Charles Beagrie report on Keeping Research data safe http://www.beagrie.com/jisc.php, Neil Begrie has done some work into helping people articulate the benefits through use of a benefits framework tool.

In conclusion Liz asked delegates to think about the gaps in their institution.

Digital Data Management Pilot Projects: Sarah Philips, Cardiff University

Sarah explained how at Cardiff the University had retention requirements for quite a lot of corporate records and permanent records. They also have requirements for some of their research data for 5 -30 years. The University has set up three pilot projects in response to feedback from a digital preservation policy in the cultural area, in the school of Biosciences using genomic data and in the school of history and archaeology. Work in the school of history and archaeology department is now coming to a close and this is the area Sarah would concentrate on.

Three projects within the department were used as a test bed. The South Romanian Archaelogical Project (SRAP) at the University had collected excavation data and the team have been keen to make the data available. The Magura Past and Present Project had artists coming in and creating art; because the project was an engagement project it was required that the outputs be available, though not necessarily the data. The final project was on auditory archaeology. All three projects were run by Doctor Steve Mills.

Records management audits were carried out through face-to-face interviews with staff using the DCCs Data Asset Framework. Questions included: what records and data are held? How are the records and data managed and stored? What are the member of staffs requirements? A data asset register was created that dealt with lots of IP issues, ownership issues etc. Once this data was collected potential risks were identified e.g. Dr Mills had been storing data on any other hard-drives available but he didn’t have a systematic approach to this, there was some metadata available but file structure was an issue, proprietary formats were used and there are no file naming procedures in place. Dr Mills was keen to make the data accessible so the RDM team have been looking at depositing it with the Archaeology Data Service, if this solution isn’t feasible they will have to use an institutional solution.

High Performance Computing and the Challenges of Data-intensive Research: Martyn Guest, Cardiff University

Martyn started off by giving an introduction to advanced research computing
at Cardiff (ARCAA) which was established in 2008. Chemistry and physics have been the biggest users of high performance computing so far, but the data problem is relatively new and has really arisen since the explosion of data use by the medical and humanities schools.

He sees the challenges as being technical (quality performance, metadata, security, ownership, access, location and longevity), political (centralisation vs departmental), governance, ownership) and personal, financial (sustainability), legal & ethical (DP, FOI). Martyn showed us their data intensive supercomputer (‘Gordon’) and a lot of big numbers (for file sizes) were banded about! Gordon runs large-memory applications (supermode) – 512 cores, 2 TB of RAM, and 9.6 TB of flash. It has been the case that NERC has spent a lot of time moving data leaving less effort for analysing the data.

Martyn shared a couple of case studies: Positron Emission Tomography Imaging (PET) data where the biggest issues were that the data was raw, researchers weren’t interested in patient identifiable data but want image while clinicians wanted PID and image. He talked about sequencing data , which is now relatively easy, the hard bit is using biometrics on the data. As Martyn explained it now costs more to a analyse a genome than to sequence it and the big issue is sharing that data. Martin joked that the “best way to share data is by Fedex”, many agreed that this may often be the case! The case studies showed that in HPC it’s often a computational problem. HPC Wales has three various components to it including awareness building around HPC and the creation of a welsh network that can be accessed from anywhere and globally distributed.

Martyn concluded that the main issues are around how to do the computing efficiently while the archiving issues continue to be secondary.

Research Data Storage at the University of Bristol: Caroline Gardiner, University of Bristol

Caroline Gardiner explained that at the University of Bristol her team had originally carried out a lot of high performance computing but were increasingly storing research data. She noted that the arts subjects are increasingly creating huge data sets.

Caroline admitted to collecting horror stories of lost data and using this as a way to leverage support. The Bristol solution has been BluePeta which has been created using £2m funding and is a petascale facility. This facility is purely for research data at the moment, not learning and teaching data, thought is an expandable facility.

Caroline explained that their success in this area came from many directions. Bristol already had a management structure in place for HPC and for research data storage, they had access to the strategy people and those who held the purse strings. Bristol also have a research data storage and management board, there continues to be buy in from academics.

The process in place is that the data steward (usually principal investigator PI) applies and can register one or more projects. There is then academic peer review and storage policies applied. There is a cost model in place, the data steward gets 5TB free and then have to pay £400 per TB for annum disk storage. They are encouraging PIs to factor in these costs when writing their research grant applications. The facility is more for data that needs to be stored over the long term rather than active data.

Bristol are also exploring options for offsite storage and will also be looking at an annual asset holding review. They are also looking at preparing an EPSRC roadmap and starting to address wider issues of data management.

In answer to a question Caroline explained that they had made cost analysis against 3rd party solutions but when using the big players (like Google and Amazon) the cost of moving the data was the issue. There was some discussion on peer-to-peer storage but delegates were concerned that it would kill the network.

Data.bris: Stephen Gray, University of Bristol

Following on from Caroline’s talk Stephen Grey talked about what was happening on the ground through data.bris. Stephen explained that the drivers for the project were meeting the funder requirements (not just EPSRC), also meeting the publisher requirements and using research data in the REF and to increase successful applications. Bristol have agreed a digital support role alongside the data.bris project, though this ia all initially limited to the department of arts and humanities.

The team will be initially meeting with researchers and using the DMPOnline tool to establish funder requirements and ethical, IPR and metadata issues. After the planning there will be the research application and then hopefully research funding. The projects will then have access to BluePeta storage. The curation is planned to happen at the end of the project and high valued data identified for curation. Minimal metadata should be added at this stage, though there is a balancing act here between resourcing and how much metadata is added. Bristol have a PURE research management system and data.bris repository where they can check the data and carry out metadata extraction and assign DOIs. They will then promote and monitor data use

In the future the team also want to look into external data centres use. A theme running through the project is ongoing training and guidance and advocacy and policy. Training will need to go to all staff including IT support and academic staff and they are hoping for some mandatory level of training.

Bristol are also planning on using the DCC’s CARDIO and DAF tools

In the Q&A session delegates were interested in how Bristol had received som much top-down support for this work. It was explained that the pro VC for research ws a scientist and understood the issues. While there was support for research data it was felt that there could do with being more support for outputs.

Herding Cats – Research Publishing at Swansea University: Alexander Roberts, Swansea University

Alexander Roberts started off his presentation by saying that Swansea wants it all: all data, big data, notes scribbled on the back of fag packets, ideas, searchable and mineable data. Not only this but Swansea would like it all in one place, currently they have a lot of departmental data bases and various file formats in use. Swansea looked at couple of different systems including PURE but wanted an in house content management system, they also inherited a DSpace repository. They wanted this system to integrate with their TerminalFour Web CMS, with their DSpace system Cronfa and to give RSS feeds for staff research profiles, give Twitter feeds, Facebook updates etc. There was a consultation process that allowed lots of relationships to be formed and the end users to be involved. People were concerned that if they passed over their data they wouldn’t be able to get it back. A schema was created for the system. They started off using Sharepoint and were clear that they wanted everything in a usable format for the REF. The end result was built from the ground up: a form-based research information system that allowed researchers to add their outputs as easily as possible. It is a simple form based application that integrates with the HR database and features DOI resolving, MathML. The ingest formats are RSS, XML, Excel, Acess and others. It provides Open Data Protocol (oData) endpoint which provides feeds to the Web CMS and personal RSS feeds.

Alexander ended by saying that in 2012 they would like to implement automatic updates to DSPACE via SWORD and a searchable live directory of research outputs. They also want to have enhanced data visualisation tools for adminstrators. Mobile consideration is also a high priority as Swansea have a mobile first policy.

Michael Day and Alexander

Delivering an Integrated Research Management & Administration System: Simon Foster, University of Exeter

A Research Management and Administration System (RMAS) is more about managing data about projects but can also deal with research data. The Exeter project has been funded under the UMF, funded by HEFCE through JISC and is part of the HEFCE bigger vision of cloud computing and join up of systems. HE USB is being used: a test cloud environment from Eduserv. Simon Foster described how the project had started with a feasibility study which looked at whether there was demand for a cradle to grave RMAS system, 29 higher education institutions expressed interest. The project was funded and it was worked out that 29 HEIs phased in over ten years could save £25 million. The single supplier approach was avoided after concerns that it could kill all others in the market. The steering group looked at the processes involved and these were fed into a user requirement document. It was necessary that it was cloud enabled and were compliant with CERIF data exchange. Current possible systems include Pure, Avida etc. Specific modules were suggested. The end result will be a framework in place that will allow institutions to put out a mini-tender for RMAS systems asking specific institution related questions. Institutions should be able to do this in 4 weeks rather than 6 months.

The next steps for the project are proof of concept deliverables using CERIF data standards and use of externally hosted services. They also want to work with other services, such as OSS Watch.

There followed a panel session which included questions around the cost implications of carrying out this work. One suggestion was to consider the cost of failed bids due to lack of data management plans.

What can the DCC do for You?: Michael Day, UKOLN

Michael Day finished off the day with an overview of the DCC offerings and who they are aimed at (from researchers to librarians, from funders to IT services staff). He reiterated that part of RDM is bringing together different people from disparate areas and clarifying their role in the RDM process. The DCC tools include CARDIO, DAF, DMP Online, DRAMBORA. Some of the services include policy development, training, costing, workflow assessment etc. DCC resources are available from the DCC Website.

Conclusions

So after a day talking about data deluge while listening to a deluge of the more familiar sort (loud hail and rain) we were left with a lot to think about.

One interesting insight for me were that while the data deluge had come originally from certain science areas (astronomy, physics etc.) now more and more subjects (including arts and social sciences) are creating big data sets. One possible approach, advocated by a number of the day’s presenters, is to use HPC as a starting point from which to jolt start research data management. However there will continue to be a lot of data ‘outside of the circle’. As ever, join up is very important. Getting all the stakeholders together is essential, and that is something the DCC roadshows do very well. All presentations from the day are available form the DCC Web site.

The next roadshow will take place from 7 – 8 February 2012 in Loughborough. It is free to attend.

Tags:
Posted in Conference, dcc, irg | Comments Off

Free Research DM Workshop, Cambridge, 9-11 Nov

Posted by Marieke Guy on 4th October 2011

Unfortunately the DCC Brighton Roadshow was cancelled but the next roashow isn’t far off and places are still available.

To give some background….

DCC logoThe UK Digital Curation Centre is running a series of free inter-linked regional workshops aimed at supporting institutional research data management planning and training. The DCC Roadshows are designed to allow every institution in the UK to prepare for effective research data management and understand more about how the DCC can help. The sixth DCC Roadshow is being organised in conjunction with Cambridge Library and will take place from 9th – 11th 2011 November in the Paston Brown Room at the Homerton Conference Centre, Cambridge.

The roadshow runs over three days but each workshop can be booked individually. Attendees are encouraged to select the workshops which address their own particular data management requirements. The workshops will provide advice and guidance tailored to a range of staff, including PVCs Research, University Librarians, Directors of IT/Computing Services, Repository Managers, Research Support Services and practising researchers.

Day one is an introductory day aimed at researchers, data curators, staff from library etc. It provides an introduction to the DCC and the role of the DCC in supporting research data management. Day two is a more interactive day aimed at senior managers, research PVCs/Directors, directors of Information Services etc. and looks at strategy/policy implementation. Day three is a hands-on day and consists of the Digital Curation 101 – How to manage research data: tips and tools workshop.

To find out more about the workshops take a look at the DCC Cambridge Roadshow page. Registration for the workshop is free but places are limited.

If you can’t decide if the roadshow is for you Steve Walsh from the Interoperable Geospatial Data for Biosphere Study( IGIBS) Project, Aberystwyth University, has written a review of the most recent workshop held in Oxford.

Details of further roadshows will be announced soon on the DCC Web site.

Tags:
Posted in dcc, irg, Project news | Comments Off

DCC and the Sussex Roadshow

Posted by Marieke Guy on 26th September 2011

I’ve recently been appointed as an Institutional Support Officer for the Digital Curation Centre. In this role I will be raising awareness and building capacity for institutional research data by liaising with libraries, IT services, research support staff and others.

My first step in getting myself up to speed will be attending the DCC Roadshow to be held in Brighton from the 4th – 6th October 2011 at the University of Sussex Conference Centre. I attended one day of one of the earlier roadshows held in Bath and thoroughly enjoyed it, so I am very much looking forward to it.

The roadshow consists of three days of training in research data management. Day one is an introductory day aimed at researchers, data curators, staff from library etc. It provides an introduction to the DCC and the role of the DCC in supporting research data management. Day two is a more interactive day aimed at senior managers, research PVCs/Directors, directors of Information Services etc. and looks at strategy/policy implementation. Day three is a proper hands-on day and consists of the Digital Curation 101 – How to manage research data: tips and tools workshop.

Attendees are welcome to dip in an out of the workshops and don’t have to attend the full three days. There are still places available and I’m sure it will a very useful couple of days. Might see you there!

Tags: ,
Posted in dcc, Events | 2 Comments »

Closing the Digital Curation Gap

Posted by Marieke Guy on 6th July 2011

Last week on the day before the ICE Forum (28th June 2011) I attended the Closing the Digital Curation Gap Meeting.

CDCG is an International Collaboration to Integrate Best Practice, Research & Development, and Training in Digital Curation. It has been running since October 2009 and was scheduled to finish in September this year but has just been given an extension (till September 2012). A comprehensive overview of the project is given on the Digital Curation Exchange Web site.

The Closing the Digital Curation Gap (CDCG) collaboration is designed to serve as a locus of interaction between those doing leading edge digital curation research, development, teaching, and training in academic and practitioner communities those with a professional interest in applying viable innovations within particular organizational contexts; IMLS; JISC; the DCC, charged with disseminating such innovation and best practices; and the SCA, charged to build a common information environment where users of publicly funded e-Content can realize best value by reducing the barriers that inhibit access, use and re-use of online content.

I have come along to the project at a fairly late stage but hope I can still be of use and possibly offer a new perspective (that of not being an expert!).

The June meeting was held at the JISC offices in London and was a joint meeting of the US and UK partners. The UK was represented by members from JISC, UKOLN, ULCC, HATII, the BL and the DPC, the US had people from the Bishoff Group, Penn State University Libraries, Purdue University Libraries, University of North Carolina at Chapel Hill and University of Toronto. [Thanks to Sharon McMeeking from the DPC for sharing her notes to help jig my memory].

The aims of the meeting were to discuss the outputs of the project so far and to set objectives for the continuation of the work in 2011/12. The main work so far has been staging a number of focus groups, work on decision trees and work on best practice guides. The digital curation exchange web site is the key resource that has been created. Much of the meeting involved discussion of the digital curation exchange: we were encouraged to pass on constructive critism, suggestions on process and ideas for future resources.

They have quite a lot to work on before the next meeting – good luck to them!

Tags:
Posted in dcc, Events | Comments Off

DCC Roadshow 2010 – 2011

Posted by Marieke Guy on 20th August 2010

The Digital Curation Centre have carried out digital preservation training in the past (for example the Digital Curation 101 course) but they have now committed to running a series of data curation roadshows. These are likely to be very useful anyone involved in digital curation from senior managers to researchers and librarians.

Institutional Challenges in the Data Decade

The DCC Roadshows will comprise of a series of inter-linked workshops aimed at supporting institutional data management, planning and training.

The first will take place 2-4 November in Bath and will be open to participants from Higher Education Institutes (HEIs) in the south-west of England. The roadshow will run over 3 days and comprise of a series of day and half day workshops.

For more details see the DCC Web site. Registration will open in September 2010.

Tags:
Posted in dcc, Project news | Comments Off