Overview and details of the sessions of this conference. Please select a date or room to show only sessions at that day or location. Please select a single session for detailed view (with abstracts and downloads if available).
|Date: Tuesday, 09/Jun/2015|
|9:00am - 9:30am||Welcome to OR2015|
|9:30am - 10:30am||PLN1: Opening Plenary: Leveraging the Web for Research, Kaitlin Thaney (Mozilla Science Lab)|
|11:00am - 12:30pm||P1A: Linked Open Data (LOD)|
Session Chair: Simeon Warner
Fedora4: The Open Linked Data Platform
1Duraspace, United States of America; 2Stanford University
Linked Open Data has moved from being a buzzword to a fundamental building block of modern repositories and information systems. Its explosion, taking in this domain the form of scholarly and scientific datasets, publications, annotations, cultural heritage descriptions and other repository-based content, offers unprecedented opportunity for scientific and societal advancement. It is the interconnections that integrate systems and resources, however, that turn disparate ideas into unanticipated solutions.
The Open Repositories community largely understands the value of linked data. The trouble has been in answering the question of “how?” to do this together, rather than "why?" do it at all. In order to be effective, we need to have solid guidelines, common practices, and well specified toolsets and products. A confluence of developments has opened the door to exactly this.
In October of 2012, the initial W3C working draft of the Linked Data Platform 1.0 (LDP) document was published. In July of 2012, the demand for a next-generation Fedora platform was channeled into the Fedora 4 (F4) project, which at the time was termed Fedora Futures. The alignment of these two stars set in motion events that provided both requirements and solutions for the community.
Ozmeka: extending the Omeka repository to make linked-data research data collections for all research disciplines
1University of Technology Sydney, Australia; 2Intersect limited, Australia
The Ozmeka project is an Australian open source project to extend the Omeka repository system. Our aim is to support Open Scholarship, Open Science, and Cultural Heritage via repository software than can manage a wide range of Research (and Open) Data, both Open and access-restricted, providing rich repository services for the gathering, curation and publishing of diverse data sets. The Ozmeka project places a great deal of importance in *integrating with external systems*, to ensure that research data is linked to its context, and high quality identifiers are used for as much metadata as possible. This will include links to the ‘traditional’ staples of the Open Repositories conference series, publications repositories, and to the growing number of institutional and discipline research data repositories. In this presentation we will take a critical look at how the Omeka system, extended with Ozmeka plugins and themes can be used to manage three diverse projects and talk about how this work paves the way for eResearch and repository support teams to supply similar services to researchers in a wide variety of fields. This work intended to reduce the cost of and complexity of creating new research data repository systems.
Why FRBRoo and CIDOC CRM are great for expressing (Linked, Open) Ethnographic Research Data
University of Prince Edward Island, Canada
Can we use a Linked Open Data framework in place of classic static metadata in a Fedora Repository? We describe how we catalogued a collection of interrelated ethnographic research data using the FRBRoo and CRM ontologies. These vocabularies allowed us to express detailed relationships between digital objects and entities (people, places, events, concepts) in a more nuanced way than traditional bibliographic metadata schemas such as MODS and MADS. Our implementation uses a network of entities in a Fedora repository, with the CRM and FRBRoo properties in the Mulgara triplestore, to catalogue data from an ethnographic project in a way that will drive the Islandora display, while allowing the network to be queried by researchers.
We describe the process of transforming our data set, consisting of audio and video recordings; photographs; biographical information; music notation files; and textual content, for inclusion into this RDF-centric repository, and the challenges encountered in modelling born-digital content in vocabularies designed to contain surrogates for physical objects.
|1:30pm - 3:00pm||P2A: Integrating with External Systems: the use case of ORCID|
Session Chair: Maureen Walsh
Panel: So we all have ORCID integrations, now what?
1University of Notre Dame; 2University of Missouri System; 3Dryad Digital Repository; 4Digital Repository Services Ltd; 5mire
In the past year, the major groundwork has been laid for repository systems to support ORCID identifiers. DSpace, Hydra, and EPrints all have support for storing and managing ORCIDs. However, we are still in the early stages of ORCID adoption. Only a small fraction of repository content is annotated with ORCIDs, and most end-users have not yet realized any benefit from the features based on ORCID.
This panel will bring together representatives of major repository systems to relate the current status of ORCID implementations, discuss plans for future work, and identify shared goals and challenges. The panelists will discuss how ORCID support provides practical benefits both to repository staff and end-users, with a focus on features that exist now or will exist in the next year.
Horizontal vs vertical organizations of repositories
ORCID, Inc, United States of America
Over its lifetime, an organization’s structure often transitions between a vertical (hierarchical) one and a horizontal (flat) one. Each structure has its benefits and challenges as it relates to how information is communicated, how work gets done, and how collaborative or self-contained the organization can or must be in interacting with others. This session explores these models as applied to repositories in the interconnected ecosystem in which they exist. Specifically it will consider:
* What are the attributes of a “vertical” or highly-aggregated repository? How does this differ from a “horizontal” or highly-distributed one? What are some examples of each type?
* What are the benefits and challenges that each model provides?
* In what situations does it make sense for the two models work together?
* A case study: how some traditionally “vertical” repositories are benefiting from the “horizontal” information model being provided by tools like ORCID.
|3:30pm - 5:30pm||P3A: Integrating with External Systems|
Session Chair: Maureen Walsh
[24x7] ORCID Integration: Services to Create and Use ORCID IDs at KAUST
King Abdullah University of Science and Technology, Saudi Arabia
As ORCID IDs become accepted as a global standard, the potential benefits to researchers and institutions of making use of them are multiplying. At King Abdullah University of Science and Technology (KAUST) we adopted the first institutional open access policy in the Arab region in June 2014 and are integrating with ORCID as part of services to researchers that improve the preservation and dissemination of their research. We started by using ORCID IDs to identify authors in our institutional repository. Our next step was to join ORCID as an institutional member and then set up a plan to support our faculty, researchers and students in the creation and use of ORCID identifiers. This presentation will look at the choices made and lessons learned during this process, focusing on the tools we developed to interact with the ORCID member API and the ways in which introducing ORCID has complemented other repository initiatives, such as the implementation of our institutional open access policy.
[24x7] From archives to repository: an archival collection management system and repository integration case study
Tufts University, United States of America
At the Digital Collections and Archives (DCA) at Tufts University we have designed, built, and integrated our archival collection management system and repository’s administrative interface to facilitate ingesting archival objects into our Fedora based repository. This 24x7 session briefly explores the assumptions and functional requirements we have used to guide this development work. The DCA’s unique position as an archives that is one of the key stakeholders and users of the Tufts institutional repository has enabled us to meet this integration challenge. The session describes how the integration of our archival collection management system and our repository relies on the ability to flexibly move metadata from one system to another.
[24x7] 77 Lines of XSLT and a Roll of Duct Tape: ArchivesSpace as Metadata Hub in a Multi-Repository Environment
University of Denver Libraries, United States of America
ArchivesSpace is enjoying increasing adoption among archives and special collections units as a collection management system. Its API supports a variety of metadata exports, making it a tempting place for archives engaged with digitization and digital repository management efforts to create and manage metadata for those projects. While the API is powerful, the differences in exports among the various content modules can make building this shared metadata infrastructure difficult, especially in a hosted environment.
This talk will focus on efforts at the University of Denver Libraries to build metadata pipelines connecting ArchivesSpace to other data repositories, particularly our XTF finding aid database and our Islandora-based institutional repository. Challenges we encountered along the way, successes, and ongoing thorns in our side will be addressed. An objective of this talk is to start, or contribute to, a conversation among other ArchivesSpace institutions about how we can improve its support for repository initiatives in archives and special collections.
Making Connections: The SHARE Notification Service and the Open Science Framework
Center for Open Science, United States of America
The Center for Open Science (COS) is a non-profit technology startup seeking to enhance the openness, integrity, and reproducibility of scientific research. One way COS approaches this mission is by building infrastructure to facilitate good scientific practice, disseminate scholarly communication and scientific research, and facilitate the accurate accumulation of scientific knowledge. COS looks to connect all parts of the research lifecycle: planning, execution, reporting, archiving, and discovery. Its free, open source, flagship product, The Open Science Framework, leverages application programming interfaces (APIs) from various services to unite resources and increase the efficiency of researchers. In partnership with the Association of Research Libraries, COS is building the SHared Access Research Ecosystem (SHARE) Notification Service. This talk will detail how the Open Science Framework connects various services and repositories and incorporates the SHARE notification service. It will include information about future connections, and how the community can influence these connections.
“Headless” Metadata for Library Discovery: NYU’s Ichabod Project
NYU, United States of America
From DSpace to Drupal, NYU has a variety of systems to ingest and display curated digital content. To make this content discoverable centrally, we developed a tool for metadata ingest, transformation, and discovery based on a popular open-source software stack: Fedora, Hydra, Solr, and Blacklight. Called “Ichabod,” this tool has allowed us to ingest, normalize, and enrich metadata from diverse systems of record and make it consumable by our main discovery tool, which is powered by the Ex-Libris product Primo. We developed Ichabod using the Agile methodology and involving developers from three distinct NYU Libraries groups. The software will lay the groundwork for future innovation in the areas of metadata management and discovery for repository content. The relationships we established have already made it possible for a similar collaboration arrangement on two other projects, with more to come in the future.
Integration with external systems and services: challenges and opportunities
Jisc, United Kingdom
Services that support open access need to interoperate effectively with each other and with external systems if they are to succeed in their mission to help institutions to become as effective and efficient as possible in capturing and sharing their research. Jisc, which provides shared services along these lines to UK institutions, has identified this as a particular challenge. It works with a number of partner organizations, including the University of Nottingham (Sherpa Services), EDINA, Mimas and the Open University (Knowledge Media Institute), in developing these services.
Achieving integration between these partner services and institutional systems has provided a number of benefits and opportunities, and has also highlighted a number of challenges. These include how best to integrate with local workflows, technical compatibility, reliance on data from publishers and, in turn, questions of trust in the services built on this data. This presentation looks at these issues from the point of view of a provider of shared services, but also includes examples from the institutional perspective.
|Date: Wednesday, 10/Jun/2015|
|9:00am - 10:00am||PLN2: Indexing repositories: pitfalls and best practices, Anurag Acharya (Google Scholar)|
|10:30am - 12:30pm||P4A: Managing Research (and Open) Data|
Session Chair: Holly Mercer
[24x7] Revisiting Self-Deposit of Scientific Data
Stanford University, United States of America
Sharing scientific data is increasingly valuable for reproducible science, furthering investigation, and innovation. To this end, repositories facilitate data sharing by making scholarly data available. We are at an impasse, however. Librarian-mediated approaches to self-deposit of scientific data are very resource-intensive, and the repository services provided to researchers are often limited. Self-deposit is quite a challenging use case as it encompasses data preparation, metadata description, upload, visualization, annotation, sharing, publication, access, rights, preservation, citation, and discovery services. This editorial suggests we revisit the value proposition we make for self-deposit and mitigate its resource-intensive workflows.
[24x7] CERN Open Data and Data Analysis Preservation
We present newly launched CERN Open Data Portal and related long-term Data Analysis preservation activities. Using the Invenio digital library platform and taking inspiration from OAIS preservation practices, the knowledge associated with successive data analysis steps is being captured for further reuse. The aim is to preserve not only information about research datasets, but also about the underlying user software and virtual machine platforms used to study it, together with any configuration parameters and high-level physics information associated with the analysis process. The CERN Open Data portal disseminates selected primary and reduced datasets of LHC experiments and offers several high-level tools permitting general public and general data scientists to visualise and further work with the data, such as interactive event display or histogram plotting interfaces. The ultimate goal of data analysis preservation efforts is to be able to reproduce an analysis even many years after its initial publication, permitting to extend the impact of preserved analyses through their future revalidation and recasting.
[24x7] Integration and Adoption: An ORCID story
Symplectic, United Kingdom
As academic engagement with institutional repositories moves from “why should I do this?” to “good idea, but how can the Library make this easier for me?”, the need for consistent and unambiguous metadata has never been greater.
Metadata consistency includes the unambiguous identification of authors, editors, supervisors and other contributors to repository objects but until the launch of ORCID, there wasn’t a common means of unambiguously identifying authors.
In this presentation, we will explain how Imperial College London - the first institution to integrate a research information management system with an institutional repository - enabled over 1,200 research active staff within a week to claim an ORCID and subsequently automate the harvest of data from ORCID helping to populate Imperial’s institutional repository with verified metadata.
Islandora as an access system for iRODS managed information packages
Zuse Institute Berlin (ZIB), Germany
Accessing information packages with Islandora is straight forward, albeit not so much when they reside within a federated data management environment. In our case, dissemination information packages live in the Fedora object store for immediate access. The archival information packages are stored safely in a hierarchical storage infrastructure managed by iRODS and are only accessible for administrative and preservation action purposes. We present a data model that supports both use cases utilizing just a single Islandora instance. To integrate with iRODS, we developed an Islandora module to display and deliver data and metadata from the storage location. This solution also allows us to extend the system with further preservation workflow actions that will be required in the future.
Databrary: A research-centered repository for video data
New York University, United States of America
As a research data repository, Databrary focuses specifically on the storage, discoverability, and sharing of video-based datasets within the developmental and learning sciences. Storing video presents its own unique opportunities and challenges, the latter of which include research subject privacy and difficulties in creating and storing metadata that comes from different research projects in a standardized fashion. Databrary has implemented policies and practices within a functioning web application that meets both the needs of researchers as well as the preservation and access needs to share these datasets into the future. The lessons learned thus far in developing Databrary stand to model a viable approach to establishing practices and workflows for gathering and organizing research data that lift the burden off of researchers and also have potential to feed into established library systems for broader findability and accessibility.
The Hydra Common Data Model
1University of California, San Diego; 2Stanford University; 3Princeton University
One of the many successes of the Hydra community is the fundamental notion from which its name is derived—the concept of many interfaces (“heads”) over top of a single repository (the “body”). The recent release of Fedora 4, with its internal RDF-centric model, has spurred efforts for a community-wide model of collections and works, such that the heads can be sure that the body will behave as they expect it to. That model has been designed and vetted by the Hydra community, and its architecture and initial implementations will be presented in this paper.
|1:30pm - 3:30pm||P5A: Building the Perfect Repository|
Session Chair: Richard Green
[24x7] Reimagining Institutional repositories through Inclusiveness and Collaboration
Indiana State University, United States of America
Many college and university institutional repositories (IRs) were initially established to archive and provide access to students’ scholarly works by means of Electronic Theses and Dissertations (ETDs). However, not all students’ culminating projects are written works, resulting in a lack of representation in college IRs. IR content on college campuses often exclude art students’ portfolios, performance portfolios, and exhibition projects due to IR’s narrow definition of scholarly work as written work. By adapting IR technology to include images, video, and other non-written formats, and collaborating with underrepresented groups, Indiana State University (ISU) is expanding its IR to fully represent the graduate students of the university. The expanding of Indiana State University’s institutional repository collection focus allows the university to better serve the student community while also making the IR more robust and inclusive.
[24x7] Metadata Enrichment
Boston Public Library, United States of America
The Digital Commonwealth harvests information from over a dozen institutions that have varying metadata practices. Rather than “dumb down” metadata to the lowest common denominator, we have built a bunch of tools to instead make it more intelligent. From geographic parsing against linked data sources to standardizing of dates, we have successfully created rich records from disorganized sparse ones. This talk is about our combination of tools and practices we use to make this happen.
Beyond a look at the tools we have developed or use ourselves, we will be showing samples of the cool stuff one can do once the metadata has been enhanced. For example, demonstrating how well parsed geographic data can now be used on an interactive map with records from 100+ institutions and not result in absolute chaos.
The Once and Future Repository, HKU's Scholars Hub
1University of Hong Kong, Hong Kong S.A.R. (China); 2euroCRIS, GrandIR Ltd; 3Cineca
The HKU Scholars Hub (the Hub) began service as a traditional institutional repository of The University of Hong Kong (HKU). However this format was not compelling to HKU researchers. Fortunately a subsequent reformation of the HKU statement on university mission and vision infused new life and purpose into the project. Over the next five years, in partnership with the Italian University Consortium, Cineca, the HKU Libraries transformed the Hub from an IR to a Current Research Information System. We expect that future development will see the Hub further transformed into a research information management system supporting both internal decision support and external public discovery. We will present new work developed recently to further these goals.
Large Scale Repository Auditing to ISO 16363 – The Final Audit
1University of Minho; 2KEEP SOLUTIONS; 3FCT/FCCN
This paper describes an audit process carried out on 26 digital repositories according to the recently approved standard ISO16363 (Audit and Certification of Trustworthy Digital Repositories). The 26 repositories share a common infrastructure hosted by RCAAP (Open Access Scientific Repository from Portugal), a free hosting service provided to research institutions in Portugal. It addresses the process and the strategic alignment with the project objectives integrated with other developments related to digital preservation of institutional repositories. After the preliminary audit, this work presents the final results of the analysis of the three topics: Organizational Infrastructure, Digital Object Management and Infrastructure and Security Risk Management.
The Evolving Repository Landscape: Identifying Motivations for Library DAMS Migration
1University of Illinois at Urbana-Champaign; 2University of Houston
The presentation will highlight results from a survey that traces institutions’ motivations for migrating from one DAMS to another. It will discuss themes and features desired in future DAMS, outlining specific topical areas that will inform future system and work flow development, as well as governing body/vendor relations. The presentation will also discuss preliminary reflections on the comparison between initial system selection and requirements for selecting and migrating to a new system. Researchers will conclude the presentation by reviewing lessons learned from the research methodology as well as discussing future areas of research related to this study.
|4:00pm - 5:00pm||P6A: Repository Rants and Raves|
Session Chair: Simeon Warner
Sharing with purpose: A rant about rights statements
Digital Public Library of America, United States of America
Sharing with purpose: A rant about rights statements
Rant: Are we engaging in market segmentation? And is this a good thing?
University of CIncinnati, United States of America
We are familiar with market segmentation, the business practice to go after several subsets of the same market, in the belief that the market subsets will total to more than the original undivided market. Sometimes however, the competition is so intense that it puts out of business the product that was formerly the strongest, and the company loses the market entirely.
In our institutions, the explosion of research services in what is broadly called ‘digital scholarship’ has spawned programs in the digital humanities, data management, digital publishing, digital forensics, web archiving, online records management, etc. We have different platforms tailored to these efforts. We may be building different standards emerging from disparate practices. Is this a healthy diversity or dangerous competition for scarce resources?
Are we in danger of creating an incoherent brand, and competing among ourselves for a small market of digital content creators who cannot between them feed all of our services?
I will not posit that we should serve all digital content from one interface, but that we do need to put more effort into seamless connections between systems that will encourage repositories to be a unifying layer, so we do not fail in our preservation mission.
Rave: The Development of a Research Data Management Training Programme Tailored to the needs of Early Career Researchers and Postgraduate Students in the Visual Arts
The Glasgow School of Art, United Kingdom
Research data management in the visual arts can be seen as complex and diverse with actual research data being tangible and intangible; digital and physical; heterogeneous and infinite; and complex and complicated, it does not always fit into the natural scheme of data management. Therefore the development of policies for arts related data management and systems/infrastructure and training packages should be aware of thinking outside the box, and lending themselves to being more iterative and open to interpretation. With the help of funding from JISC and the AHRC The Glasgow School of Art (GSA) has been able to partner with other arts institutes to look at the concept of RDM in the arts, and develop training packages and toolkits to help support researchers and postgraduate students. The projects were KAPTUR, which ultimately brought about VADS4R, both of which were led by the Centre for Digital Scholarship, a research centre at the University for the Creative Arts.
Rave: Integrating the Institutional Repository with the Learning Management System
Georgetown University Library, United States of America
The Georgetown University Library migrated a collection of 150,000 image records from a legacy Cold Fusion system to the institutional repository (IR). The legacy system allowed faculty to prepare course-related documents with links to relevant images from the collection.
The institutional repository accommodated the images and image records very well. The course-specific content was not an appropriate addition to the item records in the IR. During the migration, the university Learning Management System (LMS) was targeted as the home for course-related links to the image item records. This plan leveraged an existing enterprise solution for course-related content.
The Georgetown University Scholarly Systems team build an extension to Blackboard to facilitate the creation of links to content in the IR. A "DSpace mash-up tool" was created allowing faculty to search and select items for inclusion into a course document.
This project drove the development of a number of useful features within Georgetown's IR that have proven useful for other new collections. The LMS/IR integration had an additional benefit of raising the profile of the IR for faculty and students.
Rave: Repository Enlighten-ment: A Ten Year Rave!
University of Glasgow, United Kingdom
This repository rave will celebrate the success and the development of the University of Glasgow's institutional repository service, Enlighten over the last 10 years. 10 years which have seen the repository become a key part of the University's online presence, new roles develop for staff, new opportunities emerge for the library to work with the wider university community and new international conferences like Open Repositories to let us share our experiences.
During these 10 years, the repository has also evolved to support the needs of funders and national assessment exercises (UK REF2014) and been nimble enough to implement new trends like altmetrics.
It will look back at early decisions which shaped Enlighten's initial development and enabled it to become an embedded institutional repository service. It will then look forward to its ongoing development as it becomes a suite of linked repositories supporting not only Open Access, theses and research publications but also research data, digital collections, dissertations and even Tweets.
It will draw out lessons for its past success and caveats for its future challenges.
Rave: Repositories and ETDs - a success story from Finland
National Library of Finland, Finland
Electronic theses and dissertations (ETDs) are one of the major content types in the open repositories worldwide. Although green Open Access and research data have been receiving more widespread attention in recent years, many repositories have actually been far more successful with ETDs.
The work on ETDs started out in many institutions in the late 1990s, before the development of the current repository software platforms. In a way this has been a parallel movement happening at the same time with the article-oriented Open Access movement, although many of the same people have been involved in both fields.
In Finland theses and dissertations currently account for more than two thirds of all repository content. The adoption of repository software platforms for the management of these materials has been a success story, which has had a large influence on the public visibility of these materials and the processes associated with them.
|Date: Thursday, 11/Jun/2015|
|9:00am - 10:00am||PLN3: Closing Plenary Session|
|10:15am - 10:45am||DuraSpace: DuraSpace Plenary|
Session Chair: Debra A. Hanken Kurtz