Recent Stories

The 2011 Digital Library Federation (DLF) Forum is seeking proposals for presentations, panel discussions, workshops, research updates, and hands-on, problem-solving sessions. The Forum is a working meeting where DLF members come together to do better work through sharing and collaboration. Participation is open to all those interested in contributing to and playing an active part in the successful future of digital libraries, museums and archives services and collections.

Managing digital content from cradle to grave is a complex challenge for library, museum, and archives operations. It requires creative and collaborative approaches. In that spirit, and to maximize the Forum’s benefit and better facilitate the community’s work, the Forum’s schedule will provide many opportunities to actively engage and network.

For the 2011 Fall Forum, the Program Planning Committee is requesting proposals within the broad framework of digital collections and their effect on libraries, museums and archives services, infrastructure, resources, and organizational priorities. The topics under consideration are:

  • Data management and the roles institutional services play
  • Data management policies
  • Relationship between and possible convergence of digital humanities and digital libraries
  • Open-source tools
  • Preservation of born digital content
  • Examination of digital library systems
  • Linked Open Data: continuing the LOD-LAM conversation
  • The Digital Public Library of America effort and other large-scale digital libraries
  • Digital libraries and personal digital archives
  • Strategic collaboration bringing libraries, museums and other cultural heritage institutions together

We welcome proposals on these and other topics from current community members and non-members who are interested in joining the DLF community.

Session genres include:

  • Presentations and Panels: Traditional lecture format with question-and-answer sessions.
  • Workshops:
 In-depth, hands-on training about a tool, technique, workflow, etc. You can recommend a topic or trainer, or you can volunteer to share your own expertise.
  • Research Updates:
 An opportunity for those working in digital collections research efforts to present their preliminary findings for community feedback and discussion.
  • Working Sessions:
 Creative problem solvers, including project managers, developers, and/or administrators, gather to address a specific problem. This does not have to be a computational problem. The approach can be applied to workflow issues, metadata transformations, or other complex problems that would benefit from a collective, dynamic solution approach.
  • Demos: Variation on a poster session or lightning talks. Presenters will demonstrate tools or services they have developed or are using in their digital library environment.

Proposal Submission Guidelines and Evaluation Procedures

Complete proposals should be submitted using the online submission form by  July 22, 2011 
July 29, 2011. Proposals must include a title, session leader, session genre, proposal description (maximum 400 words), and proposal abstract (maximum 100 words). After an initial review by the Program Planning Committee, all proposals will be posted on the DLF website for community polling. Those submitting complete proposals will be notified of their status by September 9, 2011. Presenters will be guaranteed a registration place.

The 2011 Fall Forum will be held at the Hyatt Regency Hotel in Baltimore, MD October 31–November 2, 2011. More information about the 2011 Fall Forum can be found at http://www.diglib.org/2011forum/.

Brian on 15 June 2011 / Comments Off

 

From a concept eleven years ago to practice in 2005, the DLF/OCLC Registry of Digital Masters (RDM) was developed in a joint project with the Digital Library Federation (DLF) and OCLC for two main reasons:  To provide the coordination and organization of the digitizing of print materials between institutions and to prevent the unnecessary duplication when digitizing those materials.  While the ramp up to add bibliographic records to the Registry was initially slow, the Registry presently contains 4.5 million records with the participation of the HathiTrust Digital Library, Biodiversity Heritage Library, Cornell University, and other institutions.

By recording materials in the Registry, institutions signal their intent to preserve and maintain the accessibility of the described materials over an extended timeframe.

A Registry record also provides information about whether a specific item has already been digitized, and if so, whether the digitization has been done at an adequate level such that another digital copy is not required, what institution is responsible for the digitization, what institution is responsible for the preservation of the digital content, and what specific materials are available.

The past six years has seen many changes in the digital collections field. In response, OCLC and DLF have decided to take the time to evaluate the Registry and determine if it meets the needs of institutions as originally conceived, if those needs have changed in the interim, or what may be needed for future growth.

DLF and OCLC are reviewing the RDM’s success, through a survey and other venues, to determine if any changes are needed to meet the needs of the current and future users.

If you are interested in contributing to the RDM review, please send an email to the DLF info.

To keep updated on RDM activities and announcements, subscribe to its newsfeed.

Resources

RDM: Historical Compendium (PDF)

Use Assessment Survey: 2011

Public access to the Registry of Digital Masters

RDM: Record Creation Guidelines

 

 

Rachel on 4 May 2011 / Comments Off

Overview

The 2012 DLF Forum was held in Denver, Colorado, Sunday, November 4th – Monday, November 5th, 2012, with pre- and post-conference events on November 2nd, November 3rd, and November 6th. All sessions and events took place at the Westin Denver Downtown.

Forum participation is open to all who are interested in contributing to and playing an active part in the successful future of digital libraries, museum and archives services, and collections. The Forum featured presentations and panels, workshops, research updates, working sessions, demos, and more. Please visit our Schedule page for details.

To receive the most recent news about the forum, subscribe to the Forum Newsfeed.

Search for the hash tag #dlfforum to view tweets about the DLF Forum!

The 2012 DLF Forum is sponsored in part by…

 

Recent Forum News

  • 2013 DLF Forum Call for Proposals
    29 April 2013

    The 2013 Digital Library Federation (DLF) Forum is seeking proposals for presentations, panel discussions, workshops, research updates, and hands-on, problem-solving sessions. The Digital Library Federation is a robust and diverse community of practitioners who advance research, teaching, and learning through the application of digital library research, technology, and services. The Forum is a working meeting where ...

  • Sponsor the 2013 DLF Forum
    15 April 2013

    DLF has a number of sponsorship opportunities available to help promote your institution or corporate brand. Available opportunities support the annual Forum, which fosters collaboration across all areas of digital library operations and brings together those serving in the field. More information can be found here.

  • The 2014 DLF Forum will be in Atlanta!
    22 March 2013

    Save the date: Next year’s DLF Forum will be held October 27-29 at the Georgia Tech Hotel & Conference Center. Due to the closing of the Indianapolis Conference center, we chose Atlanta, as it had a strong showing in previous location votes. Details for the 2013 DLF Forum will be posted to the website as the planning ...

  • Announcing the 2013 DLF Forum Program Planning Committee
    20 March 2013

    We would like to thank everyone who volunteered to be a part of the 2013 DLF Forum Program Planning Committee (PPC). We had an incredible number of responses! The PPC slate was determined by several factors like including a mix of veterans and new members to represent the DLF community. There will be other opportunities ...

  • Call for 2013 DLF Forum Program Planning Committee Members
    7 February 2013

    The Digital Library Federation would like to invite members of our community, and those involved in the greater digital library community, to participate in the planning of the 2013 DLF Forum. Being part of the Program Planning Committee (PPC) will involve participation in bi-weekly planning phone calls beginning seven months before the Forum, as well ...

Rachel on 2 May 2011 / 1 Comment

Planning for the 2011 DLF Forum has begun, and we are creating opportunities to open up the process to the whole DLF community.

The DLF Program Committee seeks open community input on what you would like to hear about at the 2011 Forum Meeting in Baltimore. An open document has been created for the purpose of collecting ideas.

Please submit your suggestions for topics, formats, speakers, or other activities by Friday, May 20th.

Results will be announced soon after, which will inform a call for proposals.

For the most recent news about the 2011 Forum, you can subscribe to the Forum News Feed, or go to the 2011 DLF Forum web page.

The Program Planning Committee (PCC)  will using the web site and news feed to post updates and information about the Forum’s venue, registration, meeting logistics, and, most importantly, how to participate in the planning process.

Thanks for your time and contributions!

The 2011 Forum Program Planning Committee
Charles Blair
Declan Fleming
Rachel Frick
Kevin Hawkins
Delphine Khanna
Bethany Nowviskie
David Reynolds
Jenn Riley
Sarah Shreeves
Joan Starr
Jennifer Vinopal
Melissa Wisner

 

Rachel on 2 May 2011 / 4 Comments

The DLF community uses a number of communication channels to keep in touch.

 

Currently there are two active community listserves:

DLF-Announce, for general community news and announcements

DLF-PM, a dedicated list of the Project Managers interest group.

 

DLF community members can join the DLF LinkedIn group, Facebook group,and to follow us on Twitter.

 

News updates are available through an RSS feed on the whole website, or just updates to news items of a particular topic.
For the whole website, subscribe to : http://www.diglib.org/feed/

For a specific news stream, simply add “feed” to the end of the URL for the particular topic.
For example, for a newsfeed specific to the DLF Forums, subscribe to:  http://www.diglib.org/topics/forum/feed/

Rachel on 5 April 2011 / Comments Off

Bibliographic Indeterminacy and the Scale of Problems and Opportunities of “Rights” in Digital Collection Building

by John P. Wilkin

February 2011

PDF version

The research library community has little strong or reliable data on the number of unique books in our collections and their “rights”—for example, whether they are in the public domain or in-copyright and, if in-copyright, whether they are orphan works. At its foundation, this problem is created by the dearth of reliable bibliographic information, or what I’ve been calling bibliographic indeterminacy. For example, we’d like to know how large the “collective collection” of all (or even just all North American) research libraries is, and how many unique volumes research libraries hold in aggregate; otherwise, there’s no way to know the cost of digitizing or caring for these materials. We’d also like to have a better handle on the question of what’s in the public domain and, by extension, what’s in copyright. We’d like to know how many orphan works there are, or perhaps what proportion of the digitized content we have online is likely to be orphans. And while these questions and more are regularly part of the conversation around digital collection building, they’re also relevant to more conventional library problems such as print storage and particularly shared print storage. We don’t know what’s in the collective collection.

The fact is, we have little reliable data about most of these questions. There’s been considerable speculation in the wake of the proposed Google Books settlement and even years before, when we first considered the probable shape of the growing digital collection or the opportunities in front of us. Our biggest impediment to getting a good bearing on questions of the size, nature and rights status of research library collections is the simple lack of an authoritative bibliography.

Efforts to Date

To answer these questions, we often turn to WorldCat, but its records are overwhelmed by the noise in WorldCat: the high number of unique records that represent variations in cataloging rather than separate manifestations of a work, non-book and non-journal material masquerading as books and journals, and items with incomplete or unreliable metadata. As a database, WorldCat is by far the best thing we have, but its purposes long ago shifted away from documenting the collective collection to facilitating discovery (as a data source for WorldCat.org). Brian Lavoie and Lorcan Dempsey worked through those challenges with admirable adroitness in their “Beyond 1923: Characteristics of Potentially In-Copyright Print Books in Library Collections,” providing the best picture of post-1923 book publishing. Still, their analysis is just as certainly hampered by the chaos of the WorldCat database. And while extraordinarily helpful, much of the focus of Lavoie’s and Dempsey’s work is on the aggregate database (i.e., everything that has been cataloged) and then to a limited extent on a few Google digitization partners. Dempsey’s “Libraries and the Long Tail: Some Thoughts About Libraries in the Network Age,” which provides a picture of the shape of the collections of the first Google partners, is also worthy of note.

One of the best pieces of analysis on the likely body of orphan works is “580,388 Orphan Works — Give or Take” by Michael Cairns. Cairns does the best he can with the available data, relying significantly on publishing statistics as well as Lavoie’s and Dempsey’s analysis. The focus on publishing statistics highlights the fundamental problem caused by a lack of empirical data. Cairns relies on Bowker’s publishing data when, in fact, libraries buy many works that are never described in these types of sources. It’s likely that a sizable body of gray literature and even some scholarly literature (e.g., some monographs in series) skews the numbers and would create many opportunities for opening access to content. Moreover, because of the informal nature of the publishing process for these works, many more of them may be orphans. The numbers are indeed hard to pin down. For example, in another study, “In From the Cold: An Assessment of the Scope of ‘Orphan Works’ and its Impact on the Delivery of Services to the Public,” the Joint Information Systems Committee (JISC) estimated 503 UK institutions could hold in excess of 50 million orphan works.

New Insights Through HathiTrust

Over the past two years, HathiTrust, a partnership of major research libraries working together to ensure that the cultural record is preserved and accessible long into the future, has built a large and representative body of materials that gives us a much more reliable empirical window into a number of questions around books. By the end of October 2010, the collection contained digitized versions of slightly more than 5 million monographic volumes. Work by Constance Malpas, Roy Tennant, and others in RLG Research has demonstrated that the composition of the HathiTrust collection is remarkably representative of research library collections. Their data, much of it published in “Cloud-sourcing Research Collections: Managing Print in the Mass-digitized Library Environment,” by Constance Malpas, shows that the HathiTrust collection holds a growing percentage of titles that are also held by ARL libraries:  the median rate of overlap between HathiTrust and an ARL library was 19% in June 2009, 31% in June 2010,  and 33% in December 2010. The rate of overlap is fairly consistent across all ARLs, and grows in a fairly constant way (see Figure 1: Overlap between HathiTrust and ARL libraries). The composition of the collection, too, shows strong signs of representativeness. The HathiTrust collection contains more than 400 languages and, like so many ARLs, slightly fewer than 50% of the volumes are in English; as the collection grows, many bibliographic characteristics (e.g., language, period, subject) hold fairly constant. This large and representative collection, then, may hold the key to understanding the general parameters of some of the problems facing us.

Fig. 1: Overlap between HathiTrust and ARL libraries

Assuming the HathiTrust collection is representative or indicative, we’ve started to analyze it for characteristics to help us better understand the scope of the public domain, orphan works, and copyright challenges. Before beginning, I’d like to offer a frank apology about the US-centric analysis that follows. US copyright law affords us a relatively clear framework in which to understand these problems. The challenges I’ll identify are not specific to readers in the United States, though the US-specific analysis also helps us understand the problems for readers in other countries as well.

Distribution by Date

The first and most basic piece of analysis identifies how the collection breaks out according to boundaries of US copyright determination in the United States. Understanding the publishing patterns in relation to the major markers in US legislation helps clarify some of the issues that we should address. Specifically, we want to have a clear sense of how the corpus breaks down in the following regard:  works published before 1923, those published between 1923 and 1963, those published between 1964 and 1977, and those published after 1977. For US law and US users, we know that something approximating the following is true:

  1. All works published before 1923 can be treated as public domain for a US audience.
  2. US copyright law required a copyright notice and copyright renewal for US works published between 1923 and 1963.
  3. US copyright law required only copyright notice for US works published between 1964 and 1977. (Actually, works published until 1 March 1989 are in the public domain if published without notice and without subsequent registration within 5 years.)
  4. If the work was created after 1977 and published with notice, the work was afforded copyright protection for the life of the author plus 70 years. Thus, nearly all works created after 1977 will be given copyright protection for decades to come.

There is considerable nuance and some tricky exceptions to all of these rules, which I won’t try to supply here. Peter Hirtle’s “Copyright Term and the Public Domain in the United States” and other sources provide a fuller picture.

As shown in Figure 2, the distribution along these dates helps refine our sense of the certain and likely public domain. Currently, 21% of the HathiTrust book corpus was published before 1923, and another 21% was published between 1923 and 1963. These numbers both mirror and deviate from the Lavoie and Dempsey numbers based on the Google digitization partners:  their numbers for pre-1923 were a lower 15%, though the 1923-1963 numbers were a similar 20%. The higher HathiTrust pre-1923 numbers might be explained by the focus of some partners on digitizing public domain works; nevertheless, most of the works digitized are from Michigan and California, both of which have digitized more comprehensively. (About 60% of Michigan’s print collections are currently online in HathiTrust.)

Fig. 2: Breakdown of HathiTrust book corpus by publication date

Distribution of the Corpus by US and Non-US Publication

Whether a work was published in the United States also has a bearing on its copyright status, specifically for US users. For the periods 1923-1963 and 1964-1977, a work published in the United States is subject to different rules of copyright status interpretation than works published outside the United States.[1] Though we might expect significant variation in the distribution of US versus non-US published work over the years, if only because of the relative growth of US publishing over this vast span of time, it’s remarkably uniform in the HathiTrust collection. Applied to each of the four periods, the breakdown is as follows:

Fig. 3: HathiTrust book corpus: US vs. non-US-published holdings

Copyright Status Determination, Pre-1923 and 1923-1963

For the sake of this discussion, we will assume all pre-1923 books are in the public domain. This is, of course, an oversimplification and a very US-centric perspective, but I’d like to posit this for the sake of clarity in representing these numbers. The copyright status of books published in the United States between 1923-1963 cannot be assumed, and must be determined for each individual work. The University of Michigan was awarded a grant by the Institute for Museum and Library Services (IMLS) to undertake large-scale and systematic work to determine the copyright status of works published in this period. Over the last two years, Michigan, in collaboration with several other partners, has amassed a large and compelling picture of the likelihood of a US work published in this period being in the public domain. Month after month, regardless of the source institution or the collection being digitized, the Copyright Review Management System (CRMS) staff find 55% of the 1923-1963 works in the corpus to be in the public domain, either because those works never received copyright protection when they were published, or because their initial copyright was not renewed. Mind you, this is with more than 100,000 titles having been reviewed, not some insignificant and skewed sample. We confirmed the reliability of this work by asking the Library of Congress Copyright Office to analyze a random sample of our determinations. Most of the works in the remaining 45% are in copyright, though in some cases staff could not make a determination without more data. Consequently, we have a well-defined picture of the copyright status of works first published in US during this period and found in research libraries:

Fig. 4:  HathiTrust book corpus: Copyright status of books published pre-1923 and US works published 1923-1963

Moving From “Certainty” to “Speculation”:  1923-1963

The data presented thus far come with a high degree of confidence, as they are based on large numbers of volumes or, in the case of the CRMS work, tens of thousands of determinations. Now, however, we enter the realm of speculation. Some of this speculation makes assumptions grounded in work that has been done elsewhere. For example, Carnegie Mellon University’s project to secure rights for contemporary publications was unable to reach many rights holders to gain permission to use a work, and this was more likely to be the case for older works than for more recent works (Covey, “Acquiring Copyright Permission to Digitize and Provide Open Access to Books“). As Covey notes, “We could not find the publishers of most of the books published between 1920 and 1930 and of almost half of the books published between 1940 and 1950. Publishers of more than a third of the books published from 1950 to 1960 and 1960 to 1970 could not be found” (p. 19). Moreover, when a rights holder could be identified and an attempt was made to contact them, the CMU project received no response from 30-40% of the identifiable rights holders for works published during the periods 1930–1940 and 1970–1990; and no response from 20-30% of the rights holders from works published between 1940 and 1970.

Based on the experience of Carnegie Mellon University, let’s hypothesize the following, recognizing that more data are needed for each characterization:

  1. For non-US works published between 1923-1963, roughly 20% will be in the public domain (e.g., because the author died before 1941, as would be the case for determining public domain status for works published in countries like the US that has a term of life plus 70 years). I want to be clear that I have no basis for this assertion of how many authors died between 1923 and 1941—it’s a wild guess.
  2. For all works (i.e., both US and non-US works) published between 1923-1963, we will be able to contact only 10% of the authors, publishers, or heirs who hold rights.
  3. The remaining works published between 1923 and 1963, both US and non-US, (i.e., 35% in the United States and 70% outside of the United States) are “orphan works,” i.e., works in copyright where no rights holder can be identified or contacted.

Fig. 5:  HathiTrust book corpus: Public domain, in-copyright, and orphan works, pre-1923 and 1923-1963

Speculation Amplified:  1964-1977

Through HathiTrust we’ve made a considerable effort to get a bearing on the public domain opportunity for US works published between 1923 and 1963, but we have absolutely no data on the copyright status of works published between 1964 and 1977. In this later period, rights holders were required only to affix a copyright notice on published works to secure protection; moreover, if they included a copyright notice, they were not required to renew. Rights holders between 1964-1977 were undoubtedly more likely to be aware of the requirements for copyright protection than rights holders in previous periods (if only because of the increased public and legislative attention to copyright), and the lack of an additional renewal requirement probably also means that more volumes are in-copyright. This means that the percentage of volumes in the public domain is likely to be much smaller. We can also speculate that because the materials are closer to being contemporary, we are more likely to be able to locate the rights holder than would be the case for older materials. The CMU data again note that “Publishers of more than a third of the books published from 1950 to 1960 and 1960 to 1970 could not be found,” and response rates for those who could be located were 20-30% (p. 19).

My numbers for copyright status in the 1964-1977 period are just guesses. This is based on very little data because we have very little data to guide our speculation. Bear with me as I posit the following:

  1. 20% of US works published between 1964 and 1977 will be in the public domain.
  2. With very few exceptions, no works published outside of the United States will be in the public domain.
  3. Compared with the period 1923 and 1963, we are twice as likely to be able to identify and successfully contact authors, publishers, or heirs who hold rights for those works in copyright (i.e., 20%).
  4. The remaining works, both US and non-US, are “orphan works”—works in copyright where no rights holder can be identified or contacted (i.e., 60% of US works and 80% of non-US works).

Fig. 6:  HathiTrust book corpus: Breakdown by US/non-US and rights status, pre-1923, 1923-1963 and 1964-1977

Guesses, Pure Guesses: 1978 to the Present

If we had guesses and informed speculation for the periods before the present, we’re clearly working without a net for 1978 to the present. There are copyright wrinkles here (e.g,. some governmental publications, both US and otherwise, and occasional cases where a work may be ineligible for copyright protection or is dedicated to the public domain), but in general we can say that everything published is in copyright. In most cases, the work will enter the public domain 70 years after the death of the author, but there are many cases where the period of coverage is longer (e.g., 95 years for some works of corporate authorship, a number that may actually be lower or higher depending on circumstances). Covey notes that most of the publishers for works in this period could be located, but only 30-40% responded to inquiries (p. 19).[2] For discussion only, let’s assume that our ability to successfully contact authors, publishers, or heirs that hold rights will double again, reaching 40% of the works.

Fig. 7:  HathiTrust book corpus: Breakdown by US/non-US and rights status for all periods

Conclusion

Our data spotlight the likely scope of the public domain and the probable large role of orphans in our bibliographic landscape. The following are some key findings of our preliminary analysis:

  1. The percentage of public domain books in the collective collection—not simply the current 5+ million books, but the collection as it expands—is unlikely to grow to more than 33% of the total number of books we will put online. Using the numbers assembled here, the percentage of public domain materials, not including government documents, will be 28%.
  2. The body of orphan works—works whose rights holders we cannot locate—is likely to be extremely large, and perhaps the largest body of materials. If the guesses made here are right, 50% of the volumes will be orphan works. This 50% is comprised as follows:  12.6% will come from the years 1923-1963, 13.6% from 1964-1977, and 23.8% from 1978 and years that follow. (The percentage of orphan works relative to all works decreases as time passes; the number of orphan works increases in more recent years because more works are published in later years.) Indeed, if this speculation is right, our incomplete collection today includes more 2.5 million orphan works, of which more than 800,000 are US orphans.
  3. The likely size of the corpus of in-copyright publications for which we are able to identify a known rights holder will be roughly the same size as, or slightly smaller than, the body of public domain materials. Again, using these speculative numbers, they may comprise as little as 22% of the total number of books.

Even before we are finished digitizing our collections, the potential numbers are significant and surprising: more than 800,000 US orphans and nearly 2 million non-US orphans.

There are two important conclusions to draw from this preliminary analysis. The first and most obvious is that we still need better data to understand the extent of the problems and opportunities. In the coming years, HathiTrust and its partners hope to gather more data on orphan works in various periods, and on the extent of the public domain in works published outside the United States. Making serious progress on the matter of orphan works, however, will probably depend on a policy framework that allows us to make use of those volumes. Nevertheless—and this is critically important for those who wish to see reasonable uses made of digitized book content—most of the publications we hold in our collections and put online are likely to be those we would consider orphan works, with no clearly identifiable or contactable rights holder. In nearly all cases, there is no economic harm to any person or organization in opening access to these in-copyright works, and there is a great loss in not providing access to them. Without an effective legal or policy framework that allows us to do so, a significant portion of our cultural heritage will be underused and undervalued.

Notes

[1] Although we attempt to segregate US works that may have also been published abroad in our automatic rights determination process, as in our copyright review process, the numbers here make no attempt to take simultaneous publication into account.

[2] I include these numbers about a lack of response because of their possible bearing on the absence of copyright holders. Still, it should be noted, if a rights holder does not respond, it does not mean the rights holder does not exist.

Acknowledgments

Special thanks to Suzanne Chapman for the handsome graphics. Many people helped me clarify the points I’m making here. Friends with copyright knowledge, including Jack Bernard, Peter Hirtle, Melissa Levine and Anne Karle-Zenith were kind enough to set me straight on some points. Other friends immersed in these problems, including Constance Malpas, Jeremy York and Lynne Raughley, were generous with their feedback and corrections.

About the Author

John P. Wilkin is executive director of HathiTrust and associate university librarian for library information technology (LIT) for the University of Michigan. The LIT Division supports the library’s online catalog and related technologies, provides the infrastructure to both digitize and access digital library collections, supports the library’s web presence, and provides frameworks and systems to coordinate Library technology activities. Wilkin previously served as the head of the Digital Library Production Service at the University of Michigan. Among the units in the DLPS is the University of Michigan’s Humanities Text Initiative, an organization responsible for SGML document creation and online systems that Wilkin founded in 1994. He earned graduate degrees in English from the University of Virginia (1980) and Library Science from the University of Tennessee at Knoxville (1986). In 1992, he worked at the University of Virginia as the Systems Librarian for Information Services, where he shaped the Library’s plan for establishing a group of electronic centers and consulted for the University’s Institute for Advanced Technology in the Humanities (IATH) in textual issues.

HathiTrust, a collaborative of more than 50 research libraries worldwide, is devoted to ensuring preservation of and access to the published cultural record. HathiTrust is working on integration of publishing mechanisms into the preservation repository and is working with several university presses to open both their backlists and new publications in HathiTrust. Its robust digital preservation architecture is fully funded by partner contributions.

Rachel on 3 March 2011 / Comment

CLIR/DLF announce a new publication series, Ruminations.

The series will feature short research papers and essays that bring a new perspective to issues relevant to planning for and managing organizational and institutional change in the evolving digital environment for scholarship and teaching.

We inaugurate the new series with a report by John P. Wilkin that posits the scope of works in the public domain and probable extent of orphan works in our research library collections, based on an analysis of the HathiTrust book corpus.

Wilkin is executive director of HathiTrust and is associate university librarian for library information technology for the University of Michigan. HathiTrust, a collaborative of more than 50 research libraries worldwide, is devoted to ensuring preservation of and access to the published cultural record.

Let us know what you think!

Give us your feedback in the comment space following the report.

Rachel on 24 February 2011 / Comments Off

ARL

CNI

    D-Lib Magazine

    OCLC Research

    Planet Code4Lib

    Notes from the Field

      CLIR/DLF Tweets

      Eric (admin) on 23 February 2011 / Comments Off

      DLF Mailing ListRSS FeedTwitterFacebookLinkedInYouTubeZotero

      Eric (admin) on 23 February 2011 / Comments Off

      CURATEcamp is a series of unconference-style events focused on connecting practitioners and technologists interested in digital curation. DLF is privileged to contribute support to this dynamic enterprise. More information can be found at their website, at their Google group, and via Twitter.

       

      Recent CURATEcamp News and Announcements:

        Rachel on 16 February 2011 / Comments Off