An ongoing multiple volume set with updated index.
Publisher
West Group
Publication Location
St. Paul, MN
Critical Arguements
CA "Contains over 400 separate titles on a broad range of legal topics which, taken together, systematically describe the entire field of American legal doctrine. Documents available for each topic may include a summary, topic contents, each (TRUNCATED)
There are a large number of metadata standards and initiatives that have relevance to digital preservation, e.g. those designed to support the work of national and research libraries, archives and digitization initiatives. This paper introduces some of these, noting that the developers of some have acknowledged the importance of maintaining or re-using existing metadata. It is argued here that the implementation of metadata registries as part of a digital preservation system may assist repositories in enabling the management and re-use of this metadata and may also help interoperability, namely the exchange of metadata and information packages between repositories.
Publisher
2003 Dublin Core Conference: Supporting Communities of Discourse and Practice-Metadata Research & Applications
Publication Location
Seatle, WA
Critical Arguements
CA "This paper will introduce a range of preservation metadata initiatives including the influential Open Archival Information System (OAIS) reference model and a number of other initiatives originating from national and research libraries, digitization projects and the archives community. It will then comment on the need for interoperability between these specifications and propose that the implementation of metadata registries as part of a digital preservation system may help repositories manage diverse metadata and facilitate the exchange of metadata or information packages between repositories."
Conclusions
RQ "The plethora of metadata standards and formats that have been developed to support the management and preservation of digital objects leaves us with several questions about interoperability. For example, will repositories be able to cope with the wide range of standards and formats that exist? Will they be able to transfer metadata or information packages containing metadata to other repositories? Will they be able to make use of the 'recombinant potential' of existing metadata?" ... "A great deal of work needs to be done before this registry-based approach can be proved to be useful. While it would undoubtedly be useful to have registries of the main metadata standards developed to support preservation, it is less clear how mapping-based conversions between them would work in practice. Metadata specifications are based on a range of different models and conversions often lead to data loss. Also, much more consideration needs to be given to the practical issues of implementation." 
SOW
DC Michael Day is a research officer at UKOLN, which is based at the University of Bath. He belongs to UKOLN's research and development team, and works primarily on projects concerning metadata, interoperability and digital preservation. 
Type
Conference Proceedings
Title
Preserving the Fabric of Our Lives: A Survey of Web Preservation Initiatives
This paper argues that the growing importance of the World Wide Web means that Web sites are key candidates for digital preservation. After an [sic] brief outline of some of the main reasons why the preservation of Web sites can be problematic, a review of selected Web archiving initiatives shows that most current initiatives are based on combinations of three main approaches: automatic harvesting, selection and deposit. The paper ends with a discussion of issues relating to collection and access policies, software, costs and preservation.
Secondary Title
Research and Advanced Technology for Digital Libraries, 7th European Conference, ECDL 2003, Trondheim, Norway, August 2003 Proceedings
Publisher
Springer
Publication Location
Berlin
Critical Arguements
CA "UKOLN undertook a survey of existing Web archiving initiatives as part of a feasibility study carried out for the Joint Information Systems Committee (JISC) of the UK further and higher education funding councils and the Library of the Wellcome Trust. After a brief description of some of the main problems with collecting and preserving the Web, this paper outlines the key findings of this survey." (p. 462) Addresses technical, legal and organizational challenges to archiving the World Wide Web. Surveys major attempts that have been undertaken to archive the Web, highlights the advantages and disadvantages of each, and discusses problems that remain to be addressed.
Conclusions
RQ "It is hoped that this short review of existing Web archiving initiatives has demonstrated that collecting and preserving Web sites is an interesting area of research and development that has now begun to move into a more practical implementation phase. To date, there have been three main approaches to collection, characterised in this report as 'automatic harvesting,' 'selection' and 'deposit.' Which one of these has been implemented has normally depended upon the exact purpose of the archive and the resources available. Naturally, there are some overlaps between these approaches but the current consensus is that a combination of them will enable their relative strengths to be utilised. The longer-term preservation issues of Web archiving have been explored in less detail." (p. 470)
SOW
DC OAIS emerged out of an initiative spearheaded by NASA's Consultative Committee for Space Data Systems. It has been shaped and promoted by the RLG and OCLC. Several international projects have played key roles in shaping the OAIS model and adapting it for use in libraries, archives and research repositories. OAIS-modeled repositories include the CEDARS Project, Harvard's Digital Repository, Koninklijke Bibliotheek (KB), the Library of Congress' Archival Information Package for audiovisual materials, MIT's D-Space, OCLC's Digital Archive and TERM: the Texas Email Repository Model.
CA "Ironically, electronic records systems make it both possible to more fully capture provenance than paper recrods systems did and at the same time make it more likely that provenance will be lost and that archives, even if they are preserved, will therefore lack evidential value. This paper explores the relationship between provenance and evidence and its implications for management of paper or electronic information systems." (p. 177)
Conclusions
"Electronic information systems, therefore, present at least two challenges to archivists. The first is that the designers of these systems may have chosen to document less contextual information than may be of interest to archivists when they designed the system. The second is that the data recorded in any given information system will, someday, need to be transferred to another system. ... [A]rchivists will need to return to fundamental archival principles to determine just what they really wanted to save anyway. ... It may be that archivists will be satisfied with the degree of evidential historicity they were able to achieve in paper based record systems, in which case there are very few barriers to implementing successful electronic based archival environments. Or archivists may decide that the fuller capability of tracking the actual participation of electronic data objects in organizational activities needs to be documented by archivally satisfactory information systems, in which case they will need to define those levels of evidential historicity that must be attained, and specify the systems requirements for such environments. ... At a meeting on electronic records management research issues sponsored by the National Historical Publications and Records Commission in January 1991, participants identified the concept of technological and economic plateaux in electronic data capture and archiving as an important arena for research ... Hopefully this research will produce information to help archivists make decisions regarding the amount of contextual information they can afford to capture and the requirements of systems designed to document context along with managing data content. ... I will not be surprised as we refine our concepts of evidential historicity to discover that the concept of provenance takes on even greater granularity." (p. 192-193)
CA Discusses the ways traditional archival science can inform IT, and the ways IT can help the goals of archival science be achieved more easily and efficiently.
Conclusions
<RQ> "When archivists work with information technologies or electronic archiving specialists, they have a lot to offer. They are the ones who have the conceptual key to the analysis and design of the new archiving systems." (p. 174)
This study focuses upon access to authentic electronic records that are no longer required in day-to-day operations and that have been set aside in a recordkeeping system or storage repository for future reference. One school of thought, generally associated with computer information technology specialists, holds that long-term access to electronic records is primarily a technological issue with little attention devoted to authenticity. Another school of thought, associated generally with librarians, archivists, and records managers, contends that long-term access to electronic records is as much an intellectual issue as it is a technological issue. This latter position is clearly evident in several recent research projects and studies about electronic records whose findings illuminate the discussion of long-term access to electronic records. Therefore, a review of eight research projects highlighting findings relevant for long-term access to electronic records begins this chapter. This review is followed by a discussion, from the perspective of archival science, of nine questions that a long-term access strategy must take into account. The nine issues are: What is a document?; What is a record?; What are authentic electronic records?; What does "archiving" mean?; What is an authentic reformatted electronic record?; What is a copy of an authentic electronic record?; What is an authentic converted electronic record?; What is involved in the migration of authentic electronic records?; What is technology obsolescence?
Book Title
Authentic Electronic Records: Strategies for Long-Term Access
Publisher
Cohasset Associates, Inc.
Publication Location
Chicago
ISBN
0970064004
Critical Arguements
CA "Building upon the key concepts and concerns articulated by the studies described above, this report attempts to move the discussion of long-term access to electronic records towarad more clearly identified, generally applicable and redily im(TRUNCATED)
Conclusions
RQ
SOW
DC This book chapter was written by Charles M. Dollar for Cohasset Associates, Inc. Mr. Dollar has "twenty-five years of experience in working with electronic records as a manager at the National Archives and Records Administration, as an archival educator at the University of British Columbia, and a consultant to governments and businesses in North America, Asia, Europe, and the Middle East." Cohasset Associates Inc. is "one of the nation's foremost consulting firms specializing in document-based information management."
CA Makes a distinction between archival description of the record at hand and documentation of the context of its creation. Argues the importance of the latter in establishing the evidentiary value of records, and criticizes ISAD(G) for its failure to account for context. "(1) The subject of documentation is, first and foremost, the activity that generated the records, the organizations and individuals who used the records, and the purposes to which the records were put. (2). The content of the documentation must support requirements for the archival management of records, and the representations of data should support life cycle management of records. (3) The requirements of users of archives, especially their personal methods of inquiry, should determine the data values in documentation systems and guide archivists in presenting abstract models of their systems to users." (p. 45-46)
Phrases
<P1> [T]he ICA Principles rationalize existing practice -- which the author believes as a practical matter we cannot afford; which fail to provide direct access for most archives users; and which do not support the day-to-day information requirements of archivists themselves. These alternatives are also advanced because of three, more theoretical, differences with the ICA Principles: (1) In focusing on description rather than documentation, they overlook the most salient characteristic of archival records: their status as evidence. (2) In proposing specific content, they are informed by the bibliographic tradition rather than by concrete analysis of the way in which information is used in archives. (3) In promoting data value standardization without identifying criteria or principles by which to identify appropriate language or structural links between the objects represented by such terms, they fail adequately to recognize that the data representation rules they propose reflect only one particular, and a limiting, implementation. (p. 33-34) <P2> Archives are themselves documentation; hence I speak here of "documenting documentation" as a process the objective of which is to construct a value-added representation of archives, by means of strategic information capture and recording into carefully structured data and information access systems, as a mechanism to satisfy the information needs of users including archivists. Documentation principles lead to methods and practices which involve archivists at the point, and often at the time, of records creation. In contrast, archival description, as described in the ICA Principles[,] is "concerned with the formal process of description after the archival material has been arranged and the units or entities to be described have been determined." (1.7) I believe documentation principles will be more effective, more efficient and provide archivists with a higher stature in their organizations than the post accessioning description principles proposed by the ICA. <warrant> (p. 34) <P3> In the United States, in any case, there is still no truly theoretical formulation of archival description principles that enjoys a widespread adherence, in spite of the acceptance of rules for description in certain concrete application contexts. (p. 37) <P4> [T]he MARC-AMC format and library bibliographic practices did not adequately reflect the importance of information concerning the people, corporate bodies and functions that generated records, and the MARC Authority format did not support appropriate recording of such contexts and relations. <warrant> (p. 37) <P5> The United States National Archives, even though it had contributed to the data dictionary which led to the MARC content designation, all the data which it believed in 1983 that it would want to interchange, rejected the use of MARC two years later because it did not contain elements of information required by NARA for interchange within its own information systems. <warrant> (p. 37) <P6> [A]rchivists failed to understand then, just as the ISAD(G) standard fails to do now, that rules for content and data representation make sense in the context of the purposes of actual exchanges or implementation, not in the abstract, and that different rules or standards for end-products may derive from the same principles. (p. 38) <P7> After the Committee on Archival Information Exchange of the Society of American Archivists was confronted with proposals to adopt many different vocabularies for a variety of different data elements, a group of archivists who were deeply involved in standards and description efforts within the SAA formed an Ad Hoc Working Group on Standards for Archival Description (WGSAD) to identify what types of standards were needed in order to promote better description practices.  WSAD concluded that existing standards were especially inadequate to guide practice in documenting contexts of creation.  Since then, considerable progress has been made in developing frameworks for documentation, archival information systems architecture and user requirements analysis, which have been identified as the three legs on which the documenting documentation platform rests. <warrant> (p. 38) <P8> Documentation of organizational activity ought to begin long before records are transferred to archives, and may take place even before any records are created -- at the time records are created -- at the time when new functions are assigned to an organization. (p. 39) <P9> It is possible to identify records which will be created and their retention requirements before they are created, because their evidential value and informational content are essentially predetermined. (p. 39) <P10> Archivists can actively intervene through regulation and guidance to ensure that the data content and values depicting activities and functions are represented in such a way that will make them useful for subsequent management and retrieval of the records resulting from these activities. This information, together with systems documentation, defines the immediate information system context out of which the records were generated, in which they are stored, and from which they were retrieved during their active life. (p. 39) <P11> Documentation of the link between data content and the context of creation and use of the records is essential if records (archives or manuscripts) are to have value as evidence. (p. 39) <P12> [C]ontextual documentation capabilities can be dramatically improved by having records managers actively intervene in systems design and implementation.  The benefits of proactive documentation of the context of records creation, however, are not limited to electronic records; the National Archives of Canada has recently revised its methods of scheduling to ensure that such information about important records systems and contexts of records creation will be documented earlier. <warrant> (p. 39) <P13> Documentation of functions and of information systems can be conducted using information created by the organization in the course of its own activity, and can be used to ensure the transfer of records to archives and/or their destruction at appropriate times. It ensures that data about records which were destroyed as well as those which were preserved will be kept, and it takes advantage of the greater knowledge of records and the purposes and methods of day-to-day activity that exist closer to the events. (p. 40) <P14> The facts of processing, exhibiting, citing, publishing and otherwise managing records becomes significant for their meaning as records, which is not true of library materials. (p. 41) <P15> [C]ontent and data representation requirements ought to be derived from analysis of the uses to which such systems must be put, and should satisfy the day to day information requirements of archivists who are the primary users of archives, and of researchers using archives for primary evidential purposes. (p. 41) <P16> The ICA Commission proposes a principle by which archivists would select data content for archival descriptions, which is that "the structure and content of representations of archival material should facilitate information retrieval." (5.1) Unfortunately, it does not help us to understand how the Commission selected the twenty-five elements of information identified as its standard, or how we could apply the principle to the selection of additional data content. It does, however, serve as a prelude to the question of which principles should guide archivists in choosing data values in their representations. (p. 42) <P17> Libraries have found that subject access based on titles, tables of contents, abstracts, indexes and similar formal subject analysis by-products of publishing can support most bibliographic research, but the perspectives brought to materials by archival researchers are both more varied and likely to differ from those of the records creators. (p. 43) <P18> The user should not only be able to employ a terminology and a perspective which are natural, but also should be able to enter the system with a knowledge of the world being documented, without knowing about the world of documentation. (p. 44) <P19> Users need to be able to enter the system through the historical context of activity, construct relations in that context, and then seek avenues down into the documentation. This frees them from trying to imagine what records might have survived -- documentation assists the user to establish the non-existence of records as well as their existence -- or to fathom how archivists might have described records which did survive. (p. 44) <P20> When they departed from the practices of Brooks and Schellenberg in order to develop means for the construction of union catalogues of archival holdings, American archivists were not defining new principles, but inventing a simple experiment. After several years of experience with the new system, serious criticisms of it were being leveled by the very people who had first devised it. (p. 45)
Conclusions
RQ "In short, documentation of the three aspects of records creation contexts (activities, organizations and their functions, and information systems), together with representation of their relations, is essential to the concept of archives as evidence and is therefore a fundamental theoretical principle for documenting documentation. Documentation is a process that captures information about an activity which is relevant to locating evidence of that activity, and captures information about records that are useful to their ongoing management by the archival repository. The primary source of information is the functions and information systems giving rise to the records, and the principal activity of the archivist is the manipulation of data for reference files that create richly-linked structures among attributes of the records-generating context, and which point to the underlying evidence or record." (p. 46)
Type
Journal
Title
Archival Issues in Network Electronic Publications
"Archives are retained information systems that are developed according to professional principles to meet anticipated demands of user clienteles in the context of the changing conditions created by legal environments and electronic or digital technologies. This article addresses issues in electronic publishing, including authentication, mutability, reformatting, preservation, and standards from an archival perspective. To ensure continuing access to electronically published texts, a special emphasis is placed on policy planning in the development and implementation of electronic systems" (p.701).
Critical Arguements
<P1> Archives are established, administered, and evaluated by institutions, organizations, and individuals to ensure the retention, preservation, and utilization of archival holdings (p.701) <P2> The three principal categories of archival materials are official files of institutions and organizations, publications issued by such bodies, and personal papers of individuals. . . . Electronic information technologies have had profound effects on aspects of all these categories (p.702) <P3> The primary archival concern with regard to electronic publishing is that the published material should be transferred to archival custody. When the transfer occurs, the archivist must address the issues of authentication, appraisal, arrangement, description, and preservation or physical protection (p.702) <P4> The most effective way to satisfy archival requirements for handling electronic information is the establishment of procedures and standards to ensure that valuable material is promptly transferred to archival custody in a format which will permit access on equipment that will be readily available in the future (p.702) <P5> Long-term costs and access requirements are the crucial factors in determining how much information should be retained in electronic formats (p.703) <P6> Authentication involves a determination of the validity or integrity of information. Integrity requires the unbroked custody of a body of information by a responsible authority or individual <warrant> (p.703) <P7> From an archival perspective, the value of information is dependent on its content and the custodial responsibility of the agency that maintains it -- e.g., the source determines authenticity. The authentication of archival information requires that it be verified as to source, date, and content <warrant> (p.704) <P8> Information that is mutable, modifiable, or changeable loses its validity if the persons adding, altering, or deleting information cannot be identified and the time, place and nature of the changes is unknown (p.704) <P9> [P]reservation is more a matter of access to information than it is a question of survival of any physical information storage media (p.704) <P10> [T]o approach the preservation of electronic texts by focusing on physical threats will miss the far more pressing matter of ensuring continued accessibility to the information on such storage media (p.706) <P11> If the information is to remain accessible as long as paper, preservation must be a front-end, rather than an ex post facto, action (p.708) <P12> [T]he preservation of electronic texts is first and foremost a matter of editorial and administrative policy rather than of techniques and materials (p.708) <P13> Ultimately, the preservation of electronic publications cannot be solely an archival issue but an administrative one that can be addressed only if the creators and publishers take an active role in providing resources necessary to ensure that ongoing accesibility is part of initial system and product design (p.709) <P14> An encouraging development is that SGML has been considered to be a critical element for electronic publishing because of its transportability and because it supports multiple representations of a single text . . . (p.711) <P15> Underlying all questions of access is the fundamental consideration of cost (p.711)
Over the last decade a number of writers have encouraged archivists to develop strategies and tactics to redefine their role and to insert themselves into the process of designing recordkeeping systems. This paper urges archivists to exploit the authority inherent in the laws, regulations, standards, and professional best practices that dictate recordkeeping specifications to gain great acceptance for the requirements for electronic evidence. Furthermore, it postulates that this proactive approach could assist in gaining greater respect for the archival profession.
Critical Arguements
CA The use of authoritative sources of warrant would improve acceptance of electronic records as evidence and create greater respect for the archival profession.
Phrases
<P1> The legal, administrative, fiscal, or information value of records is dependent upon the degree of trust society places in records as reliable testimony or evidence of the acts they purport to document. In turn, this trust is dependent on society's faith in the procedures that control the creation and maintenance of the record. <P2> [S]ociety bestows some methods of recordkeeping and record creating with an authority or 'warrant' for generating reliable records. <P3> David Bearman first proposed the idea of "literary warrant." <P4> [S]tatements of warrant provide clear instructions on how records should be kept and delineate elements needed for the records to be complete. <P5> The information technology field promulgates standards, but in North America adherence to them is voluntary rather than obligatory. <P6> The University of Pittsburgh Electronic Recordkeeping Project suggested that requirements for electronic recordkeeping should derive from authoritative sources, such as the law, customs, standards, and professional best practices accepted by society and codified in the literature of different professions concerned with records and recordkeeping rather than developed in isolation. <P7> On their own, archival requirements for recordkeeping have very little authority as no authoritative agencies such as standards boards or professional associations have yet to endorse them [sic] and few archivists have the authority to insist that their organizations follow them. <P8> An NHPRC study suggested that archivists have not been involved in the process of meeting the challenges of electronic records because they are undervalued by their colleagues, or, in other words, are not viewed as a credible source.
Conclusions
RQ "By highlighting the similarity between recordkeeping requirements and the requirements delineated in authoritative statements in the law, auditing standards, and professional best practices, archivists will increase the power of their message. ... If archivists are to take their rightful place as regulators of an organization's documentary requirements, they will have to reach beyond their own professional literature and understand the requirements for recordkeeping imposed by other professions and society in general. Furthermore, they will have to study methods of increasing the accpetance of their message and the impact and power of warrant."
Type
Journal
Title
Defintions of electronic records: The European method
CA The consistent use of well-defined, agreed-upon terminology is a powerful tool for archivists. The point of view of diplomatics may be useful.
Phrases
<P1> It is very difficult for a European archivist or records manager to understand why it is necessary to use new terms to express old things. Literary warrant is one of these terms. If literary warrant simply means "best practice and professional culture" in recordkeeping, we only need to know what creators did for centuries and still do today (and probably will do also do in the future) in this area. (p. 220) <P2> Personally I am absolutely sure that without an effort at clarifying definitions in the recordkeeping environment, there is no way to obtain significant results in the field of electronic records. As the first, theoretical step, clarifying definitions will verify principles, from the juridical and technological point of view, which will be the basis of systems and particular applications. (p.220)
SOW
RQ What is the intrinsic nature of a record? How does the international community understand the concept of archival bond? Do they see it as something positive or a hindrance?
Type
Journal
Title
Building record-keeping systems: Archivists are not alone on the wild frontier
CA The digital environment offers archivists a host of new tools that can be adapted and used for recordkeeping. However, archivists must choose their tools judisciously while considering the long-term implications of their use as well as research and development. Ultimately, they must pick tools and strategies that dovetail with their institutions' specific needs while working to produce reliable and authentic records.
Phrases
<P1> Evidence from this review of emerging methods for secure and authentic electronic communications shows that the division of responsibility, accountability, and jurisdiction over recordkeeping is becoming more complex than a clear line between the records creator and the records preserver. (p.66) <P2> Storage of records in encrypted form is another area of concern because encryption adds additional levels of systems dependency on access to keys, proprietary encryption algorithims, hardware, and software. (p.62) <P3> It is important for archivists and records managers to understand parallel developments, because some new strategies and methods may support recordkeeping, while others may impede the achievement of archival objectives. (p.45) <P4> The concept of warrant and subsequent research on it by Wendy Duff is a significant contribution, because it situates the mandates for creating and maintaining records in a legal, administrative, and professional context, and it presents a methodology for locating, compiling, and presenting the rules governing proper and adequate documentation in modern organizations. (p. 48)
Conclusions
RQ Are electronic recordkeeping systems truly inherently inferior to paper-based systems in their capacity to maintain authentic records over time? How tightly can recordkeeping be integrated into normal business processes, and where does one draw the line between how a business does its work and how it does its recordkeeping?
Type
Journal
Title
How Do Archivists Make Electronic Archives Usable and Accessible?
CA In order to make electronic archives useable, archivists will need to enhance and link access systems to facilitate resource discovery while making the whole process as seamless and low-cost (or no-cost) as possible for the user.
Phrases
<P1> Rather than assuming that the archival community will succeed in transferring all valuable electronic records to archival institutions for preservation and future access, archivists must develop strategies and methods for accessibility and usability that can span a variety of custodial arrangements. (p.9) <P2> Maintaining linkages between different formats of materials will become increasingly burdensome if archvists do not find ways to develop integrated access systems. (p.10) <P3> Archivists must also think about ways to teach users the principles of a new digital diplomatics so that they can apply these principles themselves to make educated judgements about the accuracy, reliability, and authenticity of the documents they retrieve from electronic archives. (p.15)
Type
Journal
Title
Documenting digital images: Textual meta-data at the Blake Archive
The Electronic Library: The International Journal for Minicomputer, Microcomputer, and Software Applications in Libraries
Publication Year
1998
Volume
16
Issue
4
Pages
239
Critical Arguements
CA One of the critical issues in the future development of digital libraries is the provision of documentary metadata for non-textual electronic files.
Phrases
<P1> When libraries create digital image collections, however, documentation becomes more problematic. Even if an image is surrounded by a robust network of supporting materials, the functionality of client-server networks such as the World Wide Web permits the image to become detached from the documentary process." (p. 239)
Type
Journal
Title
Metadata Strategies and Archival Description: Comparing Apples to Oranges
Advocates of a "metadata systems approach" to the description of electronic records argue that metadata's capacity to provide descriptive information about the context of electronic records creation will obviate, or reduce significantly, the need for traditional archival description. This article examines the assumptions about the nature of archival description and of metadata on which metadata strategies are grounded, for the purposes of ascertaining the following: whether the skepticism concerning the capacity of traditional description to meet the challenges posed by the so-called "second generation" of electronic records is justified; whether the use of metadata as archival description is consistent with their nature and purpose; and whether metadata are capable of servinng archival descriptive purposes.
Critical Arguements
CA "Before the archival profession assigns to traditional archival description the diminished role of "added value" (i.e. accessory) or abandons it altogether, the assumptions about the nature of archival description and of metadata on which metadata strategies are grounded ought to be carefully examined. Such an examination is necessary to ascertain the following: whether the skepticism concerning the capacity of traditional description to meet the challenges posed by the so-called "second generation" of electronic records is justified, whether the use of metadata as archival description is consistent with their nature and purpose, and whether metadata are acapable of serving archival purposes."
Phrases
<P1> In an article published in Archivaria, David Wallace summarized recent writing on the subject of metadata and concluded that "[d]ata dictionaries and the types of metadata that they house and can be built to house should be seriously evaluated by archivists" because of their potential to signficantly improve and ultimately transform traditional archival practice in the areas of appraisal, arrangement, description, reference, and access. <warrant> <P2> In the area of description, specifically, advocates of "metadata management" or a "metadata systems approach" believe that metadata's capacity to provide descriptive information about the context of electronic records creation will obviate, or reduce significantly, the need for traditional description. <P3> Charles Dollar maintains that archival participation in the IRDS standard is essential to ensure that archival requirements, including descriptive requirements, are understood and adopted within it. <warrant> <P4> According to David Wallace, "archivists will need to concentrate their efforts on metadata systems creation rather than informational content descriptions, since in the electronic realm, archivists' concern for informational value will be eclipsed by concern for the evidential value of the system." <warrant> <P5> Charles Dollar, for his part, predicts that, rather than emphasize "the products of an information system," a metadata systems approach to description will focus on "an understanding of the information system context that supports organization-wide information sharing." <P6> Because their scope and context are comparitively narrow, metadata circumscribe and atomize these various contexts of records creation. Archival description, on the other hand, enlarges and integrates them. In so doing it reveals continuities and discontinuities in the matrix of function, structure, and record-keeping over time. <P7> Metadata are part of this broader context, since they constitute a series within the creator's fonds. The partial context provided by metadata should not, however, be mistaken for the whole context. <P8> Metadata, for example, may be capable of explaining contextual attributes of the data within an electronic records system, but they are incapable of describing themselves -- i.e., their own context of creation and use -- because they cannot be detached from themselves. For this reason, it is necessary to describe the context in which the metadata are created so that their meaning also will be preserved over time. <P9> A metadata system is like a diary that, in telegraphic style, records the daily events that take place in the life of an individual as they occur and from the individual's perspective. <P10> Archival description, it could be said, is the view from the plane; metadata, the view from the field as it is plowed. <P11> While a close-up shot-- such as the capture of a database view -- may be necessary for the purposes of preserving record context and system functionality, it does not follow that such a snapshot is necessary or even desirable for the purposes of description. <P12> Because the context revealed by metadata systems is so detailed, and the volume of transactions they capture is so enormous, metadata may in fact obscure, rather than illuminate, the broader administrative context and thereby bias the users' understanding of the records' meaning. In fact, parts of actions and transactions may develop entirely outside of the electronic system and never be included in the metadata. <P13> If the metadata are kept in their entirety, users searching for documents will have to wade through a great deal of irrelevant data to find what they need. If the metadata are chopped up into bits corresponding to what has been kept, how comprehensible will they be to the uesr? <P14> The tendency to describe metadata in metaphorical terms, e.g., in relation to archival inventories, has distracted attention from consideration of what metadata are in substantial, concrete terms. They are, in fact, records created and used in the conduct of affairs of which they form a part. <P15> The transactions captured by metadata systems may be at a more microscopic level than those captured in registers and the context may be more detailed, given the technological complexity of electronic record-keeping environments. Nevertheless, their function remains the same. <P16> And, like protocol registers, whose permanent retention is legislated, metadata need to be preserved in perpetuity because they are concrete evidence of what documents were made and received, who handled them, with what results, and the transactions to which they relate. <warrant> <P17> While it is true that metadata systems show or reveal the context in which transactions occur in an electronic system and therefore constitute a kind of description of it -- Jenkinson made the same observation about registers -- their real object is to record the fact of these transactions; they should be, like registers, "preserved as a [record] of the proceedings in that connection." <P18> Viewing metadata systems as tools for achieving archival purposes, rather than as tools for achieving the creators' purposes is dangerous because it encourages us to, in effect, privilege potential secondary uses of metadata over their actual primary use; in so doing, we could reshape such use for purposes other than the conduct of affairs of which they are a part. <P19> Metadata strategies risk compromising, specifically, the impartiality of the records' creation. <P20> For archivists to introduce in the formation of metadata records requirements directed toward the future needs of archivists and researchers rather than toward the current needs of the creator would contribute an element of self-consciousness into the records creation process that is inconsistent with the preservation of the records' impartiality. <P21> If the impartiality of the metadata is compromised, their value as evidence will be compromised, which means, ultimately, that the underlying objective of metadata strategies -- the preservation of evidence -- will be defeated. <P22> None of these objections should be taken to suggest that archivists do not have a role to play in the design and maintenance of metadata systems. It is, rather, to suggest that that role must be driven by our primary obligation to protect and preserve, to the extent possible, the essential characterisitcis of the archives. <P23> The proper role of an archivist in the design of a metadata system, then, is to assist the organization in identifying its own descriptive needs as well as to ensure that the identification process is driven, not by narrowly defined system requirements, but by the organization's overarching need and obligation to create and maintain complete, reliable, and authentic records. <P24> That is why it is essential that information holdings are identified and described in a meaningful way, organized in a logical manner that fascilitates their access, and preserved in a manner that permits their continuing use. <P25> Record-keeing requirements for electronic records must address the need to render documentary relationships wisible and to build in procedures for authentication and preservation; such measures will ensure that record-keeping systems meet the criteris of "intergrity, currency an relevancy" necessary to the records creator. <P26> In other words, effective description is a consequence of effective records management and intelligent appraisal, not their purpose. If the primary objectives of metadata are met, description will be fascilitated and the need for description at lower levels (e.g., below the series level) may even be obviated. <P27> Metadata systems cannot and should not replace archival description. To meet the challenges posed by electronic records, it is more important than ever that we follow the dictates of archival science, which begin from a consideration of the nature of archives. <P28> Archival participation in the design and maintenance of metadata systems must be driven by the need to preserve them as archival documents, that is, as evidence of actions and transactions, not as descriptive tools. Our role is not to promote our own intersts, but to deepen the creator's understanding of its interests in preserving the evidence of its own actions and transactions. We can contribute to that understanding because we have a broader view of the creator's needs over time. In supporting these interests, we indirectly promote our own. <P29> To ensure that our descriptive infrastructure is sound -- that is to say, comprehensible, flexible, efficient, and effective -- we need equally to analyze our own information management methods and, out of that analysis, to develop complementary systems of administrative and intellectual control that will build upon each other. By these means we will be able to accomodate the diversity and complexity of the record-keeping environments with which we must deal.
Conclusions
RQ "Since 'current metadata systems do not account for the provenancial and contextual information needed to manage archival records,' archivists are exhorted [by Margaret Hedstrom] to direct their research efforts (and research dollars) toward the identification of the types of metadata that ought to be captured and created to meet archival descriptive requirements. "
SOW
DC Dr. Heather MacNeil is an Assistant Professor at the School of Library, Archival, and Information Studies at the University of British Columbia. Dr. MacNeilÔÇÖs major areas of interests include: trends and themes in archival research & scholarship; arrangement and description of archival documents; management of current records; trustworthiness of records as evidence; protection of personal privacy; interdisciplinary perspectives on record trustworthiness; and archival preservation of authentic electronic records
This article provides an overview of evolving Australian records continuum theory and the records continuum model, which is interpreted as both a metaphor and a new worldview, representing a paradigm shift in Kuhn's sense. It is based on a distillation of research findings drawn from discourse, literary warrant and historical analysis, as well as case studies, participant observation and reflection. The article traces the emergence in Australia in the 1990s of a community of practice which has taken continuum rather than life cycle based perspectives, and adopted postcustodial approaches to recordkeeping and archiving. It "places" the evolution of records continuum theory and practice in Australia in the context of a larger international discourse that was reconceptualizing traditional theory, and "reinventing" records and archives practice.
Publisher
Kluwer Academic Publishers
Publication Location
The Netherlands
Critical Arguements
CA Looks at the development of the Australian community of practice that led to records continuum theory: an approach that, in contrast to the North American life cycle approach, sees recordkeeping and archival practices as part of the same continuum of activities. Since the 1990s, there has been a lively debate between proponents of these two different ways of thinking. The second part of the article is highly theoretical, situating records continuum theory in the larger intellectual trend toward postmodernism and postpositivism.
Phrases
<P1> The model was built on a unifying concept of records inclusive of archives, which are defined as records of continuing value. It also drew on ideas about the "fixed" and "mutable" nature of records, the notion that records are ÔÇ£always in a process of becoming." (p. 334). <P2> Continuum ideas about the nature of records and archives challenge traditional understandings which differentiate "archives" from "records" on the basis of selection for permanent preservation in archival custody, and which focus on their fixed nature. Adopting a pluralist view of recorded information, continuum thinking characterises records as a special genre of documents in terms of their intent and functionality. It emphasises their evidentiary, transactional and contextual nature, rejecting approaches to the definition of records which focus on their subject content and informational value. (p. 335) <P3> [R]ecordkeeping and archiving processes ... help to assure the accessibility of meaningful records for as long as they are of value to people, organisations, and societies ÔÇô whether that be for a nanosecond or millennia. (p. 336) <P4> [I]f North American understandings of the term record keeping, based on life cycle concepts of records management, are used to interpret the writings of members of the Australian recordkeeping community, there is considerable potential for misunderstanding. <P5> Members of the recordkeeping and archiving community have worked together, often in partnership with members of other records and archives communities, on a range of national policy and standards initiatives, particularly in response to the challenge of electronic recordkeeping. These collaborative efforts resulted in AS 4390, the Australian Standard: Records Management (1996), the Australian Council of Archives' Common Framework for Electronic Recordkeeping (1996), and the Australian Records and Archives Competency Standards (1997). In a parallel and interconnected development, individual archival organisations have been developing electronic recordkeeping policies, standards, system design methodologies, and implementation strategies for their jurisdictions, including the National Archives of Australia's suite of standards, policies, and guidelines under the e-permanence initiative launched in early 2000. These developments have been deliberately set within the broader context of national standards and policy development frameworks. Two of the lead institutions in these initiatives are the National Archives of Australia and the State Records Authority of New South Wales, which have based their work in this area on exploration of fundamental questions about the nature of records and archives, and the role of recordkeeping and archiving in society. <warrant> (p. 339) <P6> In adopting a continuum-based worldview and defining its "place" in the world, the Australian recordkeeping and archiving community consciously rejected the life cycle worldview that had dominated records management and archives practice in the latter half of the 20th century in North America. ... They were also strong advocates of the nexus between accountable recordkeeping and accountability in a democratic society, and supporters of the dual role of an archival authority as both a regulator of current recordkeeping, and preserver of the collective memory of the state/nation. (p. 343-344) <P7> [P]ost-modern ideas about records view them as dynamic objects that are fixed in terms of content and meaningful elements of their structure, but linked to ever-broadening layers of contextual metadata that manages their meanings, and enables their accessibility and useability as they move through "spacetime." (p. 349) <P8> In exploring the role of recordkeeping and archiving professionals within a postcustodial frame of reference, archival theorists such as Brothman, Brown, Cook, Harris, Hedstrom, Hurley, Nesmith, and Upward have concluded that they are an integral part of the record and archive making and keeping process, involved in society's remembering and forgetting. (p. 355) <P9> Writings on the societal context of functional appraisal have gone some way to translate into appraisal policies and strategies the implications of the shifts in perception away from seeing records managers as passive keepers of documentary detritus ... and archivists as Jenkinson's neutral, impartial custodians of inherited records. (p. 355-356)
Conclusions
RQ "By attempting to define, to categorise, pin down, and represent records and their contexts of creation, management, and use, descriptive standards and metadata schema can only ever represent a partial view of the dynamic, complex, and multi-dimensional nature of records, and their rich webs of contextual and documentary relationships. Within these limitations, what recordkeeping metadata research is reaching towards are ways to represent records and their contexts as richly and extensively as possible, to develop frameworks that recognise their mutable and contingent nature, as well as the role of recordkeeping and archiving professionals (records managers and archivists) in their creation and evolution, and to attempt to address issues relating to time and space." (p. 354)
Type
Journal
Title
Grasping the Nettle: The Evolution of Australian Archives Electronic Records Policy
CA An overview of the development of electronic records policy at the Australian Archives.
Phrases
<P1> The notion of records being independent of format and of "virtual" records opens up a completely new focus on what it is that archival institutions are attempting to preserve. (p. 136) <P2> The import of Bearman's contention that not all infomation systems are recordkeeping systems challenges archivists to move attention away from managing archival records after the fact toward involvement in the creation phase of records, i.e., in the systems design and implementation process. (p. 139) <P3> The experience of the Australian Archives is but one slice of a very large pie, but I think it is a good indication of the challenges other institutions are facing internationally. (p. 144)
Conclusions
RQ How has the Australian Archives managed the transition from paper to electronic records? What issues were raised and how were they dealt with?
One such expedient could be more structured and more integrated use of formal and institutional data on records and archives. I cannot offer any completed model of this enhanced perspective, and as far as I know, one does not exist. However, it is a new way of thinking and looking at the problems we encounter. What I would like to do is draw attention to some of the approaches now being developed in The Netherlands. In a way, this presentation will therefore be a report on the Dutch arvhival situation.
Critical Arguements
CA "In a world defined by the enormous size of archives, where the multiplicity of records is in turn driven by the growing complexity of society and its administration, and by the proliferation of types of 'information carriers', it is becoming increasingly difficult fpr archivists to fulfill their primary tasks. It is therefore necessary to study carefully the development of maintenance and control mechanisms for archives. We cannot afford waste or overlook any possibility. It is also necessary to look around us, to discover what other archivists in other countries are doing, and what others in related fields, such as libraries and museums, have accomplished. Essentially, we all deal with the same problems and must try to find new solutions to master these problems."
Phrases
<P1> Document forms can be regarded as forms of objects. We probably need to gain more experience in recognizing different forms of documents and interpreting them, but once we have this knowledge, we can use it in the same way as we now use 'form' in its archival sense: to distinguish one object from another. <P2> In fact, by extension, one can even construct and defend the thesis that all decisions in an administration are reached using standard procedures and forms. Once this is realized, one can ask: what use to archivists make of this knowledge in their daily work? What are the possibilities? <P3> Often the forms of materials created prove to be of a more consistent nature than the offices that use them. If an office ceases its activity, another will take over its tasks and for the most part will use the same or almost the same forms of material. <P4> Understanding the functions of the organization will provide archivists not only with information about the material involved, but also with knowledge of the procedures, which in turn provides information about the records and their different forms. This kind of sympathetic understanding enables archivists to make all kinds of decisions, and it is important to note that at least part of this knowledge should be provided to the users, so that they can decide which records might be of interest to them. <warrant> <P5> We are increasingly aware that we must distinguish between processing an archive (i.e. organizing records according to archival principles after appraisal) and making the contents available for users through finding aids, indexes and other means. <P6> With respect to the latter, it is clear that archivists should make use of both context- and provenance-based indexing. They should take advantage of the possiblities offered by the structures and forms of material -- something which the librarian cannot do. Furthermore, they should also use content indexing in a selective way, only when they think it necessary [to] better serve researchers. <warrant> <P7> The National Archives in The Hague has responded to these new perspectives by developing a computer programme called MAIS (Micro Archives Inventory System), which is a formal way of processing archives based on provenance. <P8> The object of this presentation has been to show that use of structure, forms of material and functions, can aid the archivist in his/her work.
Conclusions
RQ "While these initial Dutch efforts have been produced in a rather unorganized way, it should nevertheless be possible to approach the work more systematically in [the] future, building up a body of knowledge of forms for users of archives. David Bearman has offered some preliminary suggestions in this direction, in the article cited above; it is now a matter of more research required to realize something positive in this field."
SOW
DC J. Peter Sigmond is Director of Collections at the Rijksmuseum in Amsterdam, the Netherlands
Type
Journal
Title
Structuring the Records Continuum Part Two: Structuration Theory and Recordkeeping
In the previous issue of Archives and Manuscripts I presented the first part of this two part exploration. It dealt with some possible meanings for 'post' in the term postcustodial. For archivists, considerations of custody are becoming more complex because of changing social, technical and legal considerations. These changes include those occurring in relation to access and the need to document electronic business communications reliably. Our actions, as archivists, in turn become more complex as we attempt to establish continuity of custody in electronic recordkeeping environments. In this part, I continue the case for emphasising the processes of archiving in both our theory and practice. The archives as a functional structure has dominated twentieth century archival discourse and institutional ordering, but we are going through a period of transformation. The structuration theory of Anthony Giddens is used to show that there are very different ways of theorising about our professional activities than have so far been attempted within the archival profession. Giddens' theory, at the very least, provides a useful device for gaining insights into the nature of theory and its relationship with practice. The most effective use of theory is as a way of seeing issues. When seen through the prism of structuration theory, the forming processes of the virtual archives are made apparent.
Critical Arguements
CA "This part of my exploration of the continuum will continue the case for understanding 'postcustodial' as a bookmark term for a major transition in archival practice. That transition involves leaving a long tradition in which continuity was a matter of sequential control. Electronic recordkeeping processes need to incorporate continuity into the essence of recordkeeping systems and into the lifespan of documents within those systems. In addressing this issue I will present a structurationist reading of the model set out in Part 1, using the sophisticated theory contained in the work of Anthony Giddens. Structuration theory deals with process, and illustrates why we must constantly re-assess and adjust the patterns for ordering our activities. It gives some leads on how to go about re-institutionalising these new patterns. When used in conjunction with continuum thinking, Giddens' meta-theory and its many pieces can help us to understand the complexities of the virtual archives, and to work our way towards the establishment of suitable routines for the control of document management, records capture, corporate memory, and collective memory."
Phrases
<P1> Broadly the debate has started to form itself as one between those who represent the structures and functions of an archival institution in an idealised form, and those who increasingly concentrate on the actions and processes which give rise to the record and its carriage through time and space. In one case the record needs to be stored, recalled and disseminated within our institutional frameworks; in the other case it is the processes for storing, recalling, and disseminating the record which need to be placed into a suitable framework. <P2> Structure, for Giddens, is not something separate from human action. It exists as memory, including the memory contained within the way we represent, recall, and disseminate resources including recorded information. <P3> Currently in electronic systems there is an absence of recordkeeping structures and disconnected dimensions. The action part of the duality has raced ahead of the structural one; the structuration process has only just begun. <P4> The continuum model's breadth and richness as a conceptual tool is expanded when it is seen that it can encompass action-structure issues in at least three specialisations within recordkeeping: contemporary recordkeeping - current recordkeeping actions and the structures in which they take place; regulatory recordkeeping - the processes of regulation and the enabling and controlling structures for action such as policies, standards, codes, legislation, and promulgation of best practices; historical recordkeeping - explorations of provenance in which action and structure are examined forensically as part of the data sought about records for their storage, recall and dissemination. <P5> The capacity to imbibe information about recordkeeping practices in agencies will be crucial to the effectiveness of the way archival 'organisations' set up their postcustodial programs. They will have to monitor the distribution and exercise of custodial responsibilities for electronic records from before the time of their creation. <warrant> <P6> As John McDonald has pointed out, recordkeeping activities need to occur at desktop level within systems that are not dependent upon the person at the desktop understanding all of the details of the operation of that system. <P7> Giddens' more recent work on reflexivity has many parallels with metadata approaches to recordkeeping. What if the records, as David Bearman predicts, can be self-managing? Will they be able to monitor themselves? <P8> He rejects the life cycle model in sociology, based on ritualised passages through life, and writes of 'open experience thresholds'. Once societies, for example, had rites for coming of age. Coming of age in a high modern society is now a complex process involving a host of experiences and risks which are very different to that of any previous generation. Open experience threshholds replace the life cycle thresholds, and as the term infers, are much less controlled or predictable. <P9> There is a clear parallel with recordkeeping in a high modern environment. The custodial thresholds can no longer be understood in terms of the spatial limits between a creating agency and an archives. The externalities of the archives as place will decline in significance as a means of directly asserting the authenticity and reliability of records. The complexities of modern recordkeeping involve many more contextual relationships and an ever increasing network of relationships between records and the actions that take place in relation to them. We have no need for a life cycle concept based on the premise of generational repetition of stages through which a record can be expected to pass. We have entered an age of more recordkeeping choices and of open experience thresholds. <P10> It is the increase in transactionality, and the technologies being used for those transactions, which are different. The solution, easier to write about than implement, is for records to parallel Giddens' high modern individual and make reflexive use of the broader social environment in which they exist. They can reflexively monitor their own action and, with encoding help from archivists and records managers, resolve their own crises as they arise. <warrant> <P11> David Bearman's argument that records can be self-managing goes well beyond the easy stage. It is supported by the Pittsburgh project's preliminary set of metadata specifications. The seeds of self-management can be found in object oriented programming, java, applets, and the growing understanding of the importance and nature of metadata. <P12> Continuum models further assist us to conceive of how records, as metadata encapsulated objects, can resolve many of their own life crises as they thread their way through time and across space. <P13> To be effective monitors of action, archival institutions will need to be recognised by others as the institutions most capable of providing guidance and control in relation to the integration of the archiving processes involved in document management, records capture, the organisation of corporate memory and the networking of archival systems. <warrant> <P14> Signification, in the theoretical domain, refers to our interpretative schemes and the way we encode and communicate our activities. At a macro level this includes language itself; at a micro level it can include our schemes for classification and ordering. <P15> The Pittsburgh project addressed the three major strands of Giddens' theoretical domain. It explored and set out functional requirements for evidence - signification. It sought literary warrants for archival tasks - legitimation. It reviewed the acceptability of the requirements for evidence within organisational cultures - domination. <P16> In Giddens' dimensional approach, the theoretical domain is re-defined to be about coding, organising our resources, and developing norms and standards. In this area the thinking has already begun to produce results, which leads this article in to a discussion of structural properties. <P17> Archivists deal with structural properties when, for example, they analyse the characteristics of recorded information such as the document, the record, the archive and the archives. The archives as a fortress is an observable structural property, as is the archives as a physical accumulation of records. Within Giddens' structuration theory, when archivists write about their favourite features, be they records or the archives as a place, they are discussing structural properties. <P18> Postcustodial practice in Australia is already beginning to put together a substantial array of structural properties. These developments are canvassed in the article by O'Shea and Roberts in the previous issue of Archives and Manuscripts. They include policies and strategies, standards, recordkeeping regimes, and what has come to be termed distributed custody. <P19> As [Terry] Eastwood comments in the same article, we do not have adequate electronic recordkeeping systems. Without them there can be no record in time-space to serve any form of accountability. <warrant> <P20> In the Pittsburgh project, for example, the transformation of recordkeeping processes is directed towards the creation and management of evidence, and possible elements of a valid rule-resource set have emerged. Elements can include the control of recordkeeping actions, accountability, the management of risk, the development of recordkeeping regimes, the establishment of recordkeeping requirements, and the specification of metadata. <P21> In a postcustodial approach it is the role of archival institutions to foster better recordkeeping practices within all the dimensions of recordkeeping. <warrant>
Conclusions
RQ "Best practice in the defence of the authoritative qualities of records can no longer be viewed as a linear chain, and the challenge is to establish new ways of legitimating responsibilities for records storage and custody which recognise the shifts which have occurred." ... "The recordkeeping profession should seek to establish itself as ground cover, working across terrains rather than existing tree-like in one spot. Beneath the ground cover there are shafts of specialisation running both laterally and vertically. Perhaps we can, as archivists, rediscover something that a sociologist like Giddens has never forgotten. Societies, including their composite parts, are the ultimate containers of recorded information. As a place in society, as Terry Cook argues, the archives is a multiple reality. We can set in train policies and strategies that can help generate multiplicity without losing respect for particular mine shafts. Archivists have an opportunity to pursue policies which encourage the responsible exercising of a custodial role throughout society, including the professions involved in current, regulatory and historical recordkeeping. If we take up that opportunity, our many goals can be better met and our concerns will be addressed more effectively."
SOW
DC "Frank Upward is a senior lecturer in the Department of Librarianship, Archives and Records at Monash University. He is an historian of the ideas contained in the Australian records continuum approach, and an ex- practitioner within that approach." ... "These two articles, and an earlier one on Ian Maclean and the origins of Australian continuum thinking, have not, so far, contained appropriate acknowledgements. David Bearman provided the necessary detonation of certain archival practices, and much more. Richard Brown and Terry Cook drew my attention to Anthony Giddens' work and their own work has helped shape my views. I have many colleagues at Monash who encourage my eccentricities. Sue McKemmish has helped shape my ideas and my final drafts and Barbara Reed has commented wisely on my outrageous earlier drafts. Livia Iacovino has made me stop and think more about the juridical tradition in recordkeeping. Chris Hurley produced many perspectives on the continuum during the 1996 seminars which have helped me see the model more fully. Don Schauder raised a number of key questions about Giddens as a theorist. Bruce Wearne of the Sociology Department at Monash helped me lift the clarity of my sociological explanations and made me realise how obsessed Giddens is with gerunds. The structural-functionalism of Luciana Duranti and Terry Eastwood provided me with a counterpoint to many of my arguments, but I also owe them debts for their respective explorations of recordkeeping processes and the intellectual milieu of archival ideas, and for their work on the administrative-juridical tradition of recordkeeping. Glenda Acland has provided perceptive comments on my articles - and supportive ones, for which I am most grateful given how different the articles are from conventional archival theorising. Australian Archives, and its many past and present staff members, has been important to me."
Type
Journal
Title
Structuring the Records Continuum Part One: Post-custodial principles and properties
The records continuum is becoming a much used term, but has seldom been defined in ways which show it is a time/space model not a life of the records model. Dictionary definitions of a continuum describe such features as its continuity, the indescernibility of its parts, and the way its elements pass into each other. Precise definitions, accordingly, have to discern the indiscernible, identify points that are not distinct, and do so in ways which accomodate the continuity of change. This article, and a second part to be published in the next volume, will explore the continuum in time/space terms supported by a theoretical mix of archival science, postmodernity and the 'structuration theory' of Anthony Giddens. In this part the main objectives are to give greater conceptual firmness to the continuum; to clear the way for broader considerations of the nature of the continuum by freeing archivists from the need to debate custody; to show how the structural principles for archival practice are capable of different expression without losing contact with something deeper that can outlive the manner of expression.
Critical Arguements
CA "This is the first instalment of a two part article exploring the records continuum. Together the articles will build into a theory about the constitution of the virtual archives. In this part I will examine what it can mean to be 'postcustodial', outline some possible structural principles for the virtual archives, and present a logical model for the records continuum." ... "In what follows in the remainder of this article (and all of the next) , I will explore the relevance of [Anthony] Giddens' theory to the structuring of the records continuum."
Phrases
<P1> If the archival profession is to avoid a fracture along the lines of paper and electronic media, it has to be able to develop ways of expressing its ideas in models of relevance to all ages of recordkeeping, but do so in ways which are contemporaneous with our own society. <warrant> <P2> We need more of the type of construct provided by the Pittsburgh Project's functional requirements for evidence which are 'high modern' but can apply to recordkeeping over time. <P3> What is essential is for electronic records to be identified, controlled and accessible for as long as they have value to Government and the Community. <warrant> <P4> We have to face up to the complexification of ownership, possession, guardianship and control within our legal system. Even possession can be broken down into into physical possession and constructed possession. We also have to face the potential within our technology for ownership, possession, custody or control to be exercised jointly by the archives, the organisation creating the records, and auditing agencies. The complexity requires a new look at our way of allocating authorities and responsibilities. <P5> In what has come to be known as the continuum approach Maclean argued that archivists should base their profession upon studies of the characteristics of recorded information, recordkeeping systems, and classification (the way the records were ordered within recordkeeping systems and the way these were ordered through time). <P6> A significant role for today's archival institution is to help to identify and establish functional requirements for recordkeeping that enable a more systematic approach to authentication than that provided by physical custody. <warrant> <P7> In an electronic work environment it means, in part, that the objectivity, understandability, availability, and usability of records need to be inherent in the way that the record is captured. In turn the documents need to be captured in the context of the actions of which they are part, and are recursively involved. <warrant> <P8>A dimensional analysis can be constructed from the model and explained in a number of ways including a recordkeeping system reading. When the co-ordinates of the continuum model are connected, the different dimensions of a recordkeeping system are revealed. The dimensions are not boundaries, the co-ordinates are not invariably present, and things may happen simultaneously across dimensions, but no matter how a recordkeeping system is set up it can be analysed in terms such as: first dimensional analysis: a pre- communication system for document creation within electronic systems [creating the trace]; second dimensional analysis: a post- communication system, for example traditional registry functionality which includes registration, the value adding of data for linking documents and disseminating them, and the maintenance of the record including disposition data [capturing trace as record]; third dimensional analysis: a system involving building, recalling and disseminating corporate memory [organising the record as memory]; fourth dimensional analysis: a system for building, recalling and disseminating collective memory (social, cultural or historical) including information of the type required for an archival information system [pluralizing the memory]. <P9> In the high modern recordkeeping environment of the 1990's a continuum has to take into account a different array of recordkeeping tools. These tools, plucking a few out at random but ordering the list dimensionally, include: document management software, Australian records system software, the intranet and the internet. <P10> In terms of a records continuum which supports an evidence based recordkeeping approach, the second dimension is crucial. This is where the document is disembedded from the immediate contexts of the first dimension. It is this disembedding process that gives the record its value as a 'symbolic token'. A document is embedded in an act, but the document as a record needs to be validatable using external reference points. These points include the operation of the recordkeeping system into which it was received, and information pertaining to the technical, social (including business) and communication processes of which the document was part.
Conclusions
RQ "Postcustodial approaches to archives and records cannot be understood if they are treated as a dualism. They are not the opposite of custody. They are a response to opportunities for asserting the role of an archives - and not just its authentication role - in many re-invigorating ways, a theme which I will explore further in the next edition of Archives and Manuscripts."
SOW
DC "Frank Upward is a senior lecturer in the Department of Librarianship, Archives and Records at Monash University. He is an historian of the ideas contained in the Australian records continuum approach, and an ex-practitioner within that approach."
Type
Journal
Title
Archives and the information superhighway: Current status and future challenges
CA One struggle facing us is to convince the rest of society that the ÔÇ£information superhighwayÔÇØ is very much about records, evidence and ÔÇ£recordnessÔÇØ.
Phrases
<P1> It has been argued that existing computer software applications harm recordkeeping because they are remiss in capturing the full breadth of contextual information required to document transactions and create records -- records which can serve as reliable evidence of the transactions which created them. In place of records, these systems are producing data which fails to relate the who, what, when, where, and why of human communications -- attributes which are required for record evidence. This argument has found both saliency and support in other work conducted by the Netherlands and the World Bank, which have both noted that existing software applications fail to provide for the capture of the required complement of descriptive attributes required for proper recordkeeping. These examples point to the vast opportunity presented to archivists to position themselves as substantive contributors to information infrastructure discussions. Archivists are capable of pointing out what will be necessary to create records in the electronic environment which, in the words of David Bearman, meet the requirements of ÔÇ£business acceptable commincation. (p.87) <warrant>
Conclusions
RQ Can archivists provide access to information in the unstable electronic records environment we find ourselves in today?
Type
Journal
Title
Managing the Present: Metadata as Archival Description
Traditional archival description undertaken at the terminal stages of the life cycle has had two deleterious effects on the archival profession. First, it has resulted in enormous, and in some cases, insurmountable processing backlogs. Second, it has limited our ability to capture crucial contextual and structural information throughout the life cycle of record-keeping systems that are essential for fully understanding the fonds in our institutions. This shortcoming has resulted in an inadequate knowledge base for appraisal and access provision. Such complications will only become more magnified as distributed computering and complex software applications continue to expand throughout organizations. A metadata strategy for archival description will help mitigate these problems and enhance the organizational profile of archivists who will come to be seen as valuable organizational knowledge and accountability managers.
Critical Arguements
CA "This essay affirms this call for evaluation and asserts that the archival profession must embrace a metadata systems approach to archival description and management." ... "It is held here that the requirements for records capture and description are the requirements for metadata."
Phrases
<P1> New archival organizational structures must be created to ensure that records can be maintained in a usable form. <warrant> <P2> The recent report of Society of American Archivists (SAA) Committee on Automated Records and Techniques (CART) on curriculum development has argued that archivists need to "understand the nature and utility of metadata and how to interpret and use metadata for archival purposes." <warrant> <P3> The report advises archivists to acquire knowledge on the meanings of metadata, its structures, standards, and uses for the management of electronic records. Interestingly, the requirements for archival description immediately follow this section and note that archivists need to isolate the descriptive requirements, standards, documentiation, and practices needed for managing electronic records. <warrant> <P4> Clearly, archivists need to identify what types of metadata will best suit their descriptive needs, underscoring the need for the profession to develop strategies aand tactics to satisfy these requirements within active software environments. <warrant> <P5> Underlying the metadata systems strategy for describing and managing electronic information technologies is the seemingly universal agreement amongst electronic records archivists on the requirement to intervene earlier in the life cycle of electronic information systems. <warrant> <P6> Metadata has loomed over the archival management of electronic records for over five years now and is increasingly being promised as a basic control strategy for managing these records. <warrant> <P7> However, she [Margaret Hedstrom] also warns that as descriptive practices shift from creating descriptive information to capturing description along with the records, archivists may discover that managing the metadata is a much greater challenge than managing the records themselves. <P8> Archivists must seek to influence the creation of record-keeping systems within organizations by connecting the transaction that created the data to the data itself. Such a connection will link informational content, structure, and the context of transactions. Only when these conditions are met will we have records and an appropriate infrastructure for archival description. <warrant> <P9> Charles Dollar has argued that archivists increasingly will have to rely upon and shape the metadata associated with electronic records in order to fully capture provenance information about them. <warrant> <P10> Bearman proposes a metadata systems strategy, which would focus more explicitly on the context out of which records arise, as opposed to concentrating on their content. This axiom is premised on the assumption that "lifecycle records systems control should drive provenance-based description and link to top-down definitions of holdings." <warrant> <P11> Bearman and Margaret Hedstrom have built upon this model and contend that properly specified metadata capture could fully describe sytems while they are still active and eliminate the need for post-hoc description. The fundamental change wrought in this approach is the shift from doing things to records (surveying, scheduling, appraising, disposing/accessioning, describing, preserving, and accessing) to providing policy direction for adequate documentation through management of organizational behavior (analyzing organizational functions, defining business transactions, defining record metadata, indentifying control tactics, and establishing the record-keeping regime). Within this model archivists focus on steering how records will be captured (and that they will be captured) and how they will be managed and described within record-keeping systems while they are still actively serving their parent organization. <P12> Through the provision of policy guidance and oversight, organizational record-keeping is managed in order to ensure that the "documentation of organizational missions, functions, and responsibilities ... and reporting relationships within the organization, will be undertaken by the organizations themselves in their administrative control systems." <warrant> <P13> Through a metadata systems approach, archivists can realign themselves strategically as managers of authoritative information about organizational record-keeping systems, providing for the capture of information about each system, its contextual attributes, its users, its hardware configurations, its software configurations, and its data configurations. <warrant> <P14> The University of Pittsburgh's functional requirements for record-keeping provides a framework for such information management structure. These functional requirements are appropriately viewed as an absolute ideal, requiring testing within live systems and organizations. If properly implemented, however, they can provide a concrete model for metadata capture that can automatically supply many of the types of descriptive information both desired by archivists and required for elucidating the context out of which records arise. <P15> It is possible that satisfying these requirements will contribute to the development of a robust archival description process integrating "preservation of meaning, exercise of control, and provision of access'" within "one prinicipal, multipurpose descriptive instrument" hinted at by Luciana Duranti as a possible outcome of the electronic era. <P16> However, since electronic records are logical and not physical entities, there is no physical effort required to access and process them, just mental modelling. <P17> Depending on the type of metadata that is built into and linked to electronic information systems, it is possible that users can identify individual records at the lowest level of granularity and still see the top-level process it is related to. Furthermore, records can be reaggregated based upon user-defined criteria though metadata links that track every instance of their use, their relations to other records, and the actions that led to their creation. <P18> A metadata strategy for archival description will help to mitigate these problems and enhance the organizational profile of archivists, who will come to be seen as valuable organizational knowledge and accountability managers. <warrant>
Conclusions
RQ "First and foremost, the promise of metadata for archival description is contingent upon the creation of electronic record-keeping systems as opposed to a continuation of the data management orientation that seems to dominate most computer applications within organizations." ... "As with so many other aspects of the archival endeavour, these requirements and the larger metadata model for description that they are premised upon necessitate further exploration through basic research."
SOW
DC "In addition to New York State, recognition of the failure of existing software applications to capture a full compliment of metadata required for record-keeping and the need for such records management control has also been acknowledged in Canada, the Netherlands, and the World Bank." ... "In conjunction with experts in electronic records managment, an ongoing research project at the University of Pittsburgh has developed a set of thirteen functional requirements for record-keeping. These requirements provide a concrete metadata tool sought by archivists for managing and describing electronic records and electronic record-keeping systems." ... David A. Wallace is an Assistant Professor at the School of Information, University of Michigan, where he teaches in the areas of archives and records management. He holds a B.A. from Binghamton University, a Masters of Library Science from the University at Albany, and a doctorate from the University of Pittsburgh. Between 1988 and 1992, he served as Records/Systems/Database Manager at the National Security Archive in Washington, D.C., a non-profit research library of declassified U.S. government records. While at the NSA he also served as Technical Editor to their "The Making of U.S. Foreign Policy" series. From 1993-1994, he served as a research assistant to the University of Pittsburgh's project on Functional Requirements for Evidence in Recordkeeping, and as a Contributing Editor to Archives and Museum Informatics: Cultural Heritage Informatics Quarterly. From 1994 to 1996, he served as a staff member to the U.S. Advisory Council on the National Information Infrastructure. In 1997, he completed a dissertation analyzing the White House email "PROFS" case. Since arriving at the School of Information in late 1997, he has served as Co-PI on an NHPRC funded grant assessing strategies for preserving electronic records of collaborative processes, as PI on an NSF Digital Government Program funded planning grant investigating the incorporation of born digital records into a FOIA processing system, co-edited Archives and the Public Good: Accountability and Records in Modern Society (Quorum, 2002), and was awarded ARMA International's Britt Literary Award for an article on email policy. He also serves as a consultant to the South African History Archives Freedom of Information Program and is exploring the development of a massive digital library of declassified imaged/digitized U.S. government documents charting U.S. foreign policy.
Type
Electronic Journal
Title
ARTISTE: An integrated Art Analysis and Navigation Environment
This article focuses on the description of the objectives of the ARTISTE project (for "An integrated Art Analysis and Navigation environment") that aims at building a tool for the intelligent retrieval and indexing of high resolution images. The ARTISTE project will address professional users in the fine arts as the primary end-user base. These users provide services for the ultimate end-user, the citizen.
Critical Arguements
CA "European museums and galleries are rich in cultural treasures but public access has not reached its full potential. Digital multimedia can address these issues and expand the accessible collections. However, there is a lack of systems and techniques to support both professional and citizen access to these collections."
Phrases
<P1> New technology is now being developed that will transform that situation. A European consortium, partly funded by the EU under the fifth R&D framework, is working to produce a new management system for visual information. <P2> Four major European galleries (The Uffizi in Florence, The National Gallery and the Victoria and Albert Museum in London and the Louvre related restoration centre, Centre de Recherche et de Restauration des Mus├®es de France) are involved in the project. They will be joining forces with NCR, a leading player in database and Data Warehouse technology; Interactive Labs, the new media design and development facility of Italy's leading art publishing group, Giunti; IT Innovation, Web-based system developers; and the Department of Electronics and Computer Science at the University of Southampton. Together they will create web based applications and tools for the automatic indexing and retrieval of high-resolution art images by pictorial content and information. <P3> The areas of innovation in this project are as follows: Using image content analysis to automatically extract metadata based on iconography, painting style etc; Use of high quality images (with data from several spectral bands and shadow data) for image content analysis of art; Use of distributed metadata using RDF to build on existing standards; Content-based navigation for art documents separating links from content and applying links according to context at presentation time; Distributed linking and searching across multiple archives allowing ownership of data to be retained; Storage of art images using large (>1TeraByte) multimedia object relational databases. <P4> The ARTISTE approach will use the power of object-related databases and content-retrieval to enable indexing to be made dynamically, by non-experts. <P5> In other words ARTISTE would aim to give searchers tools which hint at links due to say colour or brush-stroke texture rather than saying "this is the automatically classified data". <P6> The ARTISTE project will build on and exploit the indexing scheme proposed by the AQUARELLE consortia. The ARTISTE project solution will have a core component that is compatible with existing standards such as Z39.50. The solution will make use of emerging technical standards XML, RDF and X-Link to extend existing library standards to a more dynamic and flexible metadata system. The ARTISTE project will actively track and make use of existing terminology resources such as the Getty "Art and Architecture Thesaurus" (AAT) and the "Union List of Artist Names" (ULAN). <P7> Metadata will also be stored in a database. This may be stored in the same object-relational database, or in a separate database, according to the incumbent systems at the user partners. <P8> RDF provides for metadata definition through the use of schemas. Schemas define the relevant metadata terms (the namespace) and the associated semantics. Individual RDF queries and statements may use multiple schemas. The system will make use of existing schemas such as the Dublin Core schema and will provide wrappers for existing resources such as the Art and Architecture thesaurus in a RDF schema wrapper. <P9> The Distributed Query and Metadata Layer will also provide facilities to enable queries to be directed towards multiple distributed databases. The end user will be able to seamlessly search the combined art collection. This layer will adhere to worldwide digital library standards such as Z39.50, augmenting and extending as necessary to allow the richness of metadata enabled by the RDF standard.
Conclusions
RQ "In conclusion the Artiste project will result into an interesting and innovative system for the art analysis, indexing storage and navigation. The actual state of the art of content-based retrieval systems will be positively influenced by the development of the Artiste project, which will pursue the following goals: A solution which can be replicated to European galleries, museums, etc.; Deep-content analysis software based on object relational database technology.; Distributed links server software, user interfaces, and content-based navigation software.; A fully integrated prototype analysis environment.; Recommendations for the exploitation of the project solution by European museums and galleries. ; Recommendations for the exploitation of the technology in other sectors.; "Impact on standards" report detailing augmentations of Z39.50 with RDF." ... ""Not much research has been carried out worldwide on new algorithms for style-matching in art. This is probably not a major aim in Artiste but could be a spin-off if the algorithms made for specific author search requirements happen to provide data which can be combined with other data to help classify styles." >
SOW
DC "Four major European galleries (The Uffizi in Florence, The National Gallery and the Victoria and Albert Museum in London and the Louvre related restoration centre, Centre de Recherche et de Restauration des Mus├®es de France) are involved in the project. They will be joining forces with NCR, a leading player in database and Data Warehouse technology; Interactive Labs, the new media design and development facility of Italy's leading art publishing group, Giunti; IT Innovation, Web-based system developers; and the Department of Electronics and Computer Science at the University of Southampton. Together they will create web based applications and tools for the automatic indexing and retrieval of high-resolution art images by pictorial content and information."
Type
Electronic Journal
Title
A Spectrum of Interoperability: The Site for Science Prototype for the NSDL
"Currently, NSF is funding 64 projects, each making its own contribution to the library, with a total annual budget of about $24 million. Many projects are building collections; others are developing services; a few are carrying out targeted research.The NSDL is a broad program to build a digital library for education in science, mathematics, engineering and technology. It is funded by the National Science Foundation (NSF) Division of Undergraduate Education. . . . The Core Integration task is to ensure that the NSDL is a single coherent library, not simply a set of unrelated activities. In summer 2000, the NSF funded six Core Integration demonstration projects, each lasting a year. One of these grants was to Cornell University and our demonstration is known as Site for Science. It is at http://www.siteforscience.org/ [Site for Science]. In late 2001, the NSF consolidated the Core Integration funding into a single grant for the production release of the NSDL. This grant was made to a collaboration of the University Corporation for Atmospheric Research (UCAR), Columbia University and Cornell University. The technical approach being followed is based heavily on our experience with Site for Science. Therefore this article is both a description of the strategy for interoperability that was developed for Site for Science and an introduction to the architecture being used by the NSDL production team."
ISBN
1082-9873
Critical Arguements
CA "[T]his article is both a description of the strategy for interoperability that was developed for the [Cornell University's NSF-funded] Site for Science and an introduction to the architecture being used by the NSDL production team."
Phrases
<P1> The grand vision is that the NSDL become a comprehensive library of every digital resource that could conceivably be of value to any aspect of education in any branch of science and engineering, both defined very broadly. <P2> Interoperability among heterogeneous collections is a central theme of the Core Integration. The potential collections have a wide variety of data types, metadata standards, protocols, authentication schemes, and business models. <P3> The goal of interoperability is to build coherent services for users, from components that are technically different and managed by different organizations. This requires agreements to cooperate at three levels: technical, content and organizational. <P4> Much of the research of the authors of this paper aims at . . . looking for approaches to interoperability that have low cost of adoption, yet provide substantial functionality. One of these approaches is the metadata harvesting protocol of the Open Archives Initiative (OAI) . . . <P5> For Site for Science, we identified three levels of digital library interoperability: Federation; Harvesting; Gathering. In this list, the top level provides the strongest form of interoperability, but places the greatest burden on participants. The bottom level requires essentially no effort by the participants, but provides a poorer level of interoperability. The Site for Science demonstration concentrated on the harvesting and gathering, because other projects were exploring federation. <P6> In an ideal world all the collections and services that the NSDL wishes to encompass would support an agreed set of standard metadata. The real world is less simple. . . . However, the NSDL does have influence. We can attempt to persuade collections to move along the interoperability curve. <warrant> <P7> The Site for Science metadata strategy is based on two principles. The first is that metadata is too expensive for the Core Integration team to create much of it. Hence, the NSDL has to rely on existing metadata or metadata that can be generated automatically. The second is to make use of as much of the metadata available from collections as possible, knowing that it varies greatly from none to extensive. Based on these principles, Site for Science, and subsequently the entire NSDL, developed the following metadata strategy: Support eight standard formats; Collect all existing metadata in these formats; Provide crosswalks to Dublin Core; Assemble all metadata in a central metadata repository; Expose all metadata records in the repository for service providers to harvest; Concentrate limited human effort on collection-level metadata; Use automatic generation to augment item-level metadata. <P8> The strategy developed by Site for Science and now adopted by the NSDL is to accumulate metadata in the native formats provided by the collections . . . If a collection supports the protocols of the Open Archives Initiative, it must be able to supply unqualified Dublin Core (which is required by the OAI) as well as the native metadata format. <P9> From a computing viewpoint, the metadata repository is the key component of the Site for Science system. The repository can be thought of as a modern variant of the traditional library union catalog, a catalog that holds comprehensive catalog records from a group of libraries. . . . Metadata from all the collections is stored in the repository and made available to providers of NSDL service.
Conclusions
RQ 1 "Can a small team of librarians manage the collection development and metadata strategies for a very large library?" RQ 2 "Can the NSDL actually build services that are significantly more useful than the general web search services?"
Type
Electronic Journal
Title
Electronic Records Research: Working Meeting May 28-30, 1997
CA Archivists are specifically concerned with records that are not easy to document -- records that are full of secret, proprietary or sensitive information, not to mention hardware and software dependencies. This front end of recordmaking and keeping must be addressed as we define what electronic records are and are not, and how we are to deal with them.
Phrases
<P1> Driven by pragmatism, the University of Pittsburgh team looked for "warrant" in the sources considered authoritative by the practicioners of ancillary professions on whom archivists rely -- lawyers, auditors, IT personnel , etc. (p.3) <P2> If the record creating event and the requirements of 'recordness' are both known, focus shifts to capturing the metadata and binding it to the record contents. (p.7) <P3> A strong business case is still needed to justify the role of archivists in the creation of electronic record management systems. (p.10)
Conclusions
RQ Warrant needs to be looked at in different countries. Does the same core definition of what constitutes a record cut across state borders? What role do specific user needs play in complying to regulation and risk management?
Type
Electronic Journal
Title
The National Digital Information Infrastructure Preservation Program: Expectations, Realities, Choices and Progress to Date
CA The goals of this plan include the continued collecting of materials regardless of evolving digital formats, the long-term preservation of said materials and ensuring access to them for the American people.
Phrases
<P1> There is widespread support for a national initiative in long term preservation of digital content across a very broad range of stakeholder groups outside the traditional scholarly community (p.4) <warrant> <P2> Approaching persistent archiving from the perspective of infrastructure allows system designers to decouple the data storage from the various components that allow users to manage the data. Potentially, any component can be "swapped out" without affecting the rest of the system. Theoretically, many of the technical problems in archiving can be separated into their components, and as innovation occurs, those components can be updated so that the archive remains persistent in the context of rapid change. Similarly, as the storage media obsolesce, the data can be migrated without affecting the overall integrity of the system. (p.7)
Conclusions
RQ Scenario-planning exercises may help expose assumptions that could ultimately be limiting in the future.
The Semantic Web activity is a W3C project whose goal is to enable a 'cooperative' Web where machines and humans can exchange electronic content that has clear-cut, unambiguous meaning. This vision is based on the automated sharing of metadata terms across Web applications. The declaration of schemas in metadata registries advance this vision by providing a common approach for the discovery, understanding, and exchange of semantics. However, many of the issues regarding registries are not clear, and ideas vary regarding their scope and purpose. Additionally, registry issues are often difficult to describe and comprehend without a working example.
ISBN
1082-9873
Critical Arguements
CA "This article will explore the role of metadata registries and will describe three prototypes, written by the Dublin Core Metadata Initiative. The article will outline how the prototypes are being used to demonstrate and evaluate application scope, functional requirements, and technology solutions for metadata registries."
Phrases
<P1> Establishing a common approach for the exchange and re-use of data across the Web would be a major step towards achieving the vision of the Semantic Web. <warrant> <P2> The Semantic Web Activity statement articulates this vision as: 'having data on the Web defined and linked in a way that it can be used for more effective discovery, automation, integration, and reuse across various applications. The Web can reach its full potential if it becomes a place where data can be shared and processed by automated tools as well as by people.' <P3> In parallel with the growth of content on the Web, there have been increases in the amount and variety of metadata to manipulate this content. An inordinate amount of standards-making activity focuses on metadata schemas (also referred to as vocabularies or data element sets), and yet significant differences in schemas remain. <P4> Different domains typically require differentiation in the complexity and semantics of the schemas they use. Indeed, individual implementations often specify local usage, thereby introducing local terms to metadata schemas specified by standards-making bodies. Such differentiation undermines interoperability between systems. <P5> This situation highlights a growing need for access by users to in-depth information about metadata schemas and particular extensions or variations to schemas. Currently, these 'users' are human  people requesting information. <warrant> <P6> It would be helpful to make available easy access to schemas already in use to provide both humans and software with comprehensive, accurate and authoritative information. <warrant> <P7> The W3C Resource Description Framework (RDF) has provided the basis for a common approach to declaring schemas in use. At present the RDF Schema (RDFS) specification offers the basis for a simple declaration of schema. <P8> Even as it stands, an increasing number of initiatives are using RDFS to 'publish' their schemas. <P9> Registries provide 'added value' to users by indexing schemas relevant to a particular 'domain' or 'community of use' and by simplifying the navigation of terms by enabling multiple schemas to be accessed from one view. <warrant> <P10> Additionally, the establishment of registries to index terms actively being used in local implementations facilitates the metadata standards activity by providing implementation experience transferable to the standards-making process. <warrant> <P11> The overriding goal has been the development of a generic registry tool useful for registry applications in general, not just useful for the DCMI. <P12> The formulation of a 'definitive' set of RDF schemas within the DCMI that can serve as the recommended, comprehensive and accurate expression of the DCMI vocabulary has hindered the development of the DCMI registry. To some extent, this has been due to the changing nature of the RDF Schema specification and its W3C candidate recommendation status. However, it should be recognized that the lack of consensus within the DCMI community regarding the RDF schemas has proven to be equally as impeding. <P13> The automated sharing of metadata across applications is an important part of realizing the goal of the Semantic Web. Users and applications need practical solutions for discovering and sharing semantics. Schema registries provide a viable means of achieving this. <warrant>
Conclusions
RQ "Many of the issues regarding metadata registries are unclear and ideas regarding their scope and purpose vary. Additionally, registry issues are often difficult to describe and comprehend without a working example. The DCMI makes use of rapid prototyping to help solve these problems. Prototyping is a process of quickly developing sample applications that can then be used to demonstrate and evaluate functionality and technology."
SOW
DC "New impetus for the development of registries has come with the development activities surrounding creation of the Semantic Web. The motivation for establishing registries arises from domain and standardization communities, and from the knowledge management community." ... "The original charter for the DCMI Registry Working Group was to establish a metadata registry to support the activity of the DCMI. The aim was to enable the registration, discovery, and navigation of semantics defined by the DCMI, in order to provide an authoritative source of information regarding the DCMI vocabulary. Emphasis was placed on promoting the use of the Dublin Core and supporting the management of change and evolution of the DCMI vocabulary." ... "Discussions within the DCMI Registry Working Group (held primarily on the group's mailing list) have produced draft documents regarding application scope and functionality. These discussions and draft documents have been the basis for the development of registry prototypes and continue to play a central role in the iterative process of prototyping and feedback." ... The overall goal of the DCMI Registry Working Group (WG) is to provide a focus for continued development of the DCMI Metadata Registry. The WG will provide a forum for discussing registry-related activities and facilitating cooperation with the ISO 11179 community, the Semantic Web, and other related initiatives on issues of common interest and relevance.
Type
Electronic Journal
Title
Primary Sources, Research, and the Internet: The Digital Scriptorium at Duke
First Monday, Peer Reviewed Journal on the Internet
Publication Year
1997
Volume
2
Issue
9
Critical Arguements
CA "As the digital revolution moves us ever closer to the idea of the 'virtual library,' repositories of primary sources and other archival materials have both a special opportunity and responsibility. Since the materials in their custody are, by definition, often unique, these institutions will need to work very carefully with scholars and other researchers to determine what is the most effective way of making this material accessible in a digital environment."
Phrases
<P1> The matter of Internet access to research materials and collections is not one of simply doing what we have always done -- except digitally. It represents instead an opportunity to rethink the fundamental triangular relationship between libraries and archives, their collections, and their users. <P2> Digital information as it exists on the Internet today requires more navigational, contextual, and descriptive data than is currently provided in traditional card catalogs or their more modern electronic equivalent. One simply cannot throw up vast amounts of textual or image-based data onto the World Wide Web and expect existing search engines to make much sense of it or users to be able to digest the results. ... Archivists and manuscript curators have for many years now been providing just that sort of contextual detail in the guides, finding aids, and indexes that they have traditionally prepared for their holdings. <P3> Those involved in the Berkeley project understood that HTML was essentially a presentational encoding scheme and lacked the formal structural and content-based encoding that SGML would offer. <P4> Encoded Archival Description is quickly moving towards become an internationally embraced standard for the encoding of archival metadata in a wide variety of archival repositories and special collections libraries. And the Digital Scriptorium at Duke has become one of the early implementors of this standard. <warrant>
Conclusions
RQ "Duke is currently involved in a project that is funded through NEH and also involves the libraries of Stanford, the University of Virginia, and the University of California-Berkeley. This project (dubbed the "American Heritage Virtual Digital Archives Project") will create a virtual archive of encoded finding aids from all four institutions. This archive will permit seamless searching of these finding aids -- at a highly granular level of detail -- through a single search engine on one site and will, it is hoped, provide a model for a more comprehensive national system in the near future."
Type
Electronic Journal
Title
Review: Some Comments on Preservation Metadata and the OAIS Model
CA Criticizes some of the limitations of OAIS and makes suggestions for improvements and clarifications. Also suggests that OAIS may be too library-centric, to the determinent of archival and especially recordkeeping needs. "In this article I have tried to articulate some of the main requirements for the records and archival community in preserving (archival) records. Based on this, the conclusion has to be that some adaptations to the [OAIS] model and metadata set would be necessary to meet these requirements. This concerns requirements such as the concept of authenticity of records, information on the business context of records and on relationships between records ('documentary context')."(p. 20)
Phrases
<P1> It requires records managers and archivists (and perhaps other information professionals) to be aware of these differences [in terminology] and to make a translation of such terms to their own domain. (p. 15) <P2> When applying the metadata model for a wider audience, more awareness of the issue of terminology is required, for instance by including clear definitions of key terms. (p. 15) <P3> The extent to which the management of objects can be influenced differs with respect to the type of objects. In the case of (government) records, legislation governs their creation and management, whereas, in the case of publications, the influence will be mostly based on agreements between producers, publishers and preservers. (p. 16) <P4> [A]lthough the suggestion may sometimes be otherwise, preservation metadata do not only apply to what is under the custody of a cultural or other preserving institution, but should be applied to the whole lifecycle of digital objects. ... Preservation can be viewed as part of maintenance. <warrant> (p. 16) <P5> [B]y taking library community needs as leading (albeit implicitly), the approach is already restricting the types of digital objects. Managing different types of 'digital objects', e.g. publications and records, may require not entirely similar sets of metadata. (p. 16) <P6> Another issue is that of the requirements governing the preservation processes. ... There needs to be insight and, as a consequence, also metadata about the preservation strategies, policies and methods, together with the context in which the preservation takes place. <warrant> (p. 16) <P7> [W]hat do we want to preserve? Is it the intellectual content with the functionality it has to have in order to make sense and achieve its purpose, or is it the digital components that are necessary to reproduce it or both? (p. 16-17) <P8> My view is that 'digital objects' should be seen as objects having both conceptual and technical aspects that are closely interrelated. As a consequence of the explanation given above, a digital object may consist of more than one 'digital component'. The definition given in the OAIS model is therefore insufficient. (p. 17) <P9> [W]e have no fewer than five metadata elements that could contain information on what should be rendered and presented on the screen. How all these elements relate to each other, if at all, is unclear. (p. 17) <P10> What we want to achieve ... is that in the future we will still be able to see, read and understand the documents or other information entities that were once produced for a certain purpose and in a certain context. In trying to achieve this, we of course need to preserve these digital components, but, as information technology will evolve, these components have to be migrated or in some cases emulated to be usable on future hard- and software platforms. (p. 17) <P11> I would like to suggest including an element that reflects the original technical environment. (p. 18) <P12> Records, according to the recently published ISO records management standard 15489, are 'information created, received and maintained as evidence and information by an organisation or person, in pursuance of legal obligations or in the transaction of business'. ... The main requirements for records to serve as evidence or authoritative information sources are ... authenticity and integrity, and knowledge about the business context and about the interrelationship between records (e.g. in a case file). <warrant> (p. 18) <P13> It would have been helpful if there had been more acknowledgement of the issue of authenticity and the requirements for it, and if the Working Group had provided some background information about its view and considerations on this aspect and to what extent it is included or not. (p. 19) <P14> In order to be able to preserve (archival) records it will ... be necessary to extend the information model with another class of information that refers to business context. Such a subset could provide a structure for describing what in archival terminology is called information about 'provenance' (with a different meaning from that in OAIS). (p. 19) <P15> In order to accommodate the identified complexity it is necessary to distinguish at least between the following categories of relationships: relationships between intellectual objects ... in the archival context this is referred to as 'documentary context'; relationships between the (structural) components of one intellectual object ... ; [and] relationships between digital components. (p. 19-20) <P16> [T]he issue of appraisal and disposition of records has to be included. In this context the recently published records management standard (ISO 15489) may serve as a useful framework. It would make the OAIS model even more widely applicable. (p. 20)
Conclusions
RQ "There are some issues ... which need further attention. They concern on the one hand the scope and underlying concepts of the OAIS model and the resulting metadata set as presented, and on the other hand the application of the model and metadata set in a records and archival environment. ... [T]he distinction between physical and conceptual or intellectual aspects of a digital object should be made more explicit and will probably have an impact on the model and metadata set also. More attention also needs to be given to the relationship between the (preservation) processes and the metadata. ... In assessing the needs of the records and archival community, the ISO records management standard 15489 may serve as a very useful framework. Such an exercise would also include a test for applicability of the model and metadata set for record-creating organisations and, as such, broaden the view of the OAIS model." (p. 20)
SOW
DC OAIS emerged out of an initiative spearheaded by NASA's Consultative Committee for Space Data Systems. It has been shaped and promoted by the RLG and OCLC. Several international projects have played key roles in shaping the OAIS model and adapting it for use in libraries, archives and research repositories. OAIS-modeled repositories include the CEDARS Project, Harvard's Digital Repository, Koninklijke Bibliotheek (KB), the Library of Congress' Archival Information Package for audiovisual materials, MIT's D-Space, OCLC's Digital Archive and TERM: the Texas Email Repository Model.
Type
Electronic Journal
Title
Computer Records and the Federal Rules of Evidence
See also U.S. Federal Rules of Evidence. Rule 803. Hearsay Exceptions; Availability of Declarant Immaterial.
Publisher
U.S. Department of Justice Executive Office for United States Attorneys
Critical Arguements
CA "This article explains some of the important issues that can arise when the government seeks the admission of computer records under the Federal Rules of Evidence. It is an excerpt of a larger DOJ manual entitled 'Searching and Seizing Computers and Obtaining Electronic Evidence in Criminal Investigations,' which is available on the internet at www.cybercrime.gov/searchmanual.htm." Cites cases dealing with Fed. R. Evid. 803(6).
Phrases
<P1>Most federal courts that have evaluated the admissibility of computer records have focused on computer records as potential hearsay. The courts generally have admitted computer records upon a showing that the records fall within the business records exception, Fed. R. Evid. 803(6). <P2> See, e.g., United States v. Cestnik, 36 F.3d 904, 909-10 (10th Cir. 1994); United States v. Moore, 923 F.2d 910, 914 (1st Cir. 1991); United States v. Briscoe, 896 F.2d 1476, 1494 (7th Cir. 1990); United States v. Catabran, 836 F.2d 453, 457 (9th Cir. 1988); Capital Marine Supply v. M/V Roland Thomas II, 719 F.2d 104, 106 (5th Cir. 1983). <P3> Applying this test, the courts have indicated that computer records generally can be admitted as business records if they were kept pursuant to a routine procedure for motives that tend to assure their accuracy. <warrant>
Conclusions
RQ "The federal courts are likely to move away from this 'one size fits all' approach as they become more comfortable and familiar with computer records. Like paper records, computer records are not monolithic: the evidentiary issues raised by their admission should depend on what kind of computer records a proponent seeks to have admitted. For example, computer records that contain text often can be divided into two categories: computer-generated records, and records that are merely computer-stored. See People v. Holowko, 486 N.E.2d 877, 878-79 (Ill. 1985). The difference hinges upon whether a person or a machine created the records' contents. ... As the federal courts develop a more nuanced appreciation of the distinctions to be made between different kinds of computer records, they are likely to see that the admission of computer records generally raises two distinct issues. First, the government must establish the authenticity of all computer records by providing 'evidence sufficient to support a finding that the matter in question is what its proponent claims.' Fed. R. Evid. 901(a). Second, if the computer records are computer-stored records that contain human statements, the government must show that those human statements are not inadmissible hearsay."
Type
Electronic Journal
Title
Collection-Based Persistent Digital Archives - Part 1
The preservation of digital information for long periods of time is becoming feasible through the integration of archival storage technology from supercomputer centers, data grid technology from the computer science community, information models from the digital library community, and preservation models from the archivistÔÇÖs community. The supercomputer centers provide the technology needed to store the immense amounts of digital data that are being created, while the digital library community provides the mechanisms to define the context needed to interpret the data. The coordination of these technologies with preservation and management policies defines the infrastructure for a collection-based persistent archive. This paper defines an approach for maintaining digital data for hundreds of years through development of an environment that supports migration of collections onto new software systems.
ISBN
1082-9873
Critical Arguements
CA "Supercomputer centers, digital libraries, and archival storage communities have common persistent archival storage requirements. Each of these communities is building software infrastructure to organize and store large collections of data. An emerging common requirement is the ability to maintain data collections for long periods of time. The challenge is to maintain the ability to discover, access, and display digital objects that are stored within an archive, while the technology used to manage the archive evolves. We have implemented an approach based upon the storage of the digital objects that comprise the collection, augmented with the meta-data attributes needed to dynamically recreate the data collection. This approach builds upon the technology needed to support extensible database schema, which in turn enables the creation of data handling systems that interconnect legacy storage systems."
Phrases
<P1> The ultimate goal is to preserve not only the bits associated with the original data, but also the context that permits the data to be interpreted. <warrant> <P2> We rely on the use of collections to define the context to associate with digital data. The context is defined through the creation of semi-structured representations for both the digital objects and the associated data collection. <P3>A collection-based persistent archive is therefore one in which the organization of the collection is archived simultaneously with the digital objects that comprise the collection. <P4> The goal is to preserve digital information for at least 400 years. This paper examines the technical issues that must be addressed and presents a prototype implementation. <P5>Digital object representation. Every digital object has attributes that define its structure, physical context, and provenance, and annotations that describe features of interest within the object. Since the set of attributes (such as annotations) will vary across all objects within a collection, a semi-structured representation is needed. Not all digital objects will have the same set of associated attributes. <P6> If possible, a common information model should be used to reference the attributes associated with the digital objects, the collection organization, and the presentation interface. An emerging standard for a uniform data exchange model is the eXtended Markup Language (XML). <P7> A particular example of an information model is the XML Document Type Definition (DTD) which provides a description for the allowed nesting structure of XML elements. Richer information models are emerging such as XSchema (which provides data types, inheritance, and more powerful linking mechanisms) and XMI (which provides models for multiple levels of data abstraction). <P8> Although XML DTDs were originally applied to documents only, they are now being applied to arbitrary digital objects, including the collections themselves. More generally, OSDs can be used to define the structure of digital objects, specify inheritance properties of digital objects, and define the collection organization and user interface structure. <P9> A persistent collection therefore needs the following components of an OSD to completely define the collection context: Data dictionary for collection semantics; Digital object structure; Collection structure; and User interface structure. <P10> The re-creation or instantiation of the data collection is done with a software program that uses the schema descriptions that define the digital object and collection structure to generate the collection. The goal is to build a generic program that works with any schema description. <P11> The information for which driver to use for access to a particular data set is maintained in the associated Meta-data Catalog (MCAT). The MCAT system is a database containing information about each data set that is stored in the data storage systems. <P12> The data handling infrastructure developed at SDSC has two components: the SDSC Storage Resource Broker (SRB) that provides federation and access to distributed and diverse storage resources in a heterogeneous computing environment, and the Meta-data Catalog (MCAT) that holds systemic and application or domain-dependent meta-data about the resources and data sets (and users) that are being brokered by the SRB. <P13> A client does not need to remember the physical mapping of a data set. It is stored as meta-data associated with the data set in the MCAT catalog. <P14> A characterization of a relational database requires a description of both the logical organization of attributes (the schema), and a description of the physical organization of attributes into tables. For the persistent archive prototype we used XML DTDs to describe the logical organization. <P15> A combination of the schema and physical organization can be used to define how queries can be decomposed across the multiple tables that are used to hold the meta-data attributes. <P16> By using an XML-based database, it is possible to avoid the need to map between semi-structured and relational organizations of the database attributes. This minimizes the amount of information needed to characterize a collection, and makes the re-creation of the database easier. <warrant> <P17> Digital object attributes are separated into two classes of information within the MCAT: System-level meta-data that provides operational information. These include information about resources (e.g., archival systems, database systems, etc., and their capabilities, protocols, etc.) and data objects (e.g., their formats or types, replication information, location, collection information, etc.); Application-dependent meta-data that provides information specific to particular data sets and their collections (e.g., Dublin Core values for text objects). <P18> Internally, MCAT keeps schema-level meta-data about all of the attributes that are defined. The schema-level attributes are used to define the context for a collection and enable the instantiation of the collection on new technology. <P19> The logical structure should not be confused with database schema and are more general than that. For example, we have implemented the Dublin Core database schema to organize attributes about digitized text. The attributes defined in the logical structure that is associated with the Dublin Core schema contains information about the subject, constraints, and presentation formats that are needed to display the schema along with information about its use and ownership. <P20> The MCAT system supports the publication of schemata associated with data collections, schema extension through the addition or deletion of new attributes, and the dynamic generation of the SQL that corresponds to joins across combinations of attributes. <P21> By adding routines to access the schema-level meta-data from an archive, it is possible to build a collection-based persistent archive. As technology evolves and the software infrastructure is replaced, the MCAT system can support the migration of the collection to the new technology.
Conclusions
RQ Collection-Based Persistent Digital Archives - Part 2
SOW
DC "The technology proposed by SDSC for implementing persistent archives builds upon interactions with many of these groups. Explicit interactions include collaborations with Federal planning groups, the Computational Grid, the digital library community, and individual federal agencies." ... "The data management technology has been developed through multiple federally sponsored projects, including the DARPA project F19628-95-C-0194 "Massive Data Analysis Systems," the DARPA/USPTO project F19628-96-C-0020 "Distributed Object Computation Testbed," the Data Intensive Computing thrust area of the NSF project ASC 96-19020 "National Partnership for Advanced Computational Infrastructure," the NASA Information Power Grid project, and the DOE ASCI/ASAP project "Data Visualization Corridor." Additional projects related to the NSF Digital Library Initiative Phase II and the California Digital Library at the University of California will also support the development of information management technology. This work was supported by a NARA extension to the DARPA/USPTO Distributed Object Computation Testbed, project F19628-96-C-0020."
Type
Electronic Journal
Title
Collection-Based Persistent Digital Archives - Part 2
"Collection-Based Persistent Digital Archives: Part 2" describes the creation of a one million message persistent E-mail collection. It discusses the four major components of a persistent archive system: support for ingestion, archival storage, information discovery, and presentation of the collection. The technology to support each of these processes is still rapidly evolving, and opportunities for further research are identified.
ISBN
1082-9873
Critical Arguements
CA "The multiple migration steps can be broadly classified into a definition phase and a loading phase. The definition phase is infrastructure independent, whereas the loading phase is geared towards materializing the processes needed for migrating the objects onto new technology. We illustrate these steps by providing a detailed description of the actual process used to ingest and load a million-record E-mail collection at the San Diego Supercomputer Center (SDSC). Note that the SDSC processes were written to use the available object-relational databases for organizing the meta-data. In the future, it may be possible to go directly to XML-based databases."
Phrases
<P1> The processes used to ingest a collection, transform it into an infrastructure independent form, and store the collection in an archive comprise the persistent storage steps of a persistent archive. The processes used to recreate the collection on new technology, optimize the database, and recreate the user interface comprise the retrieval steps of a persistent archive. <P2> In order to build a persistent collection, we consider a solution that "abstracts" all aspects of the data and its preservation. In this approach, data object and processes are codified by raising them above the machine/software dependent forms to an abstract format that can be used to recreate the object and the processes in any new desirable forms. <P3> The SDSC infrastructure uses object-relational databases to organize information. This makes data ingestion more complex by requiring the mapping of the XML DTD semi-structured representation onto a relational schema. <P4> The SDSC infrastructure uses object-relational databases to organize information. This makes data ingestion more complex by requiring the mapping of the XML DTD semi-structured representation onto a relational schema. <P5> The steps used to store the persistent archive were: (1) Define Digital Object: define meta-data, define object structure (OBJ-DTD) --- (A), define object DTD to object DDL mapping --- (B) (2) Define Collection: define meta-data, define collection structure (COLL-DTD) --- (C), define collection DTD structure to collection DDL mapping --- (D) (3) Define Containers: define packing format for encapsulating data and meta-data (examples are the AIP standard, Hierarchical Data Format, Document Type Definition) <P5> In the ingestion phase, the relational and semi-structured organization of the meta-data is defined. No database is actually created, only the mapping between the relational organization and the object DTD. <P6> Note that the collection relational organization does not have to encompass all of the attributes that are associated with a digital object. Separate information models are used to describe the objects and the collections. It is possible to take the same set of digital objects and form a new collection with a new relational organization. <P7> Multiple communities across academia, the federal government, and standards groups are exploring strategies for managing very large archives. The persistent archive community needs to maintain interactions with these communities to track development of new strategies for data management and storage. <warrant> <P8>
Conclusions
RQ "The four major components of the persistent archive system are support for ingestion, archival storage, information discovery, and presentation of the collection. The first two components focus on the ingestion of data into collections. The last two focus on access to the resulting collections. The technology to support each of these processes is still rapidly evolving. Hence consensus on standards has not been reached for many of the infrastructure components. At the same time, many of the components are active areas of research. To reach consensus on a feasible collection-based persistent archive, continued research and development is needed. Examples of the many related issues are listed below:
Type
Electronic Journal
Title
Buckets: A new digital technology for preserving NASA research
CA Buckets are information objects designed to reduce dependency on traditional archives and database systems thereby making them more resilent to the transient nature of information systems.
Phrases
Another focus of aggregation was including the metadata with data. Through experiences NASA researchers found that metadata tended to "drift" over time, becoming decoupled from the data it described or locked in specific DL systems and hard to extract or share with other systems. (p. 377) Buckets are designed to imbue the information objects with certain responsibilities, such as display, dissemination, protection, and maintenance of its contents. As such, buckets should be able to work with many DL systems simultaneously, and minimize or eliminate the necessary modification of DL systems to work with buckets. Ideally, buckets should work with everything and break nothing. This philosophy is formalized in the SODA DL model. the objects become "smarter" at the expense of the archives (that become "dumber"), as functionalities generally associated with archives are moved into the data objects themselves. (p. 390)
Conclusions
RQ The creation of high quality tools for bucket creation and administration is absolutely necessary. The extension of authentication and security measures is key to supporting more technologies. Many applications of this sort of information object independence remains to be explored.
Type
Electronic Journal
Title
Search for Tomorrow: The Electronic Records Research Program of the U.S. National Historical Publications and Records Commission
The National Historical Publications and Records Commission (NHPRC) is a small grant-making agency affiliated with the U.S. National Archives and Records Administration. The Commission is charged with promoting the preservation and dissemination of documentary source materials to ensure an understanding of U.S. history. Recognizing that the increasing use of computers created challenges for preserving the documentary record, the Commission adopted a research agenda in 1991 to promote research and development on the preservation and continued accessibility of documentary materials in electronic form. From 1991 to the present the Commission awarded 31 grants totaling $2,276,665 for electronic records research. Most of this research has focused on two issues of central concern to archivists: (1) electronic record keeping (tools and techniques to manage electronic records produced in an office environment, such as word processing documents and electronic mail), and (2) best practices for storing, describing, and providing access to all electronic records of long-term value. NHPRC grants have raised the visibility of electronic records issues among archivists. The grants have enabled numerous archives to begin to address electronic records problems, and, perhaps most importantly, they have stimulated discussion about electronic records among archivists and records managers.
Publisher
Elsevier Science Ltd
Critical Arguements
CA "The problem of maintaining electronic records over time is big, expensive, and growing. A task force on digital archives established by the Commission on Preservation and Access in 1994 commented that the life of electronic records could be characterized in the same words Thomas Hobbes once used to describe life: ÔÇ£nasty, brutish, and shortÔÇØ [1]. Every day, thousands of new electronic files are created on federal, state, and local government computers across the nation. A small but important portion of these records will be designated for permanent retention. Government agencies are increasingly relying on computers to maintain information such as census files, land titles, statistical data, and vital records. But how should electronic records with long-term value be maintained? Few government agencies have developed comprehensive policies for managing current electronic records, much less preserving those with continuing value for historians and other researchers. Because of this serious and growing problem, the National Historical Publications and Records Commission (NHPRC), a small grantmaking agency affiliated with the U.S. National Archives and Records Administration (NARA), has been making grants for research and development on the preservation and use of electronic documentary sources. The program is conducted in concert with NARA, which in 1996 issued a strategic plan that gives high priority to mastering electronic records problems in partnership with federal government agencies and the NHPRC.
Phrases
<P1> How can data dictionaries, information resource directory systems, and other metadata systems be used to support electronic records management and archival requirements? <P2> In spite of the number of projects the Commission has supported, only four questions from the research agenda have been addressed to date. Of these, the question relating to requirements for the development of data dictionaries and other metadata systems (question number four) has produced a single grant for a state information locator system in South Carolina, and the question relating to needs for archival education (question 10) has led to two grants to the Society of American Archivists for curricular materials. <P3> Information systems created without regard for these considerations may have deficiencies that limit the usefulness of the records contained on them. <warrant> <P4> The NHPRC has awarded major grants to four institutions over the past five years for projects to develop and test requirements for electronic record keeping: University of Pittsburgh (1993): A working set of functional requirements and metadata specifications for electronic record keeping systems; City of Philadelphia (1995, 1996, and 1997): A project to incorporate a subset of the Pittsburgh metadata specifications into a new human resources information system and other city systems as test cases and to develop comprehensive record keeping policies and standards for the cityÔÇÖs information technology systems; Indiana University (1995): A project to develop an assessment tool and methodology for analyzing existing electronic records systems, using the Pittsburgh functional requirements as a model and the student academic record system and a financial system as test cases; Research Foundation of the State University of New York-Albany, Center for Technology in Government (1996): A project to identify best practices for electronic record keeping, including work by the U.S. Department of Defense and the University of British Columbia in addition to the University of Pittsburgh. The Center is working with the stateÔÇÖs Adirondack Parks Agency in a pilot project to develop a system model for incorporating record keeping and archival considerations into the creation of networked computing and communications applications. <P5> No definitive solution has yet been identified for the problems posed by electronic records, although progress has been made in learning what will be needed to design functional electronic record keeping systems. <P6> With the proliferation of digital libraries, the need for long-term storage, migration and retrieval strategies for electronic information has become a priority for a wide variety of information providers. <warrant>
Conclusions
RQ "How best to preserve existing and future electronic formats and provide access to them over time has remained elusive. The answers cannot be found through theoretical research alone, or even through applied research, although both are needed. Answers can only emerge over time as some approaches prove able to stand the test of time and others do not. The problems are large because the costs of maintaining, migrating, and retrieving electronic information continue to be high." ... "Perhaps most importantly, these grants have stimulated widespread discussion of electronic records issues among archivists and record managers, and thus they have had an impact on the preservation of the electronic documentary record that goes far beyond the CommissionÔÇÖs investment."
SOW
DC The National Historic Publications and Records Commission (NHPRC) is the outreach arm of the National Archives and makes plans for and studies issues related to the preservation, use and publication of historical documents. The Commission also makes grants to non-Federal archives and other organizations to promote the preservation use of America's documentary heritage.
Type
Electronic Journal
Title
The Dublin Core Metadata Inititiative: Mission, Current Activities, and Future Directions
Metadata is a keystone component for a broad spectrum of applications that are emerging on the Web to help stitch together content and services and make them more visible to users. The Dublin Core Metadata Initiative (DCMI) has led the development of structured metadata to support resource discovery. This international community has, over a period of 6 years and 8 workshops, brought forth: A core standard that enhances cross-disciplinary discovery and has been translated into 25 languages to date; A conceptual framework that supports the modular development of auxiliary metadata components; An open consensus building process that has brought to fruition Australian, European and North American standards with promise as a global standard for resource discovery; An open community of hundreds of practitioners and theorists who have found a common ground of principles, procedures, core semantics, and a framework to support interoperable metadata.
Type
Report
Title
Mapping of the Encoded Archival Description DTD Element Set to the CIDOC CRM
The CIDOC CRM is the first ontology designed to mediate contents in the area of material cultural heritage and beyond, and has been accepted by ISO TC46 as work item for an international standard. The EAD Document Type Definition (DTD) is a standard for encoding archival finding aids using the Standard Generalized Markup Language (SGML). Archival finding aids are detailed guides to primary source material which provide fuller information than that normally contained within cataloging records. 
Publisher
Institute of Computer Science, Foundation for Research and Technology - Hellas
Publication Location
Heraklion, Crete, Greece
Language
English
Critical Arguements
CA "This report describes the semantic mapping of the current EAD DTD Version 1.0 Element Set to the CIDOC CRM and its latest extension. This work represents a proof of concept for the functionality the CIDOC CRM is designed for." 
Conclusions
RQ "Actually, the CRM seems to do the job quite well ÔÇô problems in the mapping arise more from underspecification in the EAD rather than from too domain-specific notions. "┬á... "To our opinion, the archival community could benefit from the conceptualizations of the CRM to motivate more powerful metadata standards with wide interoperability in the future, to the benefit of museums and other disciplines as well."
SOW
DC "As a potential international standard, the EAD DTD is maintained in the Network Development and MARC Standards Office of the Library of Congress in partnership with the Society of American Archivists." ... "The CIDOC Conceptual Reference Model (see [CRM1999], [Doerr99]), in the following only referred to as ┬½CRM┬╗, is outcome of an effort of the Documentation Standards Group of the CIDOC Committee (see ┬½http:/www.cidoc.icom.org┬╗, ÔÇ£http://cidoc.ics.forth.grÔÇØ) of ICOM, the International Council of Museums beginning in 1996."
Type
Report
Title
RLG Best Practice Guidelines for Encoded Archival Description
These award-winning guidelines, released in August 2002, were developed by the RLG EAD Advisory Group to provide practical, community-wide advice for encoding finding aids. They are designed to: facilitate interoperability of resource discovery by imposing a basic degree of uniformity on the creation of valid EAD-encoded documents; encourage the inclusion of particular elements, and; develop a set of core data elements. 
Publisher
Research Libraries Group
Publication Location
Mountain View, CA, USA
Language
English
Critical Arguements
<CA> The objectives of the guidelines are: 1. To facilitate interoperability of resource discovery by imposing a basic degree of uniformity on the creation of valid EAD-encoded documents and to encourage the inclusion of elements most useful for retrieval in a union index and for display in an integrated (cross-institutional) setting; 2. To offer researchers the full benefits of XML in retrieval and display by developing a set of core data elements to improve resource discovery. It is hoped that by identifying core elements and by specifying "best practice" for those elements, these guidelines will be valuable to those who create finding aids, as well as to vendors and tool builders; 3. To contribute to the evolution of the EAD standard by articulating a set of best practice guidelines suitable for interinstitutional and international use. These guidelines can be applied to both retrospective conversion of legacy finding aids and the creation of new finding aids.  
Conclusions
<RQ>
SOW
<DC> "RLG organized the EAD working group as part of our continuing commitment to making archival collections more accessible on the Web. We offer RLG Archival Resources, a database of archival materials; institutions are encouraged to submit their finding aids to this database." ... "This set of guidelines, the second version promulgated by RLG, was developed between October 2001 and August 2002 by the RLG EAD Advisory Group. This group consisted of ten archivists and digital content managers experienced in creating and managing EAD-encoded finding aids at repositories in the United States and the United Kingdom."
1. Also at http://www.access.gpo.gov/uscode/title28a/28a_8_6_.html2. As amended Feb. 28, 1966, eff. July 1, 1966; Mar. 2, 1987, eff. Aug. 1, 1987; Apr. 30, 1991, eff. Dec. 1, 1991.
Type
Web Page
Title
The Electronic Records Strategies Task Force Report: An Australian Perspective
CA The archival profession has a brief window of opportunity to become stakeholders in the realm of electronic records. In order to accomplish that, they must answer not only the "what" but the "why" of recordkeeping in all of its implications.
Conclusions
RQ How will American archivists deal with the re-invention of professional roles that have traditionally been bifurcated by records on one side and archives on the other? Where does continuum thinking leave SAA and its primary constituency of historical archivists?
This portal page provides links to all EAD-related information as applicable to those institutional members of the U.K. Archives Hub. It provides links to Creating EAD records, More about EAD, Reference, and More resources.
Publisher
The Archives Hub
Publication Location
Manchester, England, U.K.
Language
English
Critical Arguements
CA "These pages have been designed to hold links and information which we hope will be useful to archivists and librarians working in the UK Higher and Further Education sectors."
SOW
DC The Archives Hub provides a single point of access to 17,598 descriptions of archives held in UK universities and colleges. At present these are primarily at collection-level, although complete catalogue descriptions are provided where they are available. The Archives Hub forms one part of the UK's National Archives Network, alongside related networking projects. A Steering Committee which includes representatives of contributing institutions, the National Archives and the other archive networks guides the progress of the project. There is also a Contributors' and Users' Forum which provides feedback to aid the development of the service. The service is hosted at MIMAS on behalf of the Consortium of University Research Libraries (CURL) and is funded by the Joint Information Systems Committee (JISC). Systems development work is undertaken at the University of Liverpool.
Type
Web Page
Title
An Assessment of Options for Creating Enhanced Access to Canada's Audio-Visual Heritage
CA "This project was conducted by Paul Audley & Associates to investigate the feasibility of single window access to information about Canada's audio-visual heritage. The project follows on the recommendations of Fading Away, the 1995 report of the Task Force on the Preservation and Enhanced Use of Canada's Audio-Visual Heritage, and the subsequent 1997 report Search + Replay. Specific objectives of this project were to create a profile of selected major databases of audio-visual materials, identify information required to meet user needs, and suggest models for single-window access to audio-visual databases. Documentary research, some 35 interviews, and site visits to organizations in Vancouver, Toronto, Ottawa and Montreal provided the basis upon which the recommendations of this report were developed."
Type
Web Page
Title
JISC/NPO studies on the preservation of electronic materials: A framework of data types and formats, and issues affecting the long term preservation of digital material
CA Proposes a framework for preserving digital objects and discusses steps in the preservation process. Addresses a series of four questions: Why preserve? How much? How? And Where? Proposes a "Preservation Complexity Scorecard" to help identify the complexity of preservation needs and the appropriate preservation approach for a given object. "Although a great deal has been discussed and written about digital material preservation, there would appear to be no overall structure which brings together the findings of the numerous contributors to the debate, and allows them to be compared. This Report attempts to provide such a structure, whereby it should be possible to identify the essential elements of the preservation debate and to determine objectively the criticality of the other unresolved issues. This Report attempts to identify the most critical issues and employ them in order to determine their affect [sic] on preservation practice." (p. 5)
Conclusions
RQ "The study concludes that the overall management task in long term preservation is to moderate the pressure to preserve (Step 1) with the constraints dictated by a cost-effective archive (Step 3). This continuing process of moderation is documented through the Scorecard." (p. 6) "The Study overall recommends that a work programme should be started to: (a) Establish a Scorecard approach (to measure preservation complexity), (b) Establish an inventory of archive items (with complexity ratings) and (c) Establish a Technology Watch (to monitor shifts in technology), in order to be able to manage technological change. And in support of this, (a) establish a programme of work to explore the interaction of stakeholders and a four level contextual mode in the preservation process." (p. 6) A four level contextual approach, with data dictionary entry definitions, should be built in order to provide an information structure that will permit the successful retrieval and interpretation of an object in 50 years time. A study should be established to explore the principle of encapsulating documentsusing the four levels of context, stored in a format, possibly encrypted, that can be transferred across technologies and over time. <warrant> (p. 31) A more detailed study should be made of the inter-relationships of the ten stakeholders, and how they can be made to support the long term preservation of digital material. This will be linked to the economics of archive management (the cost model), changes in legislation (Legal Deposit, etc.), the risks of relying on links between National Libraries to maintain collections (threats of wholesale destruction of collections), and loss through viruses (technological turbulence). (p. 36) A technology management trail (within the Scorecard -- see Step 2 of the Framework) should be established before the more complex digital material is stored. This is to ensure that, for an item of digital material, the full extent of the internal interrelationships are understood, and the implications for long term preservation in a variety of successive environments are documented. (p. 37)
SOW
DC "The study is part of a wider programme of studies, funded by the Joint Information Systems Committee ("JISC"). The programme was initiated as a consequence of a two day workshop at Warwick University, in late November 1995. The workshop addressed the Long Term Preservation of Electronic Materials. The attendees represented an important cross-section of academic, librarian, curatorial, managerial and technological interests. 18 potential action points emerged, and these were seen as a basis for initiating further activity. After consultation, JISC agreed to fund a programme of studies." (p. 7) "The programme of studies is guided by the Digital Archive Working Group, which reports to the Management Committee of the National Preservation Office. The programme is administered by the British Library Research and Innovation Centre." (p. 2)
CA Discussion of the challenges faced by librarians and archivists who must determine which and how much of the mass amounts of digitally recorded sound materials to preserve. Identifies various types of digital sound formats and the varying standards to which they are created. Specific challenges discussed include copyright issues; technologies and platforms; digitization and preservation; and metadata and other standards.
Conclusions
RQ "Whether between record companies and archives or with others, some type of collaborative approach to audio preservation will be necessary if significant numbers of audio recordings at risk are to be preserved for posterity. ... One particular risk of preservation programs now is redundancy. ... Inadequate cataloging is a serious impediment to preservation efforts. ... It would be useful to archives, and possibly to intellectual property holders as well, if archives could use existing industry data for the bibliographic control of published recordings and detailed listings of the music recorded on each disc or tape. ... Greater collaboration between libraries and the sound recording industry could result in more comprehensive catalogs that document recording sessions with greater specificity. With access to detailed and authoritative information about the universe of published sound recordings, libraries could devote more resources to surveying their unpublished holdings and collaborate on the construction of a preservation registry to help reduce preservation redundancy. ... Many archivists believe that adequate funding for preservation will not be forthcoming unless and until the recordings preserved can be heard more easily by the public. ... If audio recordings that do not have mass appeal are to be preserved, that responsibility will probably fall to libraries and archives. Within a partnership between archives and intellectual property owners, archives might assume responsibility for preserving less commercial music in return for the ability to share files of preserved historical recordings."
Type
Web Page
Title
CDL Digital Object Standard: Metadata, Content and Encoding
This document addresses the standards for digital object collections for the California Digital Library 1. Adherence to these standards is required for all CDL contributors and may also serve University of California staff as guidelines for digital object creation and presentation. These standards are not intended to address all of the administrative, operational, and technical issues surrounding the creation of digital object collections.
Critical Arguements
CA These standards describe the file formats, storage and access standards for digital objects created by or incorporated into the CDL as part of the permanent collections. They attempt to balance adherence to industry standards, reproduction quality, access, potential longevity and cost.
Conclusions
RQ not applicable
SOW
DC "This is the first version of the CDL Digital Object Standard. This version is based upon the September 1, 1999 version of the CDL's Digital Image Standard, which included recommendations of the Museum Educational Site Licensing Project (MESL), the Library of Congress and the MOA II participants." ... "The Museum Educational Site Licensing Project (MESL) offered a framework for seven collecting institutions, primarily museums, and seven universities to experiment with new ways to distribute visual information--both images and related textual materials. " ... "The Making of America (MoA II) Testbed Project is a Digital Library Federation (DLF) coordinated, multi-phase endeavor to investigate important issues in the creation of an integrated, but distributed, digital library of archival materials (i.e., digitized surrogates of primary source materials found in archives and special collections). The participants include Cornell University, New York Public Library, Pennsylvania State University, Stanford University and UC Berkeley. The Library of Congress white papers and standards are based on the experience gained during the American Memory Pilot Project. The concepts discussed and the principles developed still guide the Library's digital conversion efforts, although they are under revision to accomodate the capabilities of new technologies and new digital formats." ... "The CDL Technical Architecture and Standards Workgroup includes the following members with extensive experience with digital object collection and management: Howard Besser, MESL and MOA II digital imaging testbed projects; Diane Bisom, University of California, Irvine; Bernie Hurley, MOA II, University of California, Berkeley; Greg Janee, Alexandria Digital Library; John Kunze, University of California, San Francisco; Reagan Moore and Chaitanya Baru, San Diego Supercomputer Center, ongoing research with the National Archives and Records Administration on the long term storage and retrieval of digital content; Terry Ryan, University of California, Los Angeles; David Walker, California Digital Library"
This is one of a series of guides produced by the Cedars digital preservation project. This guide concentrates on the technical approaches that Cedars recommends as a result of its experience. The accent is on preservation, without which continued access is not possible. The time scale is at least decades, i.e. way beyond the lifetime of any hardware technology. The overall preservation strategy is to remove the data from its medium of acquisition and to preserve the digital content as a stream of bytes. There is good reason to be confident that data held as a stream of bytes can be preserved indefinitely. Just as there is no access without preservation, preservation with no prospect of future access is a very sterile exercise. As well as preserving the data as a byte-stream, Cedars adds in metadata. This includes reference to facilities (called technical metadata in this document) for accessing the intellectual content of the preserved data. This technical metadata will usually include actual software for use in accessing the data. It will be stored as a preserved object in the overall archive store, and will be revised as technology evolves making new methods of access to preserved objects appropriate. There will be big economies of scale, as most, if not all, objects of the same type will share the same technical metadata. Cedars recommends against repeated format conversions, and instead argues for keeping the preserved byte-stream, while tracking evolving technology by maintaining the technical metadata. It is for this reason that Cedars includes only a reference to the technical metadata in the preserved data object. Thus future users of the object will be pointed to information appropriate to their own era, rather than that of the object's preservation. The monitoring and updating of this aspect of the technical metadata is a vital function of the digital library. In practice, Cedars expects that very many preserved digital objects will be in the same format, and will reference the same technical metadata. Access to a preserved object then involves Migration on Request, in that any necessary migration from an obsolete format to an appropriate current day format happens at the point of request. As well as recommending actions to be taken to preserve digital objects, Cedars also recommends the use of a permanent naming scheme, with a strong recommendation that such a scheme should be infinitely extensible.
Critical Arguements
CA "This document is intended to inform technical practitioners in the actual preservation of digital materials, and also to highlight to library management the importance of this work as continuing their traditional scholarship role into the 21st century."
This document provides some background on preservation metadata for those interested in digital preservation. It first attempts to explain why preservation metadata is seen as an essential part of most digital preservation strategies. It then gives a broad overview of the functional and information models defined in the Reference Model for an Open Archival Information System (OAIS) and describes the main elements of the Cedars outline preservation metadata specification. The next sections take a brief look at related metadata initiatives, make some recommendations for future work and comment on cost issues. At the end there are some brief recommendations for collecting institutions and the creators of digital content followed by some suggestions for further reading.
Critical Arguements
CA "This document is intended to provide a brief introduction to current preservation metadata developments and introduce the outline metadata specifications produced by the Cedars project. It is aimed in particular at those who may have responsibility for digital preservation in the UK further and higher education community, e.g. senior staff in research libraries and computing services. It should also be useful for those undertaking digital content creation (digitisation) initiatives, although it should be noted that specific guidance on this is available elsewhere. The guide may also be of interest to other kinds of organisations that have an interest in the long-term management of digital resources, e.g. publishers, archivists and records managers, broadcasters, etc. This document aimes to provide: A rationale for the creation and maintenance of preservation metadata to support digital preservation strategies, e.g. migration or emulation; An introduction to the concepts and terminology used in the influential ISO Reference Model for an Open Archival Information System (OAIS); Brief information on the Cedars outline preservation metadata specification and the outcomes of some related metadata initiatives; Some notes on the cost implications of preservation metadata and how these might be reduced.
Conclusions
RQ "In June 2000, a group of archivists, computer scientists and metadata experts met in the Netherlands to discuss metadata developments related to recordkeeping and the long-term preservation of archives. One of the key conclusions made at this working meeting was that the recordkeeping metadata communities should attempt to co-operate more with other metatdata initiatives. The meeting also suggested research into the contexts of creation and use, e.g. identifying factors that might encourage or discourage creators form meeting recordkeeping metadata requirements. This kind of research would also be useful for wider preservation metadata developments. One outcome of this meeting was the setting up of an Archiving Metadata Forum (AMF) to form the focus of future developments." ... "Future work on preservation metadata will need to focus on several key issues. Firstly, there is an urgent need for more practical experience of undertaking digital preservation strategies. Until now, many preservation metadata initiatives have largely been based on theoretical considerations or high-level models like the OAIS. This is not in itself a bad thing, but it is now time to begin to build metadata into the design of working systems that can test the viability of digital preservation strategies in a variety of contexts. This process has already begun in initiatives like the Victorian Electronic Records Stategy and the San Diego Supercomputer Center's 'self-validating knowledge-based archives'. A second need is for increased co-operation between the many metadata initiatives that have an interest in digital preservation. This may include the comparison and harmonisation of various metadata specifications, where this is possible. The OCLC/LG working group is an example of how this has been taken forward whitin a particular domain. There is a need for additional co-operation with recordkeeping metadata specialists, computing scientists and others in the metadata research community. Thirdly, there is a need for more detailed research into how metadata will interact with different formats, preservation strategies and communities of users. This may include some analysis of what metadata could be automatically extracted as part of the ingest process, an investigation of the role of content creators in metadata provision, and the production of user requirements." ... "Also, thought should be given to the development of metadata standards that will permit the easy exchange of preservation metadata (and information packages) between repositories." ... "As well as ensuring that digital repositories are able to facilitate the automatic capture of metadata, some thought should also be given to how best digital repositories could deal with any metadata that might already exist."
SOW
DC "Funded by JISC (the Joint Information Systems Committee of the UK higher education funding councils), as part of its Electronic Libraries (eLib) Programme, Cedars was the only project in the programme to focus on digital preservation." ... "In the digitial library domain, the development of a recommendation on preservation metadata is being co-ordinated by a working group supported by OCLC and the RLG. The membership of the working group is international, and inlcudes key individuals who were involved in the development of the Cedars, NEDLIB and NLA metadata specifications."
Type
Web Page
Title
Deliberation No. 11/2004 of 19 February 2004: "Technical Rules for Copying and Preserving Electronic Documents on Digital Media which are Suitable to Guarantee Authentic Copies"
CA Recognizes that preservation of authentic electronic records means preservation of authentic/true copies. Thus the preservation process is called substitute preservation, and the authenticity of a preserved document is not established on the object itself (as it was with traditional media), but through the authority of the preserver (and possibly a notary), who would attest to the identity and integrity of the whole of the reproduced documents every time a migration occurs. The preserver's task list is also noteworthy. Archival units description stands out as an essential activity (not replaceable by the metadata which are associated to each single document) in order to maintain intellectual control over holdings.
SOW
DC CNIPA (Centro Nazionale per l'Informatica nella Pubblica Amministrazione) replaced AIPA (Autorita' per l'Informatica nella Pubblica Amministrazione) in 2003. Such an Authority (established in 1993 according to art. 4 of the Legislative Decree 39/1993, as amended by art. 176 of the Legislative Decree 196/2003) operates as a branch of the Council of Ministers' Presidency with the mandate to put the Ministry for Innovation and Technologies' policies into practice. In particular, CNIPA is responsible for bringing about reforms relevant to PA's modernization, the spread of e-government and the development of nationwide networks to foster better communication among public offices and between citizens and the State. In the Italian juridical system, CNIPA's deliberations have a lower enabling power, but they nevertheless are part of the State's body of laws. The technical rules provided in CNIPA's deliberation 11/2004 derive from art. 6, par. 2 of the DPR 445/2000, which says: "Preservation obligations are fully satisfied, both for administrative and probative purposes, also with the use of digital media when the employed procedures comply with the technical rules provided by AIPA." In order to keep those rules up to date according to the latest technology, AIPA's deliberation no. 42 of 13 December 2001 on "Technical rules for documents reproduction and preservation on digital media that are suitable to guarantee true copies of the original documents" has been replaced by the current CNIPA deliberation.
This document outlines the best practices guidelines for creation of EAD-encoded finding aids for submission to the Archives Hub in the U.K. It includes sections on Mandatory Fields, Access Points, Manual Encoding, Multilevel Descriptions, Saving and Submitting Files, and Links.
Notes
This is a downloadable .pdf file. Also available in Rich Text Format (.rtf).
Publisher
Archives Hub, U.K.
Publication Location
Manchester, England, U.K.
Language
English
Critical Arguements
CA "These pages have been designed to hold links and information which we hope will be useful to archivists and librarians working in the UK Higher and Further Education sectors."
Conclusions
RQ
SOW
DC The Archives Hub provides a single point of access to 17,598 descriptions of archives held in UK universities and colleges. At present these are primarily at collection-level, although complete catalogue descriptions are provided where they are available. The Archives Hub forms one part of the UK's National Archives Network, alongside related networking projects. A Steering Committee which includes representatives of contributing institutions, the National Archives and the other archive networks guides the progress of the project. There is also a Contributors' and Users' Forum which provides feedback to aid the development of the service. The service is hosted at MIMAS on behalf of the Consortium of University Research Libraries (CURL) and is funded by the Joint Information Systems Committee (JISC). Systems development work is undertaken at the University of Liverpool.
Type
Web Page
Title
Schema Registry: activityreports: Recordkeeping Metadata Standard for Commonwealth Agencies
CA "The Australian SPIRT Recordkeeping Metadata Project was initially a project funded under a programme known as the Strategic Partnership with Industry -- Research and Training (SPIRT) Support Grant -- partly funded by the Australian Research Council. The project was concerned with developing a framework for standardising and defining recordkeeping metadata and produced a metadata element set eventually known as the Australian Recordkeeping Metadata Schema (RKMS). The conceptual frame of reference in the project was based in Australian archival practice, including the Records Continuum Model and the Australian Series System. The RKMS also inherits part of the Australian Government Locator Service (AGLS) metadata set."
The creation and use of metadata is likely to become an important part of all digital preservation strategies whether they are based on hardware and software conservation, emulation or migration. The UK Cedars project aims to promote awareness of the importance of digital preservation, to produce strategic frameworks for digital collection management policies and to promote methods appropriate for long-term preservation - including the creation of appropriate metadata. Preservation metadata is a specialised form of administrative metadata that can be used as a means of storing the technical information that supports the preservation of digital objects. In addition, it can be used to record migration and emulation strategies, to help ensure authenticity, to note rights management and collection management data and also will need to interact with resource discovery metadata. The Cedars project is attempting to investigate some of these issues and will provide some demonstrator systems to test them.
Notes
This article was presented at the Joint RLG and NPO Preservation Conference: Guidelines for Digital Imaging, held September 28-30, 1998.
Critical Arguements
CA "Cedars is a project that aims to address strategic, methodological and practical issues relating to digital preservation (Day 1998a). A key outcome of the project will be to improve awareness of digital preservation issues, especially within the UK higher education sector. Attempts will be made to identify and disseminate: Strategies for collection management ; Strategies for long-term preservation. These strategies will need to be appropriate to a variety of resources in library collections. The project will also include the development of demonstrators to test the technical and organisational feasibility of the chosen preservation strategies. One strand of this work relates to the identification of preservation metadata and a metadata implementation that can be tested in the demonstrators." ... "The Cedars Access Issues Working Group has produced a preliminary study of preservation metadata and the issues that surround it (Day 1998b). This study describes some digital preservation initiatives and models with relation to the Cedars project and will be used as a basis for the development of a preservation metadata implementation in the project. The remainder of this paper will describe some of the metadata approaches found in these initiatives."
Conclusions
RQ "The Cedars project is interested in helping to develop suitable collection management policies for research libraries." ... "The definition and implementation of preservation metadata systems is going to be an important part of the work of custodial organisations in the digital environment."
SOW
DC "The Cedars (CURL exemplars in digital archives) project is funded by the Joint Information Systems Committee (JISC) of the UK higher education funding councils under Phase III of its Electronic Libraries (eLib) Programme. The project is administered through the Consortium of University Research Libraries (CURL) with lead sites based at the Universities of Cambridge, Leeds and Oxford."
Type
Web Page
Title
Metadata for preservation : CEDARS project document AIW01
This report is a review of metadata formats and initiatives in the specific area of digital preservation. It supplements the DESIRE Review of metadata (Dempsey et al. 1997). It is based on a literature review and information picked-up at a number of workshops and meetings and is an attempt to briefly describe the state of the art in the area of metadata for digital preservation.
Critical Arguements
CA "The projects, initiatives and formats reviewed in this report show that much work remains to be done. . . . The adoption of persistent and unique identifiers is vital, both in the CEDARS project and outside. Many of these initiatives mention "wrappers", "containers" and "frameworks". Some thought should be given to how metadata should be integrated with data content in CEDARS. Authenticity (or intellectual preservation) is going to be important. It will be interesting to investigate whether some archivists' concerns with custody or "distributed custody" will have relevance to CEDARS."
Conclusions
RQ Which standards and initiatives described in this document have proved viable preservation metadata models?
SOW
DC OAIS emerged out of an initiative spearheaded by NASA's Consultative Committee for Space Data Systems. It has been shaped and promoted by the RLG and OCLC. Several international projects have played key roles in shaping the OAIS model and adapting it for use in libraries, archives and research repositories. OAIS-modeled repositories include the CEDARS Project, Harvard's Digital Repository, Koninklijke Bibliotheek (KB), the Library of Congress' Archival Information Package for audiovisual materials, MIT's D-Space, OCLC's Digital Archive and TERM: the Texas Email Repository Model.
Type
Web Page
Title
Approaches towards the Long Term Preservation of Archival Digital Records
The Digital Preservation Testbed is carrying out experiments according to pre-defined research questions to establish the best preservation approach or combination of approaches. The Testbed will be focusing its attention on three different digital preservation approaches - Migration; Emulation; and XML - evaluating the effectiveness of these approaches, their limitations, costs, risks, uses, and resource requirements.
Language
English; Dutch
Critical Arguements
CA "The main problem surrounding the preservation of authentic electronic records is that of technology obsolescence. As changes in technology continue to increase exponentially, the problem arises of what to do with records that were created using old and now obsolete hardware and software. Unless action is taken now, there is no guarantee that the current computing environment (and thus also records) will be accessible and readable by future computing environments."
Conclusions
RQ "The Testbed will be conducting research to discover if there is an inviolable way to associate metadata with records and to assess the limitations such an approach may incur. We are also working on the provision of a proposed set of preservation metadata that will contain information about the preservation approach taken and any specific authenticity requirements."
SOW
DC The Digital Preservation Testbed is part of the non-profit organisation ICTU. ICTU is the Dutch organisation for ICT and government. ICTU's goal is to contribute to the structural development of e-government. This will result in improving the work processes of government organisations, their service to the community and interaction with the citizens. Government institutions, such as Ministries, design the policies in the area of e-government, and ICTU translates these policies into projects. In many cases, more than one institution is involved in a single project. They are the principals in the projects and retain control concerning the focus of the project. In case of the Digital Preservation Testbed the principals are the Ministry of the Interior and the Dutch National Archives.
Type
Web Page
Title
Appendix N to Proceedings of The Uniform Law Conference of Canada, Proposals for a Uniform Electronic Evidence Act
CA "First, there is a great deal of uncertainty about how the [Canada Evidence Act], particularly s. 30(6), will be applied, and this makes it difficult for the parties to prepare for litigation and for businesses to know how they should keep their records. Second, there are risks to the integrity of records kept on a computer that do not exist with respect to other forms of information processing and storage, and if alterations are made, either negligently or deliberately, they can be extremely difficult to detect. Third, s. 30(1) provides little assurance that the record produced to the court is the same as the one that was originally made in the usual and ordinary course of business, for while self-interest may be an adequate guarantee that most businesses will maintain accurate and truthful records, it is not true for many others. The second and third problems combined place the party opposing the introduction of computer-produced business records in a difficult situation."
SOW
DC The Uniform Law Conference of Canada undertook to adopt uniform legislation to ensure that computer records could be used appropriately in court.
Type
Web Page
Title
Softening the borderlines of archives through XML - a case study
Archives have always had troubles getting metadata in formats they can process. With XML, these problems are lessening. Many applications today provide the option of exporting data into an application-defined XML format that can easily be post-processed using XSLT, schema mappers, etc, to fit the archives┬┤ needs. This paper highlights two practical examples for the use of XML in the Swiss Federal Archives and discusses advantages and disadvantages of XML in these examples. The first use of XML is the import of existing metadata describing debates at the Swiss parliament whereas the second concerns preservation of metadata in the archiving of relational databases. We have found that the use of XML for metadata encoding is beneficial for the archives, especially for its ease of editing, built-in validation and ease of transformation.
Notes
The Swiss Federal Archives defines the norms and basis of records management and advises departments of the Federal Administration on their implementation. http://www.bar.admin.ch/bar/engine/ShowPage?pageName=ueberlieferung_aktenfuehrung.jsp
Critical Arguements
CA "This paper briefly discusses possible uses of XML in an archival context and the policies of the Swiss Federal Archives concerning this use (Section 2), provides a rough overview of the applications we have that use XML (Section 3) and the experiences we made (Section 4)."
Conclusions
RQ "The systems described above are now just being deployed into real world use, so the experiences presented here are drawn from the development process and preliminary testing. No hard facts in testing the sustainability of XML could be gathered, as the test is time itself. This test will be passed when we can still access the data stored today, including all metadata, in ten or twenty years." ... "The main problem area with our applications was the encoding of the XML documents and the non-standard XML document generation of some applications. When dealing with the different encodings (UTF-8, UTF-16, ISO-8859-1, etc) some applications purported a different encoding in the header of the XML document than the true encoding of the document. These errors were quickly identified, as no application was able to read the documents."
SOW
DC The author is currently a private digital archives consultant, but at the time of this article, was a data architect for the Swiss Federal Archives. The content of this article owes much to the work being done by a team of architects and engineers at the Archives, who are working on an e-government project called ARELDA (Archiving of Electronic Data and Records).
Type
Web Page
Title
The Making and the Keeping of Records: (2) The Tyranny of Listing
CA Listing is tantamount to traditional recordkeeping methodology. This paradigm needs to be reconsidered to allow for better-designed archival systems.
Conclusions
RQ Can we ultimately abandon the traditional concern of ensuring records' persistence and still keep records?
Type
Web Page
Title
Report of the Ad Hoc Committee for Development of a Standardized Tool for Encoding Finding Aids
This report focuses on the development of tools for the description and intellectual control of archives and the discovery of relevant resources by users. Other archival functions, such as appraisal, acquisition, preservation, and physical control, are beyond the scope for this project. The system developed as a result of this report should be useable on stand-alone computers in small institutions, by multiple users in larger organisations, and by local, regional, national, and international networks. The development of such a system should take into account the strategies, experiences, and results of other initiatives such as the European Union Archival Network (EUAN), the Linking and Exploring Authority Files (LEAF) initiative, the European Visual Archives (EVA) project, and the Canadian Archival Information Network (CAIN). This report is divided into five sections. A description of the conceptual structure of an archival information system, described as six layers of services and protocols, follows this introduction. Section three details the functional requirements for the software tool and is followed by a discussion of the relationship of these requirements to existing archival software application. The report concludes with a series of recommendations that provide a strategy for the successful development, deployment, and maintenance of an Open Source Archival Resource Information System (OSARIS). There are two appendices: a data model and a comparison of the functional requirements statements to several existing archival systems.
Notes
3. Functional Requirements Requirements for Information Interchange 3.2: The system must support the current archival standards for machine-readable data communication, Encoded Archival Description (EAD) and Encoded Archival Context (EAC). A subset of elements found in EAD may be used to exchange descriptions based on ISAD(G) while elements in EAC may be used to exchange ISAAR(CPF)-based authority data.
Publisher
International Council on Archives Committee on Descriptive Standards
Critical Arguements
CA The Ad Hoc Committee agrees that it would be highly desirable to develop a modular, open source software tool that could be used by archives worldwide to manage the intellectual control of their holdings through the recording of standardized descriptive data. Individual archives could combine their data with that of other institutions in regional, national or international networks. Researchers could access this data either via a stand-alone computerized system or over the Internet. The model for this software would be the successful UNESCO-sponsored free library program, ISIS, which has been in widespread use around the developing world for many years. The software, with appropriate supporting documentation, would be freely available via an ICA or UNESCO web site or on CD-ROM. Unlike ISIS, however, the source code and not just the software should be freely available.
Conclusions
RQ "1. That the ICA endorses the functional requirements presented in this document as the basis for moving the initiative forward. 2. That the functional desiderata and technical specifications for the software applications, such as user requirements, business rules, and detailed data models, should be developed further by a team of experts from both ICA/CDS and ICA/ITC as the next stage of this project. 3. That following the finalization of the technical specifications for OSARIS, the requirements should be compared to existing systems and a decision made to adopt or adapt existing software or to build new applications. At that point in time, it will then be possible to estimate project costs. 4. That a solution that incorporates the functional requirements result in the development of several modular software applications. 5. That the implementation of the system should follow a modular strategy. 6. That the development of software applications must include a thorough investigation and assessment of existing solutions beginning with those identified in section four and Appendix B of this document. 7. That the ICA develop a strategy for communicating the progress of this project to members of the international archival community on a regular basis. This would include the distribution of progress reports in multiple languages. The communication strategy must include a two-way exchange of ideas. The project will benefit strongly from the ongoing comments, suggestions, and input of the members of the international archival community. 8. That a test-bed be developed to allow the testing of software solutions in a realistic archival environment. 9. That the system specifications, its documentation, and the source codes for the applications be freely available. 10. That training courses for new users, ongoing education, and webbased support groups be established. 11. That promotion of the software be carried out through the existing regional infrastructure of ICA and through UNESCO. 12. That an infrastructure for ongoing maintenance, distribution, and technical support be developed. This should include a web site to download software and supporting documentation. The ICA should also establish and maintain a mechanism for end-users to recommend changes and enhancements to the software. 13. That the ICA establishes and maintains an official mechanism for regular review of the software by an advisory committee that includes technical and archival experts. "
SOW
DC "The development of such a system should take into account the strategies, experiences, and results of other initiatives such as the European Union Archival Network (EUAN), the Linking and Exploring Authority Files (LEAF) initiative, the European Visual Archives (EVA) project, and the Canadian Archival Information Network (CAIN)."
This document is a draft version 1.0 of requirements for a metadata framework to be used by the International Press Telecommunications Council for all new and revised IPTC standards. It was worked on and agreed to by members of the IPTC Standards Committee, who represented a variety of newspaper, wire agencies, and other interested members of the IPTC.
Notes
Misha Wolf is also listed as author.
Publisher
International Press Telecommunications Council (IPTC)
Critical Arguements
CA "This Requirements document forms part of the programme of work called ITPC Roadmap 2005. The Specification resulting from these Requirements will define the use of metadata by all new IPTC standards and by new major versions of existing IPTC standards." (p. 1) ... "The purpose of the News Metadata Framework (NMDF) WG is to specify how metadata will be expressed, referenced, and managed in all new major versions of IPTC standards. The NMF WG will: Gather, discuss, agree and document functional requirements for the ways in which metadata will be expressed, referenced and managed in all new major versions of IPTC standards; Discuss, agree and document a model, satisfying these requirements; Discuss, agree and document possible approaches to expressing this model in XML, and select those most suited to the tasks. In doing so, the NMDF WG will, where possible, make use of the work of other standards bodies. (p. 2)
Conclusions
RQ "Open issues include: The versioning of schemes, including major and minor versions, and backward compatibility; the versioning of TopicItems; The design of URIs for TopicItem schemes and TopicItem collections, including the issues of: versions (relating to TopicItems, schemes, and collections); representations (relating to TopicItems and collections); The relationship between a [scheme, code] pair, the corresponding URI and the scheme URI." (p. 17)
SOW
DC The development of this framework came out of the 2003 News Standards Summit, which was attended by representatives from over 80 international press and information agencies ... "The News Standards Summit brings together major players--experts on news metadata standards as well as commercial news providers, users, and aggregators. Together, they will analyze the current state and future expectations for news and publishing XML and metadata efforts from both the content and processing model perspectives. The goal is to increase understanding and to drive practical, productive convergence." ... This is a draft version of the standard.
Type
Web Page
Title
NHPRC: Minnesota State Archives Strategic Plan: Electronic Records Consultant Project
National Historical Publications and Records Commission Grant No. 95-030
Critical Arguements
CA "The Electronic Records Consultant Project grant was carried out in conjunction with the strategic planning effort for the Minnesota Historical Society's State Archives program. The objective was to develop a plan for a program that will be responsive to the changing nature of government records." ... "The strategic plan that was developed calls for specific actions to meet five goals: 1) strengthening partnerships, 2) facilitating the identification of historically valuable records, 3) integrating electronic records into the existing program, 4) providing quality public service, and 5) structuring the State Archives Department to meet the demands of this plan."
"The ERMS Metadata Standard forms Part 2 of the National Archives' 'Requirements for Electronic Records Management Systems' (commonly known as the '2002 Requirements'). It is specified in a technology independent manner, and is aligned with the e-Government Metadata Standard (e-GMS) version 2, April 2003. A version of e-GMS v2 including XML examples was published in the autumn of 2003. This Guide should be read in conjunction with the ERMS Metadata Standard. Readers may find the GovTalk Schema Guidelines (available via http://www.govtalk.gov.uk ) helpful regarding design rules used in building the schemas."
Conclusions
RQ Electronically enabled processes need to generate appropriate records, according to established records management principles. These records need to reach the ERMS that captures them with enough information to enable the ERMS to classify them appropriately, allocate an appropriate retention policy, etc.
SOW
DC This document is a draft.
Type
Web Page
Title
Use of Encoded Archival Description (EAD) for Manuscript Collection Finding Aids
Presented in 1999 to the Library's Collection Development & Management Committee, this report outlines support for implementing EAD in delivery of finding aids for library collections over the Web. It describes the limitations of HTML, provides an introduction to SGML, XML, and EAD, outlines the advantages of conversion from HTML to EAD, the conversion process, the proposed outcome, and sources for further information.
Publisher
National Library of Australia
Critical Arguements
CA As use of the World Wide Web has increased, so has the need of users to be able to discover web-based information resources easily and efficiently, and to be able to repeat that discovery in a consistent manner. Using SGML to mark up web-based documents facilitates such resource discovery.
Conclusions
RQ To what extent have the mainstream web browser companies fulfilled their committment to support native viewing of SGML/XML documents?
This guide is optimized for creation of EAD-encoded finding aids for the collections of New York University and New York Historical Society. The links on the page list tools and files that may be downloaded and referenced for production of NYU-conformant finding aids.
Publisher
New York University
Critical Arguements
CA "This guide is optimized for creation of EAD-encoded finding aids for the collections of New York University and New York Historical Society. Instructions assume the use of NoteTab as the XML editor, utilizing template files that serve as base files for the different collections." 
Conclusions
RQ
SOW
DC This guide serves both New York University and the New York Historical Society.
Type
Web Page
Title
Preservation Metadata and the OAIS Information Model: A Metadata Framework to Support the Preservation of Digital Objects
CA "In March 2000, OCLC and RLG sponsored the creation of a working group to explore consensus-building in the area of preservation metadata. ... The charge of the group was to pool their expertise and experience to develop a preservation metadata framework applicable to a broad range of digital preservation activities." (p.1) "The OAIS information model offers a broad categorization of the types of information falling under the scope of preservation metadata; it falls short, however, of providing a decomposition of these information types into a list of metadata elements suitable for practical implementation. It is this need that the working group addressed in the course of its activities, the results of which are reported in this paper." (p. 47)
Conclusions
RQ "The metadata framework described in this paper can serve as a foundation for future work in the area of preservation metadata. Issues of particular importance include strategies and best practices for implementing preservation metadata in an archival system; assessing the degree of descriptive richness required by various types of digital preservation activities; developing algorithms for producing preservation metadata automatically; determining the scope for sharing preservation metadata in a cooperative environment; and moving beyond best practice towards an effort at formal standards building in this area." (47)
SOW
DC "[The OCLC and RLG working group] began its work by publishing a white paper entitled Preservation Metadata for Digital Objects: A Review of the State of the Art, which defined and discussed the concept of preservation metadata, reviewed current thinking and practice in the use of preservation metadata, and identified starting points for consensus-building activity in this area. The group then turned its attention to the main focus of its activity -- the collaborative development of a preservation metadata framework. This paper reports the results of the working groupÔÇÖs efforts in that regard." (p. 1-2)
During the past decade, the recordkeeping practices in public and private organizations have been revolutionized. New information technologies from mainframes, to PC's, to local area networks and the Internet have transformed the way state agencies create, use, disseminate, and store information. These new technologies offer a vastly enhanced means of collecting information for and about citizens, communicating within state government and between state agencies and the public, and documenting the business of government. Like other modern organizations, Ohio state agencies face challenges in managing and preserving their records because records are increasingly generated and stored in computer-based information systems. The Ohio Historical Society serves as the official State Archives with responsibility to assist state and local agencies in the preservation of records with enduring value. The Office of the State Records Administrator within the Department of Administrative Services (DAS) provides advice to state agencies on the proper management and disposition of government records. Out of concern over its ability to preserve electronic records with enduring value and assist agencies with electronic records issues, the State Archives has adapted these guidelines from guidelines created by the Kansas State Historical Society. The Kansas State Historical Society, through the Kansas State Historical Records Advisory Board, requested a program development grant from the National Historical Publications and Records Commission to develop policies and guidelines for electronic records management in the state of Kansas. With grant funds, the KSHS hired a consultant, Dr. Margaret Hedstrom, an Associate Professor in the School of Information, University of Michigan and formerly Chief of State Records Advisory Services at the New York State Archives and Records Administration, to draft guidelines that could be tested, revised, and then implemented in Kansas state government.
Notes
These guidelines are part of the ongoing effort to address the electronic records management needs of Ohio state government. As a result, this document continues to undergo changes. The first draft, written by Dr. Margaret Hedstrom, was completed in November of 1997 for the Kansas State Historical Society. That version was reorganized and updated and posted to the KSHS Web site on August 18, 1999. The Kansas Guidelines were modified for use in Ohio during September 2000
Critical Arguements
CA "This publication is about maintaining accountability and preserving important historical records in the electronic age. It is designed to provide guidance to users and managers of computer systems in Ohio government about: the problems associated with managing electronic records, special recordkeeping and accountability concerns that arise in the context of electronic government; archival strategies for the identification, management and preservation of electronic records with enduring value; identification and appropriate disposition of electronic records with short-term value, and
Type
Web Page
Title
Online Archive of California Best Practice Guidelines for Encoded Archival Description, Version 1.1
These guidelines were prepared by the OAC Working Group's Metadata Standards Subcommittee during the spring and summer of 2003. This version of the OAC BPG EAD draws substantially on the
Language
Anonymous
Type
Web Page
Title
Descriptive Metadata Guidelines for RLG Cultural Materials
To ensure that the digital collections submitted to RLG Cultural Materials can be discovered and understood, RLG has compiled these Descriptive Metadata Guidelines for contributors. While these guidelines reflect the needs of one particular service, they also represent a case study in information sharing across community and national boundaries. RLG Cultural Materials engages a wide range of contributors with different local practices and institutional priorities. Since it is impossible to find -- and impractical to impose -- one universally applicable standard as a submission format, RLG encourages contributors to follow the suite of standards applicable to their particular community (p.1).
Critical Arguements
CA "These guidelines . . . do not set a new standard for metadata submission, but rather support a baseline that can be met by any number of strategies, enabling participating institutions to leverage their local descriptions. These guidelines also highlight the types of metadata that enhance functionality for RLG Cultural Materials. After a contributor submits a collection, RLG maps that description into the RLG Cultural Materials database using the RLG Cultural Materials data model. This ensures that metadata from the various participant communities is integrated for efficient searching and retrieval" (p.1).
Conclusions
RQ Not applicable.
SOW
DC RLG comprises more than 150 research and cultural memory institutions, and RLG Cultural Materials elicits contributions from countless museums, archives, and libraries from around the world that, although they might retain local descriptive standards and metadata schemas, must conform to the baseline standards prescribed in this document in order to integrate into RLG Cultural Materials. Appendix A represents and evaluates the most common metadata standards with which RLG Cultural Materians is able to work.
Expanded version of the article "Ensuring the Longevity of Digital Documents" that appeared in the January 1995 edition of Scientific American (Vol. 272, Number 1, pp. 42-7).
Publisher
Council on Library and Information Resources
Critical Arguements
CA "It is widely accepted that information technology is revolutionizing our concepts of documents and records in an upheaval at least as great as the introduction of printing, if not of writing itself. The current generation of digital records therefore has unique historical significance; yet our digital documents are far more fragile than paper. In fact, the record of the entire present period of history is in jeopardy. The content and historical value of many governmental, organizational, legal, financial, and technical records, scientific databases, and personal documents may be irretrievably lost to future generations if we do not take steps to preserve them."
Conclusions
RQ "We must develop evolving standards for encoding explanatory annotations to bootstrap the interpretation of digital documents that are saved in nonstandard forms. We must develop techniques for saving the bit streams of software-dependent documents and their associated systems and application software. We must ensure that the hardware environments necessary to run this software are described in sufficient detail to allow their future emulation. We must save these specifications as digital documents, encoded using the bootstrap standards developed for saving annotations so that they can be read without special software (lest we be recursively forced to emulate one system in order to learn how to emulate another). We must associate contextual information with our digital documents to provide provenance as well as explanatory annotations in a form that can be translated into successive standards so as to remain easily readable. Finally, we must ensure the systematic and continual migration of digital documents onto new media, preserving document and program bit streams verbatim, while translating their contextual information as necessary."
Type
Web Page
Title
Archiving of Electronic Digital Data and Records in the Swiss Federal Archives (ARELDA): e-government project ARELDA - Management Summary
The goal of the ARELDA project is to find long-term solutions for the archiving of digital records in the Swiss Federal Archives. This includes the accession, the long-term storage, preservation of data, description, and access for the users of the Swiss Federal Archives. It is also coordinated with the basic efforts of the Federal Archives to realize a uniform records management solution in the federal administration and therefore to support the pre-archival creation of documents of archival value for the benefits of the administration as well as of the Federal Archives. The project is indispensable for the long-term execution of the Federal Archives Act; Older IT systems are being replaced by newer ones. A complete migration of the data is sometimes not possible or too expensive; A constant increase of small database applications, built and maintained by people with no IT background; More and more administrative bodies are introducing records and document management systems.
Publisher
Swiss Federal Archives
Publication Location
Bern
Critical Arguements
CA "Archiving in general is a necessary prerequisite for the reconstruction of governmental activities as well as for the principle of legal certainty. It enables citizens to understand governmental activities and ensures a democratic control of the federal administration. And finally are archives a prerequisite for the scientific research, especially in the social and historical fields and ensure the preservation of our cultural heritage. It plays a vital role for an ongoing and efficient records management. A necessary prerequisite for the Federal Archives in the era of the information society will be the system ARELDA (Archiving of Electronic Data and Records)."
Conclusions
RQ "Because of the lack of standard solutions and limited or lacking personal resources for an internal development effort, the realisation of ARELDA will have to be outsourced and the cooperation with the IT division and the Federal Office for Information Technology, Systems and Telecommunication must be intensified. The guidelines for the projects are as follows:
SOW
DC ARELDA is one of the five key projects in the Swiss government's e-government strategy.
Museums and the Online Archive of California (MOAC) builds on existing standards and their implementation guidelines provided by the Online Archive of California (OAC) and its parent organization, the California Digital Library (CDL). Setting project standards for MOAC consisted of interpreting existing OAC/CDL documents and adapting them to the projects specific needs, while at the same time maintaining compliance with OAC/CDL guidelines. The present overview over the MOAC technical standards references both the OAC/CDL umbrella document and the MOAC implementation / adaptation document at the beginning of each section, as well as related resources which provide more detail on project specifications.
Critical Arguements
CA The project implements specifications for digital image production, as well as three interlocking file exchange formats for delivering collections, digital images and their respective metadata. Encoded Archival Description (EAD) XML describes the hierarchy of a collection down to the item-level and traditionally serves for discovering both the collection and the individual items within it. For viewing multiple images associated with a single object record, MOAC utilizes Making of America 2 (MOA2) XML. MOA2 makes the images representing an item available to the viewer through a navigable table of contents; the display mimics the behavior of the analog item by e.g. allowing end-users to browse through the pages of an artist's book. Through the further extension of MOA2 with Text Encoding Initiative (TEI) Lite XML, not only does every single page of the book display in its correct order, but a transcription of its textual content also accompanies the digital images.
Conclusions
RQ "These two instances of fairly significant changes in the project's specifications may serve as a gentle reminder that despite its solid foundation in standards, the MOAC information architecture will continue to face the challenge of an ever-changing technical environment."
SOW
DC The author is Digital Media Developer at the UC Berkeley Art Museum & Pacific Film Archives, a member of the MOAC consortium.
CA One problem in the field of radio archives is the tendency to view anything that is not audio or video (specifically this leaves text) as metadata. However, all text is not metadata. While all text can be seen as potentially useful due to the information it represents, the creators of P/FRA recommend standardizing only the essential information needed to describe and retrieve radio archive information.
Conclusions
RQ Rules need to be drafted specifying the content of metadata fields. While the authors extol the value of ÔÇ£good metadataÔÇØ for resource discovery, proscribing the content of metadata containers is a problem here as in every other filed.