1. Introduction
The Giorgio Cini Foundation photo library in Venice preserves one of the largest historical art history photo archives in Italy and one of the most important in Europe.
In 2015, the Swiss Federal Institute of Technology in Lausanne (henceforth ETH Lausanne), the curators of the Giorgio Cini photo library and Adam Lowe of the Factum Arte Foundation began to discuss how to proceed with a massive and rapid digitization of the most substantial corpus of the photo library’s photos, i.e., the index cards, about 1 million items.
It was a synergy of undoubted value and heterogeneity between the big players, since the photographic archive could boast a cultural heritage of exceptional value and eminently representative of centuries-old visual artistic memory, the ETH Lausanne embodied cutting-edge expertise in the field of computer science, in particular computer vision, capable of dealing with the historical and technical challenges posed by the corpus in question with the tools of state-of-the-art algorithms, and Factum Arte was the leading foundation in the field of facsimile reproductions of artistic objects, be they archeological, pictorial or sculptural.
This article explains the reasons for such a collaboration, describing the ‘Replica’ project that these institutions have co-designed and led, creating a virtuous and mutually enriching partnership, which involves digitally enhancing the Foundation’s historical heritage, from digitization to valorization and online publication for communities of young scholars from all over the world.
This contribution also aims to show how this mode of co-operation can become a model for other cases where digitization of heritage, conducted jointly between a heritage institution, a research institution and a stakeholder involved in the vanguard of digitization techniques, can follow a development pipeline from document digitization to making the document available to the public.
The Replica project also showed how all the issues raised by the digital transition of heritage can be addressed by finding effective answers in the plurality of profiles in the field and in the involvement of the student-researcher community to build an image search engine suited, in the first place, to the needs of the research world.
As a patrimonial institution, the Giorgio Cini Foundation brought the experience of archival and curatorial practice, linked to a deep tradition of expertise, and structured according to the logic of the ministerial approach and policies linked to the need for conservation and protection over the long term. In addition to this structural mission, there was also the need for transmission to the community of researchers and more generally to citizens [
1]; it is no coincidence that the institution’s motto was also inspired by a famous phrase by Mahler “Tradition is the handing down of the flame and not the worshipping of ashes” [Tradition ist die Weitergabe des Feuers und nicht die Anbetung der Asche].
The Giorgio Cini Foundation Photo Archive has an immense intrinsic historical value hosting different funds from the great photographers and photographic campaigns of the 20th century. Over time, the Giorgio Cini Foundation has brought together different photographic funds whose main vocation is to represent images of works of art that are representative of Venice and its historical and artistic heritage in the broadest sense of the term. The vocation of the Foundation’s photo library is to provide visual encyclopedic access to Venetian cultural production, thus characterizing itself as having a very strong form of locational interest. At the same time, this characterization has always attracted scholars from all over the world who specifically seek out elements for research focused on Venetianness, at a granularity of content that more generic photo archives could hardly provide.
The ambition of the Replica project, from the point of view of the heritage institution, was to make available internationally, on a global scale, access to these sources, which are very specific documents of the Venetian heritage, in order to open a very local archive to global research, and to democratize the visual elements constituting Venetian history through the open access of the photographic corpus [
2]. For this reason, the Replica project made use of the involvement of a community of young art history scholars who, through special workshops and training courses, helped to co-develop the main features of the image search engine.
As an academic institution, ETH Lausanne, and in particular the Digital Humanities Laboratory, incorporates research and teaching, the digitization project therefore reached out to the student community, training them in new critical practices of combining technology and historical questions. In fact, digitization required from the very beginning of the planned pipeline to adapt the process to the historical research questions to be addressed. The introduction of the community of researchers and students made it possible to pose digitization problems as transdisciplinary issues to which students had to find solutions. Project-related activities also made it possible to offer work placements in the heritage institution, within the framework of the project, which were useful for training new professionals.
From a research perspective, the ETH Lausanne Digital Humanities Laboratory was interested in experimenting with the possibility of manipulating a large corpus of images by analyzing the concept of visual similarity and pattern migration. In fact, the development work of the image navigation interface thematized this aspect, focusing on the analysis of visual patterns.
In addition, the collaboration with the Factum Arte Foundation has made it possible to define a new digitization system suited to the precipitousness of the corpus or works chosen. In particular, the Giorgio Cini Foundation wanted a modular scanning system in which the scanner could be a set of components that could be easily replaced and implemented to improve performance.
Factum Arte had in fact already worked with the Cini Foundation on the pioneering project of digitizing and reproducing on a 1:1 scale Paolo Veronese’s Marriage at Cana, the original in the Louvre, in order to relocate the copy in the Refectory of the convent of San Giorgio Maggiore where it originally stood before its removal by Napoleon [
3].
From Photo Archive to Digital, Codesigning the Project Pipeline
The initial aim of the project is to convert Cini’s vast photographic archive into a digital archive. However, this is only the general concept summarizing a will, but then several aspects come into play related to the concrete elaboration of the pipeline, the organization of the work and the realization of the most interesting and useful digital enhancement system for the institution but also for researchers.
Hierarchies of roles were then decided in order to optimize the progress of the project. For each institution there was a project leader. In the case of the Cini Foundation, it was the IT manager who had to ensure a good scanning and preservation method within the parameters defined by the institution itself. In the case of the ETH Lausanne, it was a figure capable of understanding in detail the technical characteristics required and the development of digitization to answer some key research questions on the side of Digital Humanities and Art History. In the case of Factum Arte, responsible for the implementation of the fast and accurate scanning system, this was the technical manager able to understand the technical requirements of the institution but also capable of knowing the most advanced state of the art with respect to this technological challenge.
Before materially starting the operations, all partners involved had several workshops characterized by an open approach and brief to try to understand what the heritage institution’s practices had been up to then and how they would like to improve their daily work and the provision of materials to the public [
4]. At the same time, the ETH Lausanne recorded potential points to be developed technologically but also shared what was the state of the art with regard to computer vision applied to the management of photographic heritage collections and especially to art history. This exchange made it possible to put all actors on the same starting line of ‘making it possible’ [
5]. Once the project’s wish list of expectations had been established, the critical list was defined. These had different natures: technical logistical in the material process of digitization, optimization of the human resources involved in the various development phases of the pipeline, problems of conceptualization of the digital approach, implementation of the digital access interface, optimization of the digital interface for users.
Aware that the first attempt to define the list of critical issues could not be exhaustive and that others would arise during the operations, it was decided to increase the moments of confrontation and discussion by multiplying the debriefings for each phase of the project (
Figure 1).
The only element that had to meet the requirement of stability, and on which there could be no turning back after operations had begun and were already well advanced, was the technical digitization protocol. To finalize it, in fact, preliminary tests were carried out before the scanner was even operational. Since the autumn of 2015, several tests have been carried out between to understand the technical characteristics of image acquisition, the most suitable file format to store the metadata related to the archival location, the best method to index the physical items of the photographic archive, and the cardboard circumventing manual indexing.
It is very important in co-design projects to distinguish between processes that are circular and those that are one-off. The former can be characterized by multiple briefs with all parties involved because they can by nature be improved, indeed they derive their optimization from circularity. For one off, on the other hand, such as digitization, notwithstanding the fact that they could be started again, careful preliminary planning and adherence to the quality that one has set oneself very early in the process is necessary, otherwise starting everything all over again would be a great waste of time. Furthermore, in digitization projects there are phases in which quality feedback is automatic and does not require conceptual feedback, such as document processing and data linking. In-depth feedback is absolutely necessary for the final output, i.e., for the ‘product’ of valorizing and empowering the ‘heritage capital’ that the institution wants to value and the research institute to study.
This is why a lot of feedback is needed once the pipeline has reached the stage of generating the digital archive (
Figure 1). First of all, there are inalienable control parameters such as the ‘coverage’ and ‘completeness’ of digitization. In other words, it has to be checked whether all analog elements have been converted, e.g., no images to be digitized have been forgotten and the metadata of the archival location have been correctly reported. Then there is the information inferred through document processing, i.e., a step following digitization, which must strive for completeness as far as possible. These checks, which are semi-automated, serve as feedback for the heritage institution to eventually understand which documents need to be re-digitized and for which a set of information can be completed manually.
Once these aspects have been assessed and told in brief, the digital collection can be considered ‘finished’. It can then already be made available to the usual public with a simple ‘demonstrator viewer’ or disseminated through an “.api” of the institution. In our case, however, it was planned to create a specific search engine for the photographic archive that could not only be a traditional ‘collection manager’ but also allow searching through clusters of images that are similar to each other from a formal point of view without only classifying between metadata. For this several feedback from the heritage institution were necessary and as a complement also a series of workshops with art history researchers to present the idea and get opinions on which tools were useful for the scholarly community. This made it possible to build a circular approach in which at each debrief of these communities, not always in agreement with each other, the most relevant comments were implemented in the realization of the search engine [
6]. The co-design certainly allowed a more comprehensive approach to the realization of the interface. All transversal observations were implemented to make the interface more useful for conservators but also for scholars. Usually, in fact, heritage actors, in particular photographic archives, revert to IT professionals to whom they commission, often with tight budget constraints, collection management software that is rather standardized and does not use state-of-the-art technologies, ignoring the dynamics of use of researchers and the public who should instead be involved as essential and final users of the digital archive.
2. How to Massively and Quickly Digitizing a Photo Archive
2.1. Co-Designing the Scanner
The co-definition of digitization goals allowed us to prototype an effective and sustainable pipeline. The photo library’s cardboard had to be digitized at a much faster rate than was commonly practiced. But digitization was only the first step in a pipeline that also had to include, as a continuous workflow, the transmission of the images to the ETH Lausanne engineers in charge of processing them immediately in order to proceed with the extraction of the content and its study. In each part of the workflow, the skills of different professionals came into play. It needed a system capable of semi-automatically digitizing as many items as possible per minute, knowing that each cardboard contained valuable information on the front and back and that both had to be stored. The resolution had to be defined according to current international criteria and the practices already tried and tested by the conservators of the photo library. If we had been able to digitize 1000 cardboard for five hours a day, it would already have been an extraordinary achievement, but it was still not enough, we had to reach a possible average of 1500 cardboard per day to be able to finish in the two years we had given ourselves and which could be financed by the project. In addition, the need to digitize in recto and verso posed a further challenge and demanded an ergonomic system for operators. Another problematic element in contrast to speed was the manual input of metadata, something usually practiced by conservators, and evidently difficult to apply in a system that was intended to be fully automatic or semi-automatic in which the machine would have to process all the cardboard consecutively without interruption.
These elements were resolved when Adam Lowe thought to take inspiration from a famous image taken from the book ‘Le diverse et artificiose macchine’ compiled by Agostino Ramelli, an engineer at the service of the French monarchy, and published in 1588. The machine taken as an example was a writing desk at the service of humanists in libraries, capable of holding several books at once for the reader (
Figure 2).
A rotating system allowed multiple items to be read at the same time. The solution could therefore be a system similar to a rotating conveyer belt, capable of digitizing the same item simultaneously in recto/verso (
Figure 3). A further requirement was sustainability and autonomy in customizing the device. It was necessary to produce a modular scanner whose constituent elements could be replaced by choosing better ones, corresponding to evolving performances, and which were made up of interchangeable parts available on the market without having to go through specific production companies. Last aspect: the possibility of associating the scanner with scanning software produced for the occasion and specific to our collection indexing needs.
The scanner constructed consisted of a circular rotating table 2 m in diameter with four transparent glass plates on which to place the cards. The maximum possible format for scanning was 594 × 420 mm, an A2 format.
The rotary table is controlled by a precision motor with variable speed that allows the uninterrupted digitization of 1 image every 4 s.
A team of two people is required to carry out the scans, one of whom places the images on one of the glass plates and, at each complete turn of the table, two cameras, one positioned above and one below the table at a distance of 180 degrees from each other, first digitize the front and then the back of the article. The document continues its rotation and a second operator picks it up from the shelf. A sensor system calculates the position of the document and detects when it is placed on the glass surface. The data acquired by the cameras are immediately stored in a dedicated server. The cameras are replaceable if necessary: those prepared for the scanner allowed a document resolution of 400 ppi (5424 × 3616 pixels).The flash units were designed and engineered by Factum Arte to provide the lowest possible light level for a high-quality image, minimizing glare.
2.2. Co-Designing the Workflow Inventoring for Digitisation
The first part of the digitised collection, consisting of cardboards (
Figure 4), collected in drawers, did not have an inventory. This case is not unique and that indeed it is very frequent, especially for large collections, not to have inventories or to possess inventories that one may want to re-check because one no longer considers the information given to be reliable [
7].
Therefore the research activities of the ETH and the professional knowledge of the curators conservators of the Cini Foundation had to find a quick but extremely reliable method to index the images to be scanned. The curators have the archival habit of proceeding by indexing item-by-item in a semi-manual way, inserting the information considered key to their institution and/or those that meet needs defined by the need to index each item according to nationally defined criteria, for example by the Ministry of Culture. In the case of a photographic archive, there are also conservation needs as each image is in itself not only representative of the importance of the object represented, but a document of intrinsic patrimonial value [
8].
The card description is therefore often extremely detailed and impose a long and patient work on the curator and archivist to describe the physical and content of the object to be inventoried. In this sense, the practice of inventorying may be at odds with a massive digitization approach whose timeframe does not correspond at all to that required for such a descriptive practice [
9].
Imagining then that each single item could be inventoried with an archival work of 20 min per item, a time that is already very fast and not at all likely, about 6000 items per year would have been indexed (upward estimate) by one person, and it would have taken 166 years to inventory 1M items making up the collection. It is well understood that drawing up a comprehensive inventory prior to digitization makes it effectively impossible to complete the digitization in a reasonable timeframe. The inventor with his manual skills and knowledge represents the bottleneck of the digitization chain. Therefore, if the need is to digitize in order to preserve and conserve, the production of a very detailed inventory in parallel jeopardizes the efficiency of the digitization itself, thus making it impossible to complete digital preservation. It was therefore planned to proceed by successive rounds of post-digitization inventories, gradually becoming more specific and accurate. For this reason, the pre-digitization inventory consisted of documenting the archival position of each item.
It was decided to split the digitization pipeline into two different processes in which the recto and verso of each item underwent two different content extractions (
Figure 5).
The recto of the image file contains two different types of information: the image and a block of text structured in cells with some descriptive fields.
The
verso contains a barcode generated within the pre-inventory of the project, in which the archival location of the scanned item is described (
Figure 6).
This is in fact the key information: the address of the item that makes it possible, in the topographical vastness of an archive, to find it and to validate the equivalence between the scanned image and its material consistency [
10].
Cardboards are physically placed in drawers. It was therefore decided to assign a barcode that identified the archival position relative to the drawer and the position of the cardboard inside the drawer.
In this way, through the automatic reading of the barcode with the scanner software, a json file was generated containing the archival position data associated with the scanned image. The various operators were able to move forward quickly by applying barcodes on the
verso of items, only having to verify the exact correspondence between the item, the archival location and the generated code (
Figure 7).
With regard to any content information important for indexing, the descriptive elements in each file were automatically extracted after assigning semantic labels [
11].
In fact, each cardboard presented a grid in which certain information provided by the curators appeared in a structured manner: place of conservation of the object represented in the photo, photographer’s name, inventory number, title and technique of the work, author (
Figure 8).
This descriptive space was identified and then labelled into categories and then structured into metadata describing the content of each item.
Once digitisation was complete, the question arose of opening up access to the photographic archive to a community that was not local, such as the scholars who frequented the library, but global.
This is why we opted for the creation of an IIIF server that would allow access to the images and files in an international format [
12]. To do this, we hosted the images in an IIIF format by generating a manifest for each physical drawer that collects a number of corresponding documents. Thanks to the manifest, any IIIF vewer can navigate through the original cardboards in the same order in which they are physically located in the physical archive. In digitisation, it is indeed important to respect not only the arborescence of the collection holdings but also the physical ‘topography’ of the digitised objects. Their location, in fact, is not random and follows a curatorial, or historical, order imposed by the logic of the nature of the items preserved. By respecting the visualisation of the actual location in the archive, the organisation of the collection as a historical result and the manifest of the extracted photograph are better understood, effectively linking the image to its primary source in the original archive. Photographs are thus related to each other by their archival position, and this position can only be grasped by visualising the items through the manifest of each drawer, otherwise all images in the collection would be presented at the same logical level, as if it were a flat indexing system.
5. Conclusions
The trajectory of the Replica project illustrates several levels of interdisciplinary collaboration with the objective of designing a new research method in digital art history.
In the first phase, the construction of a new scanner and a new method to transform a physical archive into a digital database, researchers, each expert in their field, with different but complementary backgrounds and experiences, combine their knowledge to produce, through a form of co-design, an innovative approach.
In the second phase, the construction of a search engine fed by the explicit knowledge developed by the art history researchers allows a collective co-construction of a superior knowledge. A feedback loop offers the opportunity for everyone to continuously improve the system.
The last phase, not developed in this article, should extend this virtuous dynamic beyond the restricted community of art history researchers. The space of visual similarity progressively transformed by the information of the Morphograph can be explored by algorithms which try to extract groups of works automatically, connected by visual links, not yet documented by the Morphograph. It would not be a visual search engine, but an automatic system to “discover” pairs of linked works. These discoveries must then be validated by verifying the presence or absence of visual links in the identified works. This is where the general public can get involved. If it takes a great knowledge of art history to successfully query a visual search engine, pursuing a specific line of inquiry, the verification of the presence of a migration of form between two works is potentially more immediate. The last step of the project will be to include them in the project through an extended collaboration between the machine and a large number of communities who are curious and motivated to progress in solving art historical questions. This proves that the construction of specific tools that combine a discipline-specific approach to a more general versatility makes it possible to answer research questions that are potentially transversal to the communities that traditionally study them with their own methods. It is foremost a matter of opening up and making concepts and tools available to suggest answers in which users from different communities benefit, who learn to sharpen their eye for the specific problem, and the community of art historians who benefit from a global expertise on which they can always comment later on the relevance of their selections.
The co-design of the Replica project was characterized by many discussion phases, especially in the design of the final prototype. Finding an initial convergence point on the general pipeline was not difficult, but converging on a material prototype of the final realization required much discussion and rejection of certain interface development hypotheses. In the case of the Replica project, moreover, it was not possible to start from an already existing prototype precisely because it had to be packaged over time thanks to the feedback of the curators’ and researchers’ community, and it was precisely in these latter aspects that its intrinsic value lay. Moreover, in this case, the curators of the photographic archive would probably have welcomed as a non-critical imposition the presence of a prototype, even if only a conceptual one, which could then be perfected by adapting it to the specific case. In fact, an element not to be underestimated in co-design projects for cultural heritage is that all participants wish to feel protagonists and believe they have a specific legitimacy, by authority or knowledge, that cannot be avoided. The development of co-design between institutions that are profoundly different in their dynamic conception of heritage has also made it possible to bypass approaches whereby heritage is conserved as “an unchanging monument of the past” [
24]. and the presence of the community of end users, which in our case are mostly art history researchers, in other projects it is the general public of citizens, has made it possible to respond to a dynamic and functional vision of this heritage [
25]. Indeed, it should not be forgotten that the only way to preserve cultural heritage is to find current and interesting forms for its transmission.