by Jean-Baptiste Barreau (CNRS/CReAAH UMR 6566), Ronan Gaugne (Université de Rennes 1/IRISA-Inria) and Valérie Gouranton (INSA Rennes/ IRISA-Inria)
A point cloud is the basic raw data obtained when digitizing cultural heritage sites or monuments with laser scans or photogrammetry. These data represent a rich and faithful record provided that they have adequate tools to exploit them. Their current analyses and visualizations on PC require software skills and can create ambiguities regarding the architectural dimensions. We propose a toolbox to explore and manipulate such data in an immersive environment, and to dynamically generate 2D cutting planes usable for CH documentation and reporting.
Due to their states, complexity and archaeological interest, the two subjects studied in this work are Breton architectural sites for which the development of new analytical techniques appears quite appropriate. The first is the chapel of Languidou, built in the middle of the 13th century in the municipality of Plovan, which seems to be the “founding element” of a religious architectural style. The second building addressed in this work is the “jeu de paume” court of Rennes, built at the beginning of the 17th century and registered as a Historical Monument. For several decades, the study of these kinds of architectural styles and reorganizations has been carried out by archaeologists by producing different types of 2D documentation: plans, drawings, sections, profiles and orthophotos. 3D documentation comes from classical topographic survey, laser scanning or photogrammetry. In the case of the French Grand-Ouest, some of this documentation has been carried out within the scope of the West Digital Conservatory of Archaeological Heritage [1]. Until now, an engineer has performed the segmentations of the 3D documentation on a PC and tries to better meet the archaeologist’s expectations, who is sometimes unwilling to use new technologies and frustrated by his lack of autonomy. The objective of this work is to involve the archaeologist more deeply in this proces by immersing and allowing him to segment, in real time and on 1:1 scale, the 3D survey.
The first step consists in loading the point clouds of the architectural sites into the Virtual Reality platform Immersia [2]. The scans were done in June 2013 and June 2014, with a Leica ScanStation C10 and a Focus3D X330, and were integrated a few months later (see Figure 1). The data structure of the points displayed in our virtual reality device is a billboard with 3 coordinates, a colour and a scalar. The files are in a binary format that we designed for Unity. For a correct exploration within the Immersia, a subsampling was systematically done and two cloud loading modes were implemented. For the first mode, the cloud is distributed over several hundred octrees. They are loaded dynamically and are visible from the user’s point of view, thanks to a culling technique (cf Figure 2). The second mode consists in the use of a Level-of-Detail technique. The number of points loaded at the start in a single file is smaller and their size is fixed. The selection of cloud segmentation tools is done through a MiddleVR menu that allows the user to interact with three parameters (cf Figure 3). When the first mode is active, it is possible to modify the size of the points. On a more global scale, the user can switch between the two loading modes and change the display distance to the cloud. Concerning the “cutting plane”, it is possible to display or hide it, change its thickness and assign a unique color to the points contained in it. It is also possible to change the opacity of points outside the cutting plane. Its manipulation within Immersia is done thanks to a Flystick which modifies its translations and rotations. The display in the Immersia platform (MiddleVR) has a frame rate which can reach 30fps. This result is two to three times less fluent than that on PC (Unity / MiddleVR). The rendering of the 2D resulting plane is done with a camera orthogonal to the 3D cutting plane and can displayed on a tablet (cf Figure 3). The user can also adjust the distance, field of view and roll of the camera. A scale is also generated automatically on the resulting 2D plane, in order to provide a correct understanding of the dimensions.
Figure 1: Point clouds of the chapel of Languidou and the “jeu de paume” court.
Figure 2: Octree partitioning and culling used on “jeu de paume” court points cloud.
Figure 3:"Cutting plane" creation process by an operator in the Immersia platform.
We now need to have the tool tested by a large community of archaeologists, in particular to check if the use of the Flystick to move the cutting plane is sufficiently precise. At the same time, we still have to improve the optimization of the point clouds management. Concerning the reduction of the I/O time and disk space, it will be necessary to store the points coordinates on 2 bits (instead of 4), to downgrade the precision to the millimetre unit, to decrease the entropy, and to use zstd compression. The use of culling occlusion, at native Unity or hardware level, is also being studied to improve optimization. Finally, the management of additional data associated with points is a major issue. Their storage, visualization and linking seems indeed to be a very interesting long-term prospect.
References :
[1] J.-B. Barreau, R. Gaugne, Y. Bernard, G. Le Cloirec, and V. Gouranton, “The West Digital Conservatory of Archaelogical Heritage Project,” in DH, (France), pp. 547–554, 2013.
[2] R. Gaugne, V. Gouranton, G. Dumont, A. Chauffaut, and B. Arnaldi, “Immersia, an open immersive infrastructure: doing archaeology in virtual reality,” Archeologia e Calcolatori, supplemento 5, pp. 180-189, 2014.
Please contact:
Jean-Baptiste Barreau
CNRS/CReAAH UMR 6566, Rennes, France
Ronan Gaugne
Université de Rennes 1/IRISA UMR 6074, Rennes, France
Valérie Gouranton
INSA Rennes/IRISA UMR 6074, Rennes, France