TU Darmstadt / ULB / TUbiblio

Procedural Descriptions for Analyzing Digitized Artifacts

Ullrich, Torsten ; Schinko, Christoph ; Schiffer, Thomas ; Fellner, Dieter W. (2013)
Procedural Descriptions for Analyzing Digitized Artifacts.
In: Applied Geomatics, 5 (3)
doi: 10.1007/s12518-013-0107-7
Artikel, Bibliographie

Kurzbeschreibung (Abstract)

Within the last few years, generative modeling techniques have gained attention especially in the context of cultural heritage. As a generative model describes a rather ideal object than a real one, generative techniques are a basis for object description and classification. This procedural knowledge differs from other kinds of knowledge, such as declarative knowledge, in a significant way: It is an algorithm, which reflects the way objects are designed. Consequently, generative models are not a replacement for established geometry descriptions (based on points, triangles, etc.) but a semantic enrichment. In combination with variance analysis techniques, generative descriptions can be used to validate reconstructions. Detailed mesh comparisons can reveal smallest changes and damages. These analysis and documentation tasks are needed not only in the context of cultural heritage but also in engineering and manufacturing. Our contribution to this problem is a work flow, which automatically combines generative/procedural descriptions with reconstructed artifacts and performs a nominal/actual value comparison. The reference surface is a procedural model whose accuracy and systematics describe the semantic properties of an object, whereas the actual object is a real-world data set (laser scan or photogrammetric reconstruction) without any additional semantic information.

Typ des Eintrags: Artikel
Erschienen: 2013
Autor(en): Ullrich, Torsten ; Schinko, Christoph ; Schiffer, Thomas ; Fellner, Dieter W.
Art des Eintrags: Bibliographie
Titel: Procedural Descriptions for Analyzing Digitized Artifacts
Sprache: Englisch
Publikationsjahr: 2013
Titel der Zeitschrift, Zeitung oder Schriftenreihe: Applied Geomatics
Jahrgang/Volume einer Zeitschrift: 5
(Heft-)Nummer: 3
DOI: 10.1007/s12518-013-0107-7
Kurzbeschreibung (Abstract):

Within the last few years, generative modeling techniques have gained attention especially in the context of cultural heritage. As a generative model describes a rather ideal object than a real one, generative techniques are a basis for object description and classification. This procedural knowledge differs from other kinds of knowledge, such as declarative knowledge, in a significant way: It is an algorithm, which reflects the way objects are designed. Consequently, generative models are not a replacement for established geometry descriptions (based on points, triangles, etc.) but a semantic enrichment. In combination with variance analysis techniques, generative descriptions can be used to validate reconstructions. Detailed mesh comparisons can reveal smallest changes and damages. These analysis and documentation tasks are needed not only in the context of cultural heritage but also in engineering and manufacturing. Our contribution to this problem is a work flow, which automatically combines generative/procedural descriptions with reconstructed artifacts and performs a nominal/actual value comparison. The reference surface is a procedural model whose accuracy and systematics describe the semantic properties of an object, whereas the actual object is a real-world data set (laser scan or photogrammetric reconstruction) without any additional semantic information.

Freie Schlagworte: Business Field: Digital society, Forschungsgruppe Semantic Models, Immersive Systems (SMIS), Computer aided design (CAD), Cultural heritage, Ray tracing, Ray casting, Reverse engineering, Visualization
Fachbereich(e)/-gebiet(e): 20 Fachbereich Informatik
20 Fachbereich Informatik > Graphisch-Interaktive Systeme
Hinterlegungsdatum: 12 Nov 2018 11:16
Letzte Änderung: 04 Feb 2022 12:40
PPN:
Export:
Suche nach Titel in: TUfind oder in Google
Frage zum Eintrag Frage zum Eintrag

Optionen (nur für Redakteure)
Redaktionelle Details anzeigen Redaktionelle Details anzeigen