AUTOMATICALLY DETERMINING SCALE WITHIN UNSTRUCTURED POINT CLOUDS
Three dimensional models obtained from imagery have an arbitrary scale and therefore have to be scaled. Automatically scaling these models requires the detection of objects in these models which can be computationally intensive. Real-time object detection may pose problems for applications such as i...
Saved in:
Published in | International archives of the photogrammetry, remote sensing and spatial information sciences. Vol. XLI-B3; pp. 617 - 624 |
---|---|
Main Authors | , |
Format | Journal Article Conference Proceeding |
Language | English |
Published |
Gottingen
Copernicus GmbH
09.06.2016
Copernicus Publications |
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Three dimensional models obtained from imagery have an arbitrary scale and therefore have to be scaled. Automatically scaling these models requires the detection of objects in these models which can be computationally intensive. Real-time object detection may pose problems for applications such as indoor navigation. This investigation poses the idea that relational cues, specifically height ratios, within indoor environments may offer an easier means to obtain scales for models created using imagery. The investigation aimed to show two things, (a) that the size of objects, especially the height off ground is consistent within an environment, and (b) that based on this consistency, objects can be identified and their general size used to scale a model. To test the idea a hypothesis is first tested on a terrestrial lidar scan of an indoor environment. Later as a proof of concept the same test is applied to a model created using imagery. The most notable finding was that the detection of objects can be more readily done by studying the ratio between the dimensions of objects that have their dimensions defined by human physiology. For example the dimensions of desks and chairs are related to the height of an average person. In the test, the difference between generalised and actual dimensions of objects were assessed. A maximum difference of 3.96% (2.93cm) was observed from automated scaling. By analysing the ratio between the heights (distance from the floor) of the tops of objects in a room, identification was also achieved. |
---|---|
ISSN: | 2194-9034 1682-1750 2194-9034 |
DOI: | 10.5194/isprs-archives-XLI-B3-617-2016 |