Peter Hofmann1, Paul Lettmayer2, Thomas Blaschke1, Mariana Belgiu1, Stefan Wegenkittl2, Roland Graf2, Thomas Josef Lampoltshammer3, Vera Andrejchenko1. 1. Interfaculty Department of Geoinformatics - Z_GIS, Salzburg University , Schillerstr. 30, Salzburg 5020 , Austria. 2. Department of Information Technology & Systems Management, Salzburg University of Applied Sciences , Salzburg , Austria. 3. Interfaculty Department of Geoinformatics - Z_GIS, Salzburg University, Schillerstr. 30, Salzburg5020, Austria; Department of Information Technology & Systems Management, Salzburg University of Applied Sciences, Salzburg, Austria.
Abstract
Object-based image analysis (OBIA) as a paradigm for analysing remotely sensed image data has in many cases led to spatially and thematically improved classification results in comparison to pixel-based approaches. Nevertheless, robust and transferable object-based solutions for automated image analysis capable of analysing sets of images or even large image archives without any human interaction are still rare. A major reason for this lack of robustness and transferability is the high complexity of image contents: Especially in very high resolution (VHR) remote-sensing data with varying imaging conditions or sensor characteristics, the variability of the objects' properties in these varying images is hardly predictable. The work described in this article builds on so-called rule sets. While earlier work has demonstrated that OBIA rule sets bear a high potential of transferability, they need to be adapted manually, or classification results need to be adjusted manually in a post-processing step. In order to automate these adaptation and adjustment procedures, we investigate the coupling, extension and integration of OBIA with the agent-based paradigm, which is exhaustively investigated in software engineering. The aims of such integration are (a) autonomously adapting rule sets and (b) image objects that can adopt and adjust themselves according to different imaging conditions and sensor characteristics. This article focuses on self-adapting image objects and therefore introduces a framework for agent-based image analysis (ABIA).
Object-based image analysis (OBIA) as a paradigm for analysing remotely sensed image data has in many cases led to spatially and thematically improved classification results in comparison to pixel-based approaches. Nevertheless, robust and transferable object-based solutions for automated image analysis capable of analysing sets of images or even large image archives without any human interaction are still rare. A major reason for this lack of robustness and transferability is the high complexity of image contents: Especially in very high resolution (VHR) remote-sensing data with varying imaging conditions or sensor characteristics, the variability of the objects' properties in these varying images is hardly predictable. The work described in this article builds on so-called rule sets. While earlier work has demonstrated that OBIA rule sets bear a high potential of transferability, they need to be adapted manually, or classification results need to be adjusted manually in a post-processing step. In order to automate these adaptation and adjustment procedures, we investigate the coupling, extension and integration of OBIA with the agent-based paradigm, which is exhaustively investigated in software engineering. The aims of such integration are (a) autonomously adapting rule sets and (b) image objects that can adopt and adjust themselves according to different imaging conditions and sensor characteristics. This article focuses on self-adapting image objects and therefore introduces a framework for agent-based image analysis (ABIA).
Since the beginning of the millennium, two major new technologies have influenced the remote-sensing community: the availability of very high resolution (VHR) remote-sensing data and object-based image analysis (OBIA). Although OBIA builds on several older concepts and methods such as image segmentation, the particular combination of these concepts allow applying multi-scale concepts (Burnett and Blaschke 2003). Both developments mentioned have led to a paradigm change in analysing remote-sensing data: from pixel-based to object-based methods (Blaschke et al.
2014). A widespread assumption is that the latter allows for the analysis of remote-sensing data beyond spectral statistical parameters, using further object properties such as shape and spatial context (Benz et al.
2004, Blaschke 2010). However – while invoking Blaschke and Strobl (2001) – one may first need to ask ‘What’s wrong with pixels?’ Instead of a comprehensive answer, we may refer to the rapidly growing body of literature where a significant number of authors identified an increasing dissatisfaction with pixel-by-pixel image analysis. Although this critique is not new (Cracknell 1998), see also Blaschke and Strobl (2001), Blaschke (2010) and Blaschke et al. (2014) for a more thorough discussion; these authors described a need for applications ‘beyond pixels’ and for specific methods and methodologies that support this.This ready availability of high-resolution multi-band imagery coincided with the increasing awareness in remote-sensing literature that novel methods to extract meaningful and more accurate results were crucial. Likewise, what is fundamentally required in complex image processing tasks is a kind of ‘intelligence’. Here, the authors are very careful about venturing into artificial intelligence. Rather, what is needed is a kind of ‘geo-intelligence’ as described in Hay and Blaschke (2010).In the first years of OBIA, segmentation was regarded to be inextricably linked to this concept. Indeed, segmentation provides the building blocks of OBIA (Hay and Castilla 2008, Lang 2008). Segments are regions that are generated by one or more criteria of homogeneity in one or more dimension (of a feature space). Thus, segments have additional spectral information compared to single pixels. It is, in principle, based on the spectral statistics of the segments’ underlying pixels. However, of even greater advantage than spectral per-object statistics is the additional spatial information for objects (Benz et al.
2004, Hay and Castilla 2008). It has been frequently claimed that this spatial dimension (shape, distances, neighbourhood, topologies, etc.) is crucial to OBIA methods, and that this is a major reason for the remarkable increase in the use of segmentation-based methods in recent times, compared to the use of image segmentation in remote sensing during the 1980s and 1990s (Benz et al.
2004, Blaschke et al.
2014). Still, we can conclude that in very recent literature, segmentation has been seen as less crucial for OBIA. Authors increasingly recognise that the real potential lies in the intelligence and the chance to formulate user knowledge as ‘rule sets’ (Hofmann et al.
2011, Belgiu et al.
2014a, Lang et al.
2014).Although various studies in literature report an increase in spatial and thematic accuracies for OBIA approaches (see the meta-analyses of Blaschke (2010) and Blaschke et al. (2014)), the creation of robust, object-based solutions for automated image analysis of a set of images or even large image archives still remains extremely challenging (Pinz 2005, Walker and Blaschke 2008, Hofmann et al.
2011, Laliberte and Rango 2011, Kohli et al.
2013). Especially the highly complex content of VHR image data and the hardly predictable variability of the objects’ qualities in such diverging image data reduce the robustness and transferability of OBIA rule sets used for classification. Consequently, either the rule sets or the objects’ shape or even both need manual adaptation in order to achieve acceptable results. However, manual interaction and adaptation is deemed to be time consuming, labour-intensive and consequently error-prone. Novack et al. (2014) report on results from transferring a generic knowledge base to two different software packages for OBIA, both of them finally operating with software-specific rule sets. In order to overcome the limited transferability of OBIA rule sets, we investigate the coupling, extension and integration of OBIA with principles and methods from the agent-based paradigm. In particular, this article introduces a framework for agent-based image analysis (ABIA), which extends the existing OBIA concepts and methods by some from the agent-based paradigm.
The principles of OBIA workflow
Recently, the typical workflow of OBIA begins with a more or less arbitrary segmentation of the input data to generate a hierarchical net of image objects followed by an initial classification of the generated image objects. Then, OBIA enters an iterative process of selective segmentation improvements and re-classifications until a satisfactory result is achieved. The criteria on which the subset selection has to operate during iteration can be based on spatial criteria, non-spatial criteria or both (Baatz et al.
2008, Lang 2008). In order to be able to reapply the segmentation and classification process, all processing steps, their procedural sequences and sub-sequences are organised in a rule set. Depending on the software used, rule sets can be described in a domain-specific language (DSL), which is structured as a programming language but uses linguistic elements and concepts of the application domain (Hudak 1996, Fowler 2010). A prominent representative of such a DSL in the context of OBIA is the cognition network language (CNL), which is implemented in the software eCognition® (Athelogou et al.
2007). In CNL, two principal types of rules exist: processing rules, which either calculate values or change the objects’ shape; and classification rules, which assign objects to classes based on defined classification rules. For the latter, hierarchical fuzzy classification schemes complement the usual threshold-based classification and thereby describe each class based on fuzzy membership functions and their combination to fuzzy rules (Benz et al.
2004). The hierarchical classification schemes reflect the object classes’ ontology, which describes the appearance of the real-world objects in the image data at hand (Belgiu et al.
2014b). However, the effort to develop a rule set can be great (Arvor et al.
2013, Belgiu et al.
2014b), and its reusability is limited once the input data changes (Hofmann et al.
2011). In the remainder of this article we therefore introduce a first approximation of a solution to this problem, which aims to integrate concepts of agent-based computing for the adaption of such rule sets and image objects.
The definition of quality in OBIA
Quality of geographic information and therefore of results derived from the analysis of remote-sensing data by definition reflects the minimum properties of an acceptable classification result (ISO 19157:2013). In this context, quality is usually determined by quantifying the correctness and completeness of a classification result. In remote-sensing practice, both aspects are measured by comparing a to-be-evaluated classification result with another classification result that is assumed to be true (Congalton and Green 2009, ISO 191577:2013, Novack et al.
2014). However, this approach suggests an absolute correctness and completeness of the reference classification that is not necessarily given. Additionally, it requires semantically absolute conform class definitions, which in many cases do not exist (Hofmann and Lohmann 2007, Albrecht et al.
2010). Hence, for evaluating the quality of a classification result, it is rather sensible to focus on the requirements a classification result has to meet in order to be accepted by the user. Such quality criteria could be, for example, a minimum allowed deviation from a given – and not necessarily correct – reference map. When using fuzzy classification mechanisms, further requirements can be defined; for example, a minimum allowed ambiguity per object, per class or per scene (Benz et al.
2004, Hofmann et al.
2011).
Software agents and multi-agent systems
Software agents are defined by being flexible and capable of acting autonomously in complex environments. They are thus provided with sensors and effectors in order to interact with their environment and to achieve predefined goals. Coupling several software agents to a multi-agent system (MAS) enables them to interact, communicate and collaborate among themselves in order to achieve either individual or common goals. In such a system, individual agents with different roles can exist, with each of them having role-dependent abilities and goals and all of them being organised in a network of collaborative agents. Although being organised in a network of software agents, each agent decides individually and based on its own strategy, how to act in a particular situation (Figure 1). These fundamental abilities have been shown to allow agent-based systems to deal with complex and unpredictable situations, as well as with incomplete information in a much more flexible and robust manner compared to conventional systems (Wooldridge and Jennings 1995, Jennings 2000).
Figure 1.
Principal components of a software agent.
Principal components of a software agent.
Software agents and MAS in GIS and image analysis
In GIScience, agent-based models (ABMs) are typically used for simulating complex spatiotemporal processes, such as land-use modelling (Parker et al.
2003, Macal and North 2005, Brown et al.
2005, Koch 2007, Marceau and Moreno 2008, Yu and Peuquet 2009). Likewise, agent-based systems (ABSs) are meanwhile widely used in industry for applications, such as process automation, that require a high level of robustness and flexibility (Fazel Zarandi and Ahmadpour 2009, Göhner 2013). However, using software agents or MAS in image analysis is not very common yet. In the remote-sensing domain, Samadzadegan et al. (2009) and (2010) report on the application of ABS for building detection from LiDAR data, whereas for image analysis in the life science domain, Bovenkamp et al. (2004) describe a similar approach for object detection in intra-vascular ultra-sound (IVUS) images and Rodin et al. (2004) demonstrate the application of agent-based systems in biological image analysis. Mahmoudi et al. (2013) describe an ABS designed to improve OBIA classification results in urban areas. Here, different tasks of object recognition are distributed among respectively defined agents. The agents then simultaneously operate on the image and share their (intermediate) results. The authors demonstrate that software agents can principally be used to parallelise image processing tasks and show how simultaneously arising individual classification results can be improved and condensed by agent-based systems as compared to conventional classification techniques.
Ontologies in GIS, OBIA and the agent-based paradigm
Ontologies in GIS, OBIA and agent-based programming play a central role in the design of geo-databases, rule sets and that of software agents. In all the three domains, they explicitly describe those parts of the real world that are relevant for the respective domain and application. Especially in OBIA, the ontology acts as the foundation for the rule set, its object classes and all the classes’ semantic constraints. It describes the object classes and their semantics as independent from any underlying image data as possible. However, the image-specific appearance of particular classes in particular image data can only be described by a rule set that uses the structure of the ontology but is simultaneously designed for the data in use. Belgiu et al. (2014b) describe a method for converting an ontology expressed in the Web Ontology Language 2 (OWL 2, see Motik et al.
2009) automatically to a framework for an OBIA rule set. OWL 2 is a recommendation of the World Wide Web Consortium (W3C) that has been widely adopted by the Semantic Web Initiative. The remote-sensing ontologies extend the semantics of target classes such as land cover classes, informal settlements or refugee camps with properties that allow their detection in the remotely sensed data at hand. These properties are acquired either from literature (Belgiu et al.
2014c) or by using data mining techniques (Belgiu et al.
2014a, Maillot et al.
2004). In agent-based computing for the design of software agents, the so-called belief-desire-intention (BDI) model is very common (Rao and Georgeff 1995). In this context the ontology serves as the world-model for agents. That is, the ontology describes the agents’ beliefs. As recent studies showed (Viezzer 2006), this world model is not necessarily static. Moreover, software agents can be designed to adapt their world-model according to changing environmental conditions as a kind of an individual survival strategy.
Integrating OBIA and the agent-based paradigm
The concept for ABIA proposed in this article focuses on integrating mechanisms of agent-based control (ABC) as applied in process automation and ABM with OBIA. In particular, a framework for autonomous and adaptive control of OBIA using software agents in order to increase the robustness of particular OBIA solutions is introduced. For this purpose, two principal and independent approaches in the MAS framework are possible:Extending the image object hierarchy as known from OBIA to a hierarchically organised MAS of networked image object agents (IOAs) with autonomous adaptation and interaction capabilities.Autonomously adapting existing rule sets by means of a MAS of rule set adaptation agents (RSAAs) in order to robustly analyse varying image data without any need for further human interaction.In the following we focus on the first approach. In this approach, we extend the well-known OBIA concept of a hierarchical net of image objects to a hierarchical net of IOAs. Within this network, each image object aka image segment can act and interact in agent-based manner, that is, act and react, cooperate and negotiate with other agents in order to achieve its goals (Figure 2).
Figure 2.
Principal workflow for ABIA with MAS operating with RSAAs (top) and IOAs (bottom).
Principal workflow for ABIA with MAS operating with RSAAs (top) and IOAs (bottom).
Conceptual framework
In the context of ABIA, each MAS must balance its activities between ontology conformity and quality requirements. Consequently, in the course of adaptation, each MAS must detect to what degree the quality requirements are met by particular adaptations and autonomously decide whether the adaptation process needs to be continued or not. By logging the success of each adaptation action, ‘promising’ adaptation strategies can be learnt or even the agent’s ontology can be adapted. That is, particular actions – preferably those providing the least ontology violation together with best achievable quality – are being prioritised for future applications. In the ABIA framework, this balancing and evaluation task can be either performed on an agent individual level, or by one or several dedicated control agents (CAs). The major roles of CAs are then:Avoiding contradictions between the ontology and intended agent actions.Compare the achieved quality with the user-defined quality for acceptance.Trigger and abort adaptation processes as necessary; for example, abort if quality criteria are achieved.Learn the most promising strategies for goal achievement.
Design of a MAS with IOAs
In a MAS consisting of IOAs, it is the image objects aka IOAs themselves that adapt autonomously according to changing image data. Similar to OBIA, in this architecture, image objects evolve to a hierarchical net of IOAs wherein each IOA is connected with its neighbour agents, its higher-level (super-) agents and its lower-level (sub-) agents (Figure 3).
Figure 3.
Hierarchical net of IOAs after segmentation and classification.
Hierarchical net of IOAs after segmentation and classification.After initial segmentation and classification steps, each IOA compares its degree of compliance with the ‘antetype’ of the class it was initially assigned to and as it is defined in the ontology. The goal of each IOA is to meet its ‘antetype’ as best as possible. For this purpose, each IOA develops its individual strategy to achieve this goal, whereas action priorities can be pre-defined in general or depending on the initial class assignment of an IOA. In principle, each IOA has two options to act: (1) re-segment itself and (2) merge with neighbouring IOAs. Option (2) is sensible in situations where merging of neighbour-IOAs would improve the overall classification quality (over-segmentation), but it implies that at least one of the involved IOAs will dissolve itself. Option (1) can be manifold, ranging from sub-segmentation via shrink-and-grow methods to negotiations about border pixels with neighbour IOAs. The prioritisation of particular actions can depend on the IOA’s class assignment and its grade of goal achievement. For example, an IOA classified as a ‘house’ but with borders that are not fully straight-lined might intend to execute a respective straight-lining algorithm. Since these operations can lead to conflicts between neighbouring objects (for example, between a ‘house’ and a ‘forest’ IOA), using CAs to prioritise individual actions is sensible (Figure 4).
Figure 4.
Architecture of an IOA with sensors to detect its neighbourhood and to receive messages from other agents. Its effectors allow the IOA to send messages and to change its own shape including dissolving itself.
Architecture of an IOA with sensors to detect its neighbourhood and to receive messages from other agents. Its effectors allow the IOA to send messages and to change its own shape including dissolving itself.Orthoimage and generated DSM; red rectangular indicates subset under investigation (top). Image data (middle left) and DSM superimposed to hill shaded relief (middle right); slope and slope of slope (bottom, both in degree) for subset area.The adaptation process stops either if the user defined quality requirements are met or if a user-defined threshold for the total number of adaptations has been reached. In order to avoid the system being trapped in a cyclic sequence of adaptation steps, a snapshot mechanism is sensible, which avoids endless loops.
Subset image data (top left), manual reference (top right), initial (bottom left) and final (bottom right) result after ABIA processing, with membership degrees to ‘roof’ per object.
Data
The simulation has been applied on an orthorectified image from Weilheim, Upper Bavaria, Germany, taken in May 2010. The orthoimage together with a digital surface model (DSM) was generated based on a stereo pair of the RGB-bands captured by an UltraCam XP (www.ultracamx.com) using software from SimActive (www.simactive.com). The spatial resolution of the orthoimage is at 8 cm, that of the DSM at 35 cm. The radiometric resolution of the optical data is at 8bit. From the DSM, the slope and slope of slope (change of slope) were calculated per pixel and expressed in degree from 0° to 90°. The subset under investigation has a size of 1311 × 869 pixels (Figure 5).
Figure 5.
Orthoimage and generated DSM; red rectangular indicates subset under investigation (top). Image data (middle left) and DSM superimposed to hill shaded relief (middle right); slope and slope of slope (bottom, both in degree) for subset area.
Ontology
To describe semantically how buildings look like in general when using the above-described data, we have developed an ontology in accordance with Lampoltshammer and Heistracher (2014), Belgiu et al. (2014c) and Durand et al. (2007). In particular our ontology denotes the buildings’ roof shape, that is, their form and colour and the buildings’ relative height to their neighbouring objects. In contrast to the above-mentioned authors, in our building ontology colour and local elevation difference together with the elevation variability is included (Figure 6).
Figure 6.
Ontology for building description in the used image and elevation data. Relations with numeric values are to be understood as fuzzy relations.
For initial classification, we have developed a simple rule set to delineate roofs based on the ontology as described in Section 3.2. It starts with a multi-resolution segmentation (MRS) as described by Baatz and Schäpe (2000) whereas all three bands, the DSM and the slope have been weighted equally for the segmentation. Respectively, the homogeneity criteria of the MRS are composed equally of the RGB bands, the DSM and the slope. The brightness per object is calculated only based on the RGB-bands. To assign the initially created objects to a ‘roof’ class, a fuzzy class hierarchy has been defined with classification rules as depicted in Table 1. Applying the above-described rule set to the data leads to an initial classification result as depicted in Figure 8.
Applying the simulated IOA-MAS as described in the section before to the initial result generated with the data and rule set described in Sections 3.1 and 3.3 led to the final result as displayed in Figure 8. We have run the ABIA process for 100 tics. In order to document the development of the classification quality, after each tic the intermediate result (see Appendix) underwent a per-pixel accuracy assessment based on correctly assigned pixels (true positives and negatives) and wrongly assigned pixels (false positives and negatives) using the manual reference image as depicted in Figure 8. Additionally, some quality assessment measures based on the membership degree of each object are calculated for each intermediate result. As can be seen from the development plots, after 17 tics no significant changes are observable (Figure 7). Thus, in the following discussion we are focusing on the first 17 tics.
Figure 7.
Development of accuracy measures during processing for 0–17 tics (left) and 0–100 tics (right).
Development of precision, recall, accuracy and F-score during the ABIA process.
Development of precision, recall, accuracy and F-score during the ABIA process.As can be seen from Figure 9, the accuracy first decreases (tic No. 1–4), but then increases until it saturates at tic No. 12 at a level of 0.9037. Similarly, recall even continuously increases already from the very beginning, while precision more or less inversely decreases in the same period of tics; the F-score behaves similarly to recall. This indicates that the overall classification has increased during the ABIA process. Regarding the classification quality based on the membership degrees per object, this has improved as well (Figure 10).
Figure 10.
Development of accuracies based on membership degrees per object. Mean membership to ‘roof’ (upper left), number of objects with a membership degree of µ > 0.5 to ‘roof’ (upper right), mean membership degree of objects with more than 50% true positive pixels (bottom left) and more than 50% false positive pixels (bottom right).
Development of accuracies based on membership degrees per object. Mean membership to ‘roof’ (upper left), number of objects with a membership degree of µ > 0.5 to ‘roof’ (upper right), mean membership degree of objects with more than 50% true positive pixels (bottom left) and more than 50% false positive pixels (bottom right).The mean µ of all objects with a membership to ‘roof’ of µ > 0.0 (mean µ per class) increases from 0.0916 to its final saturated level of 0.1874, whereas the mean µ per class also includes wrongly assigned objects (false positives) and objects with a relatively low membership degree. The number of ‘roof’ objects with µ > 0.5 develops from n = 3 (tic No. 1) to n = 6 (tic No. 10) whereas in tic No. 8 and tic No. 9 it decreases to n = 5 but then returns to n = 6. This indicates that the number of classified ‘roof’-objects with low fuzziness, that is, with a relatively clear class assignment, has increased.Comparing the developments with the reference classification, obviously a per-object improvement is observable, since the number of objects with a minimum amount of 50% correctly classified pixels remained constant (n = 5) but their average membership degree to ‘roof’ has increased and remains at a relative high level (mean µ = 0.85). Vice versa, the initial number of wrongly assigned objects (more than 50% of false positive pixels) increases and they could improve their average membership degree, too, but on a far lower level (mean µ = 0.40) than the correctly assigned objects. Similarly, the obviously largest false negative object, that is, the western part of the big central building, develops positive (µ = 0.0 in tic No. 1 to µ = 0.016 in tic No. 11 and the following tics). However, this object could not improve further, because if the eastern ‘roof’ object of the centre building (µ = 0.98 in tic No. 17) would merge with it, the resulting object would have a decreased membership to ‘roof’. Similarly, a growth of this object would decrease the membership of the neighbouring ‘roof’ objects. Moreover, when comparing the memberships of the final result with those of the reference map (Figure 8), the manually classified objects (except the cut-off object in the north-east) have lower membership degrees to ‘roof’ than some of the automatically processed ‘roofs’. Especially for the building in the centre-east this is obvious. This indicates, that (1) the ontology and therefore the class hierarchy is not absolutely in compliance with our perception and (2) data quality leads to misinterpretations by the classifier. The latter is obvious for shaded higher vegetation, since here elevation information and spectral information for vegetation and buildings are similar due to lower brightness. Especially in the case of the shaded vegetation bordering north to the centre building (µ = 0.69 in tic No. 17), it has increased its membership to ‘roof’, since the shade of the bordering building improves its shape in terms of being ‘roof’-like. For this object during the ABIA process, the more shade (of the building) it has accumulated the more its shape criteria for ‘roof’ was fulfilled. Nevertheless, for all false positive objects including both wrongly assigned vegetation objects, their membership degree to ‘roof’ is relatively low. Additionally, when examining Figure 5, at some positions, simply the DSM is very inaccurate. Especially at the buildings’ borders, this effect led to a slight overgrowing of, and in the worst case to a merger with, the already wrongly assigned ‘roof’ objects, as is the case in the south-eastern jutty of the central building.
Authors: Thomas Blaschke; Geoffrey J Hay; Maggi Kelly; Stefan Lang; Peter Hofmann; Elisabeth Addink; Raul Queiroz Feitosa; Freek van der Meer; Harald van der Werff; Frieke van Coillie; Dirk Tiede Journal: ISPRS J Photogramm Remote Sens Date: 2014-01 Impact factor: 8.979