Literature DB >> 35726003

Research on the visual image-based complexity perception method of autonomous navigation scenes for unmanned surface vehicles.

Binghua Shi1, Jia Guo2, Chen Wang3, Yixin Su4, Yi Di1, Mahmoud S AbouOmar5.   

Abstract

To solve the long-tail problem and improve the testing efficiency for autonomous navigation systems of unmanned surface vehicles (USVs), a visual image-based navigation scene complexity perception method is proposed. In this paper, we intend to accurately construct a mathematical model between navigation scene complexity and visual features from the analysis and processing of image textures. First, the typical complex elements are summarized, and the navigation scenes are divided into four levels according to whether they contain these typical elements. Second, the textural features are extracted using the gray level cogeneration matrix (GLCM) and Tamura coarseness, which are applied to construct the feature vectors of the navigation scenes. Furthermore, a novel paired bare bone particle swarm clustering (PBBPSC) method is proposed to classify the levels of complexity, and the exact value of the navigation scene complexity is calculated using the clustering result and an interval mapping method. By comparing different methods on the classical and self-collected datasets, the experimental results show that our proposed complexity perception method can not only better describe the level of complexity of navigation scenes but also obtain more accurate complexity values.
© 2022. The Author(s).

Entities:  

Mesh:

Year:  2022        PMID: 35726003      PMCID: PMC9209519          DOI: 10.1038/s41598-022-14355-y

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.996


Introduction

As various new technologies related to autonomous navigation ships have emerged recently, the main development direction for ships is to be human-free, smart, and intelligent[1]. Some advanced environmental perception, path planning, and motion control methods assist the development of unmanned surface vehicles (USVs)[2-4]. However, to realize accurate remote automatic or autonomous navigation of USVs, a large amount of practical testing is also needed to verify the safety and reliability of their navigation systems[5]. If a USV directly conducts remote autonomous and unmanned driving test experiments in public waters, it will not only be extremely inefficient and costly but will create greater safety problems for themselves and their surrounding navigable vessels[6]. Therefore, it is necessary to assess various navigation scenarios for a USV before conducting real driving test experiments. Navigation scenario construction technology based on visual sensors has the characteristics of low cost, mature solutions and a high degree of scene restoration, which makes this technology widely used to reconstruct ship navigation scenes in a comprehensive and multiview way[7-9]. Vision-based navigation scene perception provides visual semantic information to human beings that approximate eyesight, which makes this method likely to be widely used in the field of autonomous navigation[10,11]. The greatest challenge in using visual images to construct navigation scenarios is that it is difficult to cover navigation scenes with a small probability of occurrence but high risk, i.e., the “long-tail problem”[12]. Usually, to essentially solve the long-tail problem of scenario coverage, it is necessary to improve the diversity and complexity of the constructed scenarios, and diversity can be obtained by artificially selecting different test conditions. Generally, one of the feasible approaches to address the long-tail problem is to increase the diversity and complexity of navigation scenes[13]. The diversity can be increased by artificially selecting different test conditions. However, the complexity of the navigation scene is a relatively subjective feeling, and a unified perception standard is urgently needed. Furthermore, the perceived complexity of an autonomous driving test system can reflect the difficulty of the construction scenes and the good or poor ability of USVs. The presentation of navigation scenes based on visual sensors is accomplished through continuous image sequences or videos. The complexity perception method can be designed by referring to the related works in the image engineering field. The main idea of these works is to let a computer simulate human visual perception to make quantitative decisions on the perceived visual complexity of images. The common complexity perception approaches include the following: information entropy[14], the average information gain[15], and the compression rate of significant regions[16]. Other methods include quantitative perception methods using image features, such as grayscale, color, edge, and texture[17], and then using machine learning, neural networks, and other classification regression methods to obtain the complexity perception results. Among the features, textural features are one of the most commonly used feature objects in image content complexity calculations. For example, Guo et al.[18] used the textural features of regularity, orientation, density, roughness, and familiarity to describe image complexity. Chen et al.[19] applied BP neural networks to train five index weight coefficients, including the energy, entropy, contrast, homogeneity, and correlation and then built an image complexity perception model. The experimental results showed that their proposed model can accurately describe the complexity of images. Zhan et al.[20] proposed an autonomous visual perception method for unmanned surface vehicle navigation based on the local color and texture features. Experiments in lakes with various scenes demonstrate that the proposed method can recognize unknown environments. To assess the complexity of an image, machine learning has two main categories of methods: regression and classification[21]. The regression-based methods include linear or nonlinear fitting[22] and random forest fitting algorithms[23], which have the disadvantage of being prone to overfitting when the size of the dataset is small. The classification-based methods include support vector machines (SVMs)[24,25], BP neural networks[21], and convolutional neural networks (CNNs)[26,27], but these methods are also more demanding on datasets and require manual labeling. Different from the above methods, unsupervised clustering-based methods for image complexity perception have also been proposed[28]. These methods use K-means or particle swarm optimization (PSO) algorithms to cluster data samples with the same characteristics together for the purpose of complexity assessment. The above approaches in the image engineering field tend to qualitatively describe images using the overall image complexity and lack results focusing on the complexity analysis of USV navigation scenes. In addition, the complexity result is easily affected by various aspects, such as light irradiation, object surface reflection and the internal performance parameters of an imaging sensor. In particular, the complexity of a scene varies greatly under different navigation conditions, which increases the difficulty of calculating and perceiving the complexity of navigation scenes. Hence, this paper considers the special scenarios for USV, analyzes the typical complex elements, proposes the complexity perception method of a navigation scene based on image texture features, optimizes the unsupervised PSO features clustering algorithm and explores the feasibility of our proposed method on classical and self-harvested datasets. The remainder of this paper is organized as follows. Section “Vision-based navigation scene” presents and analyzes some typical complex elements for vision-based navigation scenes. Section “Design of the complexity perception method” proposes our novel algorithms and their implementation, and the results and discussions are given in Sections “Experiments and results” and “Discussion”, respectively. Finally, some conclusions and future research are discussed in Section “Conclusions”.

Vision-based navigation scene

Typical complex elements

To cover as many test scenarios as possible and solve the long-tail problem when constructing navigation scenarios for USVs, the typical complex elements that need to be focused on are systematically determined. Before a USV is driven for testing, it is necessary to have knowledge of the type of water to be navigated, i.e., to determine whether the navigation scenes are in open water or restricted water. Open water usually refers to a sea with an open view, as shown in Fig. 1a. Conversely, restricted water includes inland rivers or ports with heavy traffic[29]. Typical complex elements usually occur in restricted waters because there are many uncertain features during navigation testing. A classification of these features reveals that a complex navigation scene contains the following typical elements:
Figure 1

Examples of open water and 5 typical complex elements.

Examples of open water and 5 typical complex elements. Cluttered background A cluttered background will occupy a certain proportion of the navigation scenes, and the pixel values of the corresponding area will change irregularly once the test scenarios include complex elements[30]. For example, distant city buildings and plants on the shore are shown in Fig. 1b. These cluttered backgrounds do nothing but increase the complexity of navigation scenes and interfere with the environmental perception of USVs. Dynamic obstacles. Affected by the environment or a ship’s movement, navigation scenes contain disturbances caused by dynamic obstacles and a static background. For instance, moving ships (as shown in Fig. 1c), floating obstacles and other traffic participants can increase the risk and make a navigation scene more complex[31]. Camera shaking Vision-based navigation scenes have to consider the camera shaking phenomenon, which is caused by wind, waves and currents on the water. In addition, the phenomenon is even more serious when a ship is moving[32]. Camera shaking can lead to skewed, rotated or even distorted scene images, adding to the complexity, as shown in Fig. 1d. Cast shadows The sun, clouds, ships and buildings in a navigation scene can all produce cast shadows, which can be classified as self-shadows and projected shadows. A self-shadow is formed by the part of light being blocked directly, and a projected shadow is formed by an object being projected onto the water surface, i.e., the reflection of the sun setting (as shown in Fig. 1e), clouds, trees and buildings on the shore. Special weather conditions Rain, snow, night and fog will affect the complexity of a navigation scene. Special weather conditions, such as rain and snow, lead to different scene performances. For example, raindrops and snowflakes are easily captured and thus interfere with the perception of other targets. Night or foggy conditions will lead to a decrease in imaging contrast, blurred images and loss of details. This is shown in Fig. 1f.

Division of levels of complexity

To determine the complexity of a navigation scene, we judge whether the scene is complex according to whether it contains the typical elements and how many elements are included to divide the levels of complexity. The specific case method is shown in Table 1.
Table 1

Description of the levels of complexity.

No.Abbr.LevelsDescription
Case 1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_{N}$$\end{document}LNNo complexityDoes not contain any typical elements
Case 2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_{L}$$\end{document}LLLow complexityContains any of typical elements
Case 3\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_{M}$$\end{document}LMMedium complexityContains any two typical elements
Case 4\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_{H}$$\end{document}LHHigh complexityContains three or more typical elements.
Description of the levels of complexity. The example images in Fig. 2 are taken from the marine obstacle detection dataset (MODD)[33]. The no-complexity scene in Fig. 2a shows clear weather and an open view, which is basically free of shadows and does not affect the safe navigation of the USV. Fig. 2b contains projected shadow elements on the shoreline that form reflections on the water, causing this navigation scene to appear slightly less complex. Figure 2c contains the cluttered background and cast shadow elements in which the dark areas in the image of the navigation scene are relatively large, showing a medium-complexity scenario. Figure 2d contains not only a cluttered background and cast shadow elements but also dynamic obstacles, and the navigation scene is highly complex under the influence of three typical elements.
Figure 2

Examples of levels of complexity.

Examples of levels of complexity.

Design of the complexity perception method

Methodology framework

As the integral link of autonomous navigation testing, the complexity perception of a navigation scene based on a visual sensor relies on feature extraction and processing. The presence of the abovementioned typical complex elements makes the frame images captured by visual sensors exhibit different textural features, and the extraction and clustering of these features are closely related to the results of the complex perception of the navigation scene. The complexity perception results are reflected in two aspects: level determination and complexity value calculation. As shown in Fig. 3, our complexity perception of the navigation scene includes the following steps: (1) dataset preparation, (2) feature extraction, (3) feature clustering, and (4) complexity value calculation. The specific implementation process of each step is as follows.
Figure 3

The framework of the complexity perception method.

The framework of the complexity perception method.

Texture feature parameter extraction

Image texture information is one of the important bases to reflect whether a navigation scene is complex. Referring to the related results in the field of image complexity analysis, the textural features can be obtained using the gray level cogeneration matrix (GLCM) and Tamura methods. By computing the GLCM of an image to obtain its feature information, Haralick[34] proposed 14 statistical feature parameters, including the energy, entropy, contrast, homogeneity, correlation, and variance. Tamura’s method is based on the human visual perception of texture, and Hideyuki Tamura[35] approximated six basic textural features in computational form: coarseness, contrast, directionality, line-likeness, regularity, and roughness. Combining the above methods results in 20 feature parameters describing the complexity of a navigation scene from different aspects, but there is an overlap problem and redundancy among these feature parameters. Therefore, this paper selects five textural parameters with low correlation that are easy to calculate, including the energy, contrast, inverse difference moment, correlation, and roughness, for the complexity perception of a navigation scene.

Feature parameters based on a GLCM

Assume an navigation scene I with gray grades, and are two pixel points in scene I with distance d in the direction of . Then, the GLCM of this navigation scene is calculated as follows:where denotes the number of elements in the set. represents the gray levels of two pixels. The energy (ASM) is used to describe the uniformity of the distribution of the navigation scene. When the elements are concentrated near the diagonal of the GLCM, a smaller value of ASM indicates that the grayscale distribution is more uniform and the texture is finer; conversely, it indicates that the grayscale distribution is uneven and the texture is rougher. The contrast (CON) is used to reflect the depth of image texture grooves and clarity. In a particular navigation scene, a clearer image texture means a larger value of CON, and the opposite means a smaller value. The inverse difference moment (IDM) is a statistical feature parameter that reflects the local texture of a navigation scene. When the value of IDM is large, it indicates that the textures of different regions in this navigation scene are more homogeneous. The correlation (COR) is used to measure the similarity of the GLCM elements in the row or column direction. When the row or column similarity is high, the value of COR is larger, and the complexity of the scene is smaller. The opposite also holds.where and denote the mean values of the elements along the normalized GLCM in the row and column directions, respectively. and represent their mean squared values.

Feature parameter based on Tamura coarseness

The coarseness (COA) is related to the distances of notable spatial variations of gray levels. That is, the coarseness is implicitly related to the size of the primitive elements forming the texture of the navigation scene. Since the water surface is easily rippled by wind and wave currents and other meteorological environments, wave texture is one of the elements that cannot be ignored in ship navigation scenes, as shown in Fig. 4 (referring to the classical SMD dataset[13]). The wave texture on the water surface is closely related to the complexity of the navigation scenes, and the intensity of the wave texture can be characterized by the Tamura coarseness. Therefore, when the value of COA is larger, the water surface is not calm, and the complexity of the navigation scene is high.
Figure 4

Coarseness of water surface.

Coarseness of water surface. For an navigation scene I, the calculation of coarseness of the navigation scene starts by selecting a pixel sliding window and obtaining the mean value of the pixel intensity after traversing the sliding window:where controls the size of the sliding window, and K generally takes values between 2 and 6. (x, y) denotes the pixel points within the sliding window. is the gray level at point (m, n). Then, the average intensity difference between windows that do not overlap with each other is calculated by:where denotes the intensity difference in the horizontal direction, and denotes the intensity difference in the vertical direction. For each pixel, assuming that is the value of k that maximizes the larger of and , then the corresponding optimal size is at this time. The average of the optimal sizes is calculated as follows: Therefore, for any USV navigation scene, we extract five parameters separately and combine them into a texture feature vector as follows:

Feature clustering based on the PBBPSC algorithm

To better perceive the complexity of a navigation scene, it is necessary to obtain 4 complexity-level texture feature clustering centers using a paired bare bones particle swarm clustering (PBBPSC) approach, which is modified from the classical and unsupervised particle swarm optimization (PSO) algorithm. The objective function of the feature clustering problem is: denotes the sum distance of the feature particles from the clustering centers.where represents the length of the texture feature vector . In the above equations, is the distance matrix, and denotes the distance between the ith vector data and the jth cluster center. Furthermore, is the belonging matrix, and is expressed as: The purpose of our clustering algorithm is to find that can minimize the objective function. Therefore, a sample and effective PBBPSC approach is used to investigate the best position of the cluster centers. At the beginning of clustering, initial positions for all particles are generated randomly, and then the first positions and personal best values can be obtained. During the iterative process, a paired operator (PO) is applied to cross local minimums. The PO forms particles to search in pairs. In each pair, the particle with a small personal best value is the main particle, and the other particle is the side particle. The candidate position of the main particle is calculated by:where is the personal best position of the main particle, is the global best position of the swarm, is the candidate position of the main particle, and is the Gaussian distribution with mean and standard deviation . Similarly, the candidate position of a side particle is calculated by:where and are the personal best and candidate positions of the side particle, respectively. is the Gaussian distribution with mean and standard deviation . In each iteration, the personal best positions and values and the global best position and value can be updated after all particles have calculated their candidate positions. Once the number of iterations reaches the preset value, the final global best value and the position are the solutions presented by the PBBPSC approach. The pseudocode of PBBPSC is shown in Algorithm  1.

Complexity calculation based on the interval mapping method

The level of complexity of a navigation scene can be obtained by the above clustering method, and we use a sample interval mapping function to calculate the complexity value as follows: Step 1: Compare the Euclidean distance. Calculate the distance d(E, L) between the textural features of scene E and the cluster centers at each level of complexity. Step 2: Determine the levels of complexity. If the distance from navigation scene to the cluster centers of the four classes satisfies the following equation, the level of complexity of this scene is : Step 3: The value of the complexity is distinguished. Map the four levels to the interval [0, 1], i.e., the range of complexity values corresponds to the following equation:when the navigation scene is classified as and if , then Step 4: Calculate the final complexity value. The cluster center of each level of complexity divides that complexity interval into the left part and the right part in turn, and the complexity value corresponding to each part is calculated as follows: Furthermore, can be obtained by the equation:where is the average distance from the scene complexity samples to the respective hierarchical clustering centers, and is the standard variance in the distance.

Experiments and results

Datasets

To enrich the diversity of the testing scenarios, the effectiveness of the proposed method is verified on the classical Singapore marine dataset (SMD), MODD, and our self-collected Yangtze River navigation scene dataset (YRNSD)[36]. The classical SMD contains navigation scenes in various environmental conditions obtained using a Canon 70D camera in waters around Singapore from July 2015 to May 2016[13]. The classical MODD comes from multiple platforms, most of which were small unmanned surface vehicles (USVs), which were manually operated in the port of Koper, Slovenia over a period of months at different times of the day under different weather conditions[33]. Our YRNSD dataset contains a total of 64 videos recorded in the Wuhan section of the Yangtze River and covers many types of obstacles and meteorological conditions. The complexity of the datasets was annotated using manual classification based on whether the frame images contain typical complex elements and the number of their types. The corresponding annotated results are shown in Table 2.
Table 2

Datasets and their levels of complexity (unit: frame).

DatasetsNo complexityLow complexityMedium complexityHigh complexitySize of frames (:pixels)
SMD30856224211321920 × 1080
MODD3783213806687640 × 480
YRNSD5831672277014431920 × 1080
Datasets and their levels of complexity (unit: frame).

Experimental setup

All experiments are implemented on a personal computer with an Intel Core i7-8700K @ 3.70 GHz*12 CPU, an NVIDIA GeForce GTX 1080Ti GPU, 32 GB RAM, and the 64-bit Windows 7 operating system. To ensure testing efficiency and save computational resources, the sizes of the frames in the above datasets are unified to 500 × 280 before clustering and calculation. Then, a 5-dimensional parameter vector can be obtained after textural feature extraction. The above datasets are divided into training and testing sets at rates of 70% and 30%, respectively. Figure 5 shows some randomly selected scenes in the testing set, and their corresponding textural feature parameters are shown in Table 3.
Figure 5

Random navigation scenes from the testing set.

Table 3

Test images with different complexities.

DatasetsASMCONIDMCORCOA
SMDFigure 5a0.05950.98810.17730.97080.7725
Figure 5b0.00780.09910.23020.94570.7345
Figure 5c0.90290.33790.02860.06450.9139
Figure 5d0.93730.30670.02980.03500.9025
Figure 5e0.05710.98770.42100.97270.0262
Figure 5f0.04720.98450.25730.98010.2297
MODDFigure 5g0.06070.93000.54610.90820.2044
Figure 5h0.73800.17810.35430.16960.5658
Figure 5i0.67740.42700.34590.37950.5371
Figure 5j0.92630.32050.01740.04360.9945
Figure 5k0.24790.76030.47760.76670.2792
Figure 5l0.24730.77040.36390.74860.3769
YRNSDFigure 5m0.07930.98600.28640.36180.1563
Figure 5n0.55830.44540.26540.38950.6688
Figure 5o0.50950.46740.56190.48930.4722
Figure 5p0.93790.31770.00420.02910.9703
Figure 5q0.95160.28280.02130.03620.9077
Figure 5r0.06530.98120.29040.96740.2395
Figure 5s0.02630.97210.37720.98690.1495
Figure 5t0.02050.97710.40540.98970.0390
Random navigation scenes from the testing set. Test images with different complexities.

Results comparison and analysis

To verify and compare the effectiveness of our proposed method, we conduct a comparative analysis of the feature clustering and the calculated values of the levels of complexity of the navigation scenes.

Comparison of different clustering methods

The level of complexity of each navigation scene of the training set is manually graded in advance based on its inclusion of typical elements. The manual grading results are used as references, and our proposed PBBPSC approach clustering results are used as the actual results. To evaluate the effectiveness of our proposed approach, it is compared with the K-means algorithm and the classical particle swarm clustering (PSC) algorithm using a series of mathematical parameters, such as the purity, precision, Rand index (RI), and adjusted Rand index (ARI). ARI is an improved version of RI that aims to remove the random effects on the results. These mathematical parameters are calculated as follows: m denotes the total number of testing navigation scenes, and is the number of all members in cluster i. represents the max value of , which refers to the probability that a member of cluster i belongs to class j. Among the components, TP represents the number of navigation scenes where two samples with the same level of complexity are in the same cluster. FP represents the number of navigation scenes where two samples with different levels of complexity are in the same cluster. TN denotes the number of scenes in which two samples with different levels of complexity are in two separate clusters. FN denotes the number of scenes in which two similar levels of complexity are in two separate clusters. The above mathematical parameters all have values in the interval [0, 1], and the larger the value is, the better. Table 4 shows the clustering results of these three methods.
Table 4

Evaluation parameters of the three clustering methods.

MethodsPurityPrecisionRIARIRunning time (unit:s)
K-means0.85950.76740.82540.58511.3380
PSC0.92730.90200.89740.74245.5780
PBBPSC0.97820.96190.95490.88286.2680

Significant values are in bold.

Evaluation parameters of the three clustering methods. Significant values are in bold. The results shown in Table 4 show that the novel proposed PBBPSC clustering method achieves the best performance on the above evaluation parameters in addition to the running time. However, the time complexity of all three methods is O(n) in theory. In addition, to further verify the stability of the PBBPSC clustering method, the above experiment was repeated 30 times, and the stability effect of each clustering method is shown in Fig. 6. The results demonstrate that the PBBPSC clustering method has the best mean and standard deviation over 30 independent runs, which indicates the better stability of our proposed method.
Figure 6

Stability of the three clustering methods.

Stability of the three clustering methods.

Comparison of different complexity calculation methods

The complexity value of a navigation scene can be precisely calculated according to the extracted textural features. First, specific image feature parameters are usually selected as evaluation metrics, and the selected evaluation metrics are normalized and transformed into metrics that are positively correlated with the image complexity. Then, the average weighting method[37] or BP neural network[19] is used to obtain the weight coefficients of each evaluation metric, and the final step multiplies the evaluation metrics with the weights to obtain the complexity value. In this paper, the energy, contrast, inverse difference moment, correlation, and roughness are used as evaluation indexes, and the average weighting method (AWM), BP neural network weighting method (BPNNWM), and our proposed method are used to calculate the complexity. The specific results are shown in Table 5.
Table 5

Test images of different complexity.

Sample dataLevelsReference intervalsAWMBPNNWMOur method
Figure 5a\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_N$$\end{document}LN[0, 0.25)0.59630.26410.2395
Figure 5b\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_L$$\end{document}LL[0.25, 0.5)0.40350.36880.2570
Figure 5c\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_M$$\end{document}LM[0.5, 0.75)0.44960.77270.6262
Figure 5d\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_M$$\end{document}LM[0.5, 0.75)0.44230.66180.6563
Figure 5e\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_H$$\end{document}LH[0.75, 1]0.49280.95150.7919
Figure 5f\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_H$$\end{document}LH[0.75, 1]0.4998 0.95360.9422
Figure 5g\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_N$$\end{document}LN[0, 0.25)0.53040.29570.1345
Figure 5h\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_N$$\end{document}LN[0, 0.25)0.40120.19560.1824
Figure 5i\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_L$$\end{document}LL[0.25, 0.5)0.47340.38880.4416
Figure 5j\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_M$$\end{document}LM[0.5, 0.75)0.46050.76870.7772
Figure 5k\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_H$$\end{document}LH[0.75, 1]0.50650.95980.9943
Figure 5l\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_H$$\end{document}LH[0.75, 1]0.50140.96480.9180
Figure 5m\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_N$$\end{document}LN[0, 0.25)0.37400.08970.1390
Figure 5n\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_L$$\end{document}LL[0.25, 0.5)0.47150.18690.1495
Figure 5o\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_L$$\end{document}LL[0.25, 0.5)0.50000.48410.4297
Figure 5p\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_M$$\end{document}LM[0.5, 0.75)0.45180.67080.6725
Figure 5q\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_M$$\end{document}LM[0.5, 0.75)0.43990.68010.7297
Figure 5r\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_H$$\end{document}LH[0.75, 1]0.50880.96160.8595
Figure 5s\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_H$$\end{document}LH[0.75, 1]0.50240.96630.9014
Figure 5t\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$ L_H$$\end{document}LH[0.75, 1]0.48630.96400.9117
Test images of different complexity. The values in bold in Table 5 indicate that the complexity values are within the reference intervals, which means that the scene is validly perceived. The complexity values obtained by the AWM are prone to mismatches between the calculated value and the actual level. For example, the level of complexity of the navigation scene in Fig. 5a is , but its complexity value is greater than that of the reference interval with a high level of complexity, and this result is obviously improper. The BPNNWM method largely improves the shortcomings of the AWM, but it easily falls into the local optimal solution, and each training initial value and process has a slight difference that may lead to a large difference in the complexity value of the same navigation scene; that is, there is randomness in the calculation results. Table 5 shows that the complexity values calculated by our proposed method can match the reference intervals well. First, this is because the texture parameter features are relatively fixed. Second, the accuracy of the proposed clustering results can be maintained at a high level. Therefore, the complexity values computed by interval mapping can possess good stability during the complexity perception process.

Discussion

During the USV tests, we concluded that the actual distribution of the four levels of complexity scenarios deviated from the distribution of the actual dataset used in this paper. Both the classical and self-harvested datasets in our experiments focus on scenarios with high complexity and smaller proportions of no-complexity scenes. Taking the self-collected YRNSD as an example, the Wuhan section of the Yangtze River basin is typical inland river water, and the characteristics of the collected frames are characterized by busy traffic, complex backgrounds on both sides of the river, and narrow channels. Therefore, the proportion of no-complexity scenes is very small in the data samples. To balance the data samples, we performed the data augmentation operations of rotation, translation, scaling, and mirroring for the no-complexity navigation scenes in the datasets to expand this part of the data sample by a factor of 5. However, the complexity perception results of the navigation scenes were not significantly improved by reclustering and computing the expanded datasets. The reason for this phenomenon may be due to the GLCM and Tamura coarseness methods employed in the texture extraction process, which are insensitive to rotation, translation, scaling, and mirroring operations. This is because the above operations only linearly transform the image and do not substantially change the distribution of textures in the navigation scenes. This phenomenon further demonstrates that our proposed textural feature parameters can be used to perceive and distinguish navigation scenes with different levels of complexity.

Conclusions

This paper constructs a USV navigation scene complexity perception system based on image texture features. Here, the level of complexity classification and accurate perception calculation method are obtained by modeling and analyzing the navigation scene complexity problem. The first contribution of this paper is that it sorts out the multiple elements contained in complex scenes to cover as many test scenes as possible, solves the long-tail problem, and summarizes a total of five typical elements of cluttered background, dynamic obstacles, camera shaking, cast shadows, and special weather conditions. Then, four levels of complexity are divided according to whether the current navigation scene contains the above typical complex scene elements and their quantity. These levels are no complexity , low complexity , medium complexity and high complexity . The other main contribution is that the work proposes a scene complexity perception method based on image texture features. The method first uses the GLCM and Tamura coarseness to obtain a 5-dimensional texture feature vector. Then, a novel PBBPSC algorithm is designed to cluster the scene complexity, and based on the classification results, the interval mapping method is applied to obtain the exact value of the navigation scene complexity. Finally, it is proven that the method in this paper can not only describe the level of complexity of navigation scenes but also obtain accurate complexity values, which are verified using classical and self-collected datasets. Next, based on the obtained navigation scene complexity, we will further study and improve the navigation scene testing system to provide support for the development of USV testing technology.
  5 in total

1.  Fast Image-Based Obstacle Detection From Unmanned Surface Vehicles.

Authors:  Matej Kristan; Vildana Sulíc Kenk; Stanislav Kovacic; Janez Pers
Journal:  IEEE Trans Cybern       Date:  2015-03-31       Impact factor: 11.448

2.  NIMA: Neural Image Assessment.

Authors:  Hossein Talebi; Peyman Milanfar
Journal:  IEEE Trans Image Process       Date:  2018-04-30       Impact factor: 10.856

3.  The Effects of GLCM parameters on LAI estimation using texture values from Quickbird Satellite Imagery.

Authors:  Jingjing Zhou; Rui Yan Guo; Mengtian Sun; Tajiguli Tu Di; Shan Wang; Jiangyuan Zhai; Zhong Zhao
Journal:  Sci Rep       Date:  2017-08-04       Impact factor: 4.379

4.  Autonomous Visual Perception for Unmanned Surface Vehicle Navigation in an Unknown Environment.

Authors:  Wenqiang Zhan; Changshi Xiao; Yuanqiao Wen; Chunhui Zhou; Haiwen Yuan; Supu Xiu; Yimeng Zhang; Xiong Zou; Xin Liu; Qiliang Li
Journal:  Sensors (Basel)       Date:  2019-05-14       Impact factor: 3.576

5.  Underactuated USV path following mechanism based on the cascade method.

Authors:  Mingzhen Lin; Zhiqiang Zhang; Yandong Pang; Hongsheng Lin; Qing Ji
Journal:  Sci Rep       Date:  2022-01-27       Impact factor: 4.379

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.