This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Aerial Images Meet Crowdsourced Trajectories: A New Approach to Robust Road Extraction

Lingbo Liu, Zewei Yang, Guanbin Li, Kuo Wang, Tianshui Chen and Liang Lin, Senior Member, IEEE This work was supported in part by the Guangdong Basic and Applied Basic Research Foundation under Grant No.2020B1515020048, in part by the National Natural Science Foundation of China under Grant No.61976250, No.U1811463 and No.61836012, and in part by the Guangzhou Science and Technology Project under Grant No.202102020633. (Corresponding Author: Liang Lin.)L. Liu, G. Li, K. Wang and L. Lin are with the School of Computer Science and Engineering, Sun Yat-Sen University, China, 510000 (e-mail: liulingb@mail2.sysu.edu.cn; liguanbin@mail.sysu.edu.cn; wangk229@mail2.sysu.edu.cn; linliang@ieee.org).Z. Yang is with the School of Mathematics, Sun Yat-Sen University, China, 510000 (e-mail: yangzw7@mail2.sysu.edu.cn).T. Chen is with The Guangdong University of Technology, Guangzhou, China, 510000 (e-mail: tianshuichen@gmail.com).
Abstract

Land remote sensing analysis is a crucial research in earth science. In this work, we focus on a challenging task of land analysis, i.e., automatic extraction of traffic roads from remote sensing data, which has widespread applications in urban development and expansion estimation. Nevertheless, conventional methods either only utilized the limited information of aerial images, or simply fused multimodal information (e.g., vehicle trajectories), thus cannot well recognize unconstrained roads. To facilitate this problem, we introduce a novel neural network framework termed Cross-Modal Message Propagation Network (CMMPNet), which fully benefits the complementary different modal data (i.e., aerial images and crowdsourced trajectories). Specifically, CMMPNet is composed of two deep Auto-Encoders for modality-specific representation learning and a tailor-designed Dual Enhancement Module for cross-modal representation refinement. In particular, the complementary information of each modality is comprehensively extracted and dynamically propagated to enhance the representation of another modality. Extensive experiments on three real-world benchmarks demonstrate the effectiveness of our CMMPNet for robust road extraction benefiting from blending different modal data, either using image and trajectory data or image and Lidar data. From the experimental results, we observe that the proposed approach outperforms current state-of-the-art methods by large margins. Our source code is resealed on the project page http://lingboliu.com/multimodal_road_extraction.html.

Index Terms:
Land remote sensing, Road network extraction, Aerial images, Crowdsourced trajectories.

I Introduction

Earth science [1, 2] is a complex and huge subject that has been researched for decades or even centuries. As a subbranch of geoscience, geoinformatics [3] recently has received increasing interests with the rapid development of satellite and computer technologies. Accurately obtaining land surface information (e.g., trees, lakes, buildings, roads, and so on) from remote sensing data can help us to better understand our earth. Among these objects, traffic roads are very difficult to recognize, since they are threadlike and unimpressive in aerial images. To promote land analysis, in this work we aim to recognize traffic roads automatically from remote sensing data. Such a geoinformatics task not only facilitates a series of practical applications [4, 5, 6] for urban development, but also helps to estimate the urban expansion trend to analyze potential impacts of human activities on earth lands.

In literature, numerous algorithms have been proposed to extract traffic roads from aerial images. Most early works [7, 8, 9] extracted handcrafted features (e.g., texture and contour) and applied shallow models (e.g., Support Vector Machine [10] and Markov Random Field [11]) to recognize road regions. Recently, deep convolutional networks have become the mainstream in this field and achieved remarkable progresses [12, 13, 14] due to their great capacities of representation learning. However, aerial image-based traffic road extraction remains a very challenging problem, especially in the face of the following circumstances. First, some roads are extremely occluded by trees, as shown in Fig. 1-(a). Relying solely on visual information, these roads are hard to be detected from aerial images. Second, some infrastructures (e.g., train tracks, building tops, and river walls) have similar appearances of traffic roads, as shown in Fig. 1-(b). Without extra information, it is hard to distinguish roads from these structures, which may result in false negatives and false positives. Third, in some bad meteorological conditions (e.g., thick fog/haze), it’s very difficult to recognize traffic roads due to poor visibility, as shown in Fig. 1-(c). Nevertheless, road maps have low tolerance for errors, since incorrect routes would seriously affect the transportation’s operation efficiency. Therefore, some robust methods are desired to accurately extract traffic roads.

Refer to caption

Figure 1: (a) Traffic roads are usually occluded by trees. Although crowdsourced trajectories can help discover roads, excessive noises are also introduced. (b) Train tracks and traffic roads have similar appearances, thus it is hard to distinguish them only using visual cues. When only using trajectories, some parking lots are easily mistaken for roads. (c) It’s difficult to directly recognize roads from aerial images when the studied city has poor visibility in fog/haze weather. (d) Only using local information, we may fail to recognize some road regions that are heavily occluded and have very few trajectories, as shown in the yellow box.

Fortunately, we observe that some data for non-visual modalities, such as vehicle trajectories, can also help discover traffic roads. Intuitively, a region with a large number of trajectories is likely to be a road segment [15, 16, 17]. In recent years, vehicle ownership has grown dramatically and most vehicles have been equipped with GPS devices, which greatly increases the availability of large-scale trajectory datasets and boosts the feasibility of trajectory-based road extraction. Despite substantial progress [18, 19], this research direction still suffers from many challenges. First, crowdsourced trajectories have excessive noises (e.g., positioning drift) caused by the uneven quality of GPS devices, as shown in Fig. 1-(a). Although various preprocessing techniques (e.g., clustering and K-nearest neighbors) were used [20, 21, 22], the noise problem has not been well solved. Second, some non-road areas, such as the parking lot in Fig. 1-(b), also have lots of trajectories and they are easily mistaken for roads without auxiliary information. Most conventional works [23, 24, 25, 26] have not explicitly distinguished these areas. Third, previous trajectory-based methods mainly extracted the topologies of road networks. Because of mass erratic trajectories, it is difficult to obtain the accurate width of roads, which can be easily computed in high-resolution aerial images.

In general, image-based methods and trajectory-based methods have individual strengths and weaknesses. It is very natural to incorporate aerial images and crowdsourced trajectories to extract traffic roads robustly. However, there are very limited works [27, 28] that simultaneously utilized the two modalities mentioned above. Moreover, these works directly fed the concatenation of aerial images and rendered maps of trajectories or their features into convolutional neural networks, which is a suboptimal strategy for multimodal fusion. Recently, Wu et al. [29] designed a gated fusion module to fuse multimodal features, but not refine features mutually, thus the complementarities of images and trajectories have not been fully exploited. Furthermore, all above-mentioned methods performed road extraction only with local features/information, thus may fail to recognize some road regions that are heavily occluded and meanwhile have very few trajectories, such as the yellow box in Fig. 1-(d). When considering all information of the whole image and trajectories, we can correctly infer that this region is a road segment. Therefore, both the local and global information should be explored for traffic road extraction.

To facilitate road extraction, we propose a novel framework termed Cross-Modal Message Propagation Network (CMMPNet), which fully explores the complementarities between aerial images and vehicle crowdsourced trajectories. Specifically, our CMMPNet is composed of: (i) two deep AutoEncoders for modality-specific feature learning, in which one takes an aerial image as input and the other one uses the rendered trajectory heat-map, and (ii) a Dual Enhancement Module (DEM) that refines the features of different modalities mutually with a message passing mechanism. In particular, our DEM propagates both the local detail information and global structural information dynamically with two progress propagators. First, a Non-Local Message Propagator extracts the local and global messages embedded in the features of each modality, which are utilized to refine the features of another modality. Thereby, image features and trajectories features can be enhanced mutually. Moreover, the limitation of local information is also well eliminated. Second, a Gated Message Propagator employs gate functions to dynamically determine the final propagated messages, so that the beneficial messages are transmitted and the interferential messages (e.g., visual cues of train tracks and the noises of trajectories) are abandoned. For further improving the robustness, our DEM is integrated into different layers of CMMPNet to enhance the image features and trajectory features hierarchically. Finally, the last outputs of two AutoEncoders are concatenated to accurately predict the high-resolution traffic road maps.

The proposed CMMPNet has three appealing properties. First, through refining modality-specific features mutually, our method can better explore the complementarities of aerial images and crowdsourced trajectories, compared with previous works that directly taken their concatenation as input or simply fused their features. Second, thanks to the tailor-designed DEM, our method is more robust to extract traffic roads. With the aid of visual information, some useless and noisy trajectories can be effectively eliminated, while occluded roads are easily discovered with the trajectory information and some delusive non-road regions are also well distinguished. Third, it is worth noting that our method is very general for robust road extraction by utilizing multimodal information. Furthermore, CMMPNet can also be generalized to combine image and lidar data for road extraction. Extensive comparisons on three real-world benchmarks two for image and trajectory data and the other for image and lidar data) demonstrate the advantage of our proposed method. In summary, this paper makes the following contributions:

  • It proposes a novel Cross-Modal Message Propagation Network for land remote sensing analysis, which extracts traffic roads robustly by explicitly capturing the complementarities among different modal data.

  • It introduces a Dual Refinement Module for multimodal representation learning, where the complementary information of each modality is dynamically propagated to effectively enhance other modal features based on the message passing mechanism.

  • It presents sufficient experiments and comparisons on three multimodal benchmarks for showing the superiority and generalization of our approach against existing state-of-the-art methods.

The rest of this paper is organized as follows. First, we review some related works of earth science research and traffic road extraction in Section II. We then provide some preliminaries in Section III and introduce the proposed CMMPNet in Section IV. Extensive evaluations and generalization analysis are conducted in Section V and in Section VI. Finally, we conclude this paper and discuss future works in Section VII.

II Related Works

II-A Earth Science Research

Earth science [1, 2] is a crucial subject that studies the physical, chemical, and biological characterizations of our earth for better understanding various physical phenomena and natural systems. Earth science is also a complex subject and it contains a lot of research branches [30]. For instance, meteorologists [31] study the atmosphere for dangerous storm warnings and hydrologists [32] examine hydrosphere for flood warnings. Seismologists [33] study earthquakes and forecast where they will strike, while geologists [34] study rocks and help to locate useful minerals. Among all the subbranches of geoscience, geoinformatics [3] recently has attracted widespread interests with the rapid development of satellite and computer technologies, since it can greatly facilitate other research branches, e.g., monitoring storm/flood from remote sensing data and forecasting their evolutionary trend. In this work, we inherit the research content of geoinformatics and apply computer technologies to land remote sensing analysis, e.g, extracting the traffic road network from aerial images and some complementary modalities. This problem has important applications in transportation navigation and public management. Moreover, we can also compare the road networks at different times and estimate the urban expansion tendency, thereby analyzing the potential impacts of human activities on earth lands.

II-B Traffic Road Extraction

As a crucial foundation in intelligent transportation systems, automatic road extraction has been studied for decades [35]. On the basis of the modality of input data, previous approaches can be divided into four categories and we would investigate the related works of each category.

II-B1 Aerial Image-based Road Extraction

In industrial communities, a large number of high-quality aerial images can be accessed easily, with the rapid development of remote sensing imaging technologies equipped in artificial satellites [36, 37]. Numerous methods were proposed to extract traffic roads from these aerial images. Early works [38, 39, 40, 41] usually fed hand-crafted features (e.g., texture and contour) into shallow models (e.g., deformable model and Markov Random Field) to recognize road regions. However, most of them only worked in constraint scenarios. In recent years, due to the great capacity for representation learning, deep neural networks [42] have become the mainstream in this field. For instance, Cheng et al. [12] proposed a cascaded end-to-end convolutional neural network to cope with the road detection and centerline extraction simultaneously with two cascaded CNN. Zhang et al. [43] developed a semantic segmentation neural network, which combined the residual learning and U-Net to extract road areas. Zhou et al. [44] utilized dilation convolutions to enlarge the receptive field of Linknet [45] and then employed this enhanced model to extract road regions from high-resolution aerial images. Fu et al. [46] predicted the category of each pixel with a multi-scale Fully Convolutional Network and refined the output density map with a Conditional Random Fields post-processing. Despite substantial progress, they may still fail in complex scenarios, especially in the face of extreme occlusions. As analyzed above, it’s very difficult to perfectly extract traffic roads only with the visual information of aerial images. Therefore, more complementary information should be delved from other modalities for facilitating road extraction.

Refer to caption
Figure 2: Illustration of trajectory heat-map generation. Given an aerial image, we first query all trajectory samples in the corresponding geographical region and then generate a 2D trajectory heat-map by counting the number of samples projected at every pixel. Finally, a logarithm-based normalized function and a Gaussian kernel filter are applied to refine the trajectory heat-map.

II-B2 Trajectory-based Road Extraction

Intuitively, a geographical region with mass vehicle trajectories is likely to be a road area. Based on this observation, some researchers have attempted to recognize traffic roads from crowdsourced trajectories. Since trajectory data has excessive noise, most previous works focused on how to eliminate the GPS noises and uncertainties. Conventional methods can be divided into three categories. The first category is clustering-based models [47, 24, 48]. In these works, the task of road extraction is formulated as a network alignment optimization problem where both the nodes and edges of road networks have to be inferred. Specifically, nodes or short edges are first identified from raw GPS points with spatial clustering algorithms and then connected to form the final road networks. The second category is trace-merging based models [49, 50], which either merge each trajectory to an existing road segment or generate a new segment if no existing segment is matching. The third category is KDE-based methods [19, 51], which first apply Kernel Density Estimation [52] to convert trajectories into a density map for reducing the influences of noise, and then employs image processing techniques to extract road centerlines. Recently, deep neural networks have also been applied to this task. For instance, Cheng et al. [26] proposed a deep learning-based map generation framework, which extracts features from trajectories in both spatial view and transition view to infer road centerlines. Although various techniques are used, GPS noises still can’t be well eliminated and the extracted road networks are far from satisfactory, due to the limited information of crowdsourced trajectories.

II-B3 Lidar-based Road Extraction

Compared with aerial images, Light Detection And Ranging (Lidar) data have two specialties. First, Lidar data contains depth or distance information. Second, different objects (e.g., buildings, trees, and roads) have different reflectivity to the laser. Because of these specialties, roads are mostly defined by flatness in the aerial viewpoint, which can help to distinguish the road proposals from buildings and trees. In literature, there also exist some algorithms [53, 54, 55] that identified traffic road from Lidar data. For instance, Hu et al. [56] first filtered the non-ground LiDAR points and then detected road centerlines from the remaining ground points. After obtaining the ground intensity images, Zhao and You [56] designed structure templates to search for roads and determined road widths and orientations with a subsequent voting scheme. Despite some progress, Lidar-based road extraction remains a very challenging problem and existing methods perform poorly in complex scenarios, suffering from the sparsity of Lidar data and the noise points [57].

II-B4 Multi-modal Road Extraction

As analyzed above, each modality has individual benefits and drawbacks, so it’s wise to aggregate their complementary information for extracting traffic roads effectively. In literature, numerous methods have been proposed to identify road areas using both aerial images and Lidar data, because of the accessibility of these data. For instance, Hu et al. [58] first segmented the primitives of roads from both optical images and Lidar data, and then detected road stripes with an iterative Hough transform algorithm to form the final road network by topology analysis. Parajuli et al. [59] developed a modular deep convolution network called TriSeg, in which two SegNet [60] were used to extract features respectively from aerial images and Lidar data, and another SegNet fused modular features to estimate the final road maps. However, neither of aerial images and Lidar data can provide sufficient information to discover the traffic roads heavily occluded by trees, thus some recent works incorporated aerial images and vehicle trajectories to identify road areas. For instance, Sun et al. [27] fed the concatenation of rendered trajectory heat-maps and aerial images into different backbone networks (e.g., UNet [61], Res-UNet [43], LinkNet [45] and D-LinkNet [44]) to estimate those traffic roads. In [29], trajectory maps and aerial images were first fed into different networks respectively for feature extraction, and then the modular features at different layers were fused to predict the final roads. Despite progress, such a concatenation or fusion manner can not fully exploit the complementarities of different modalities, and more effective methods are desired for multimodal road extraction.

II-C Message Passing Mechanism

In the field of machine learning, message passing [62, 63] refers to information interactions between different entities. A large number of works have shown that such a mechanism can effectively facilitate deep representation learning. For instance, Wang et al. [64] introduced an inter-view message passing module to enhance the view-specific features for action recognition, while Liu et al. [65] propagated information among multiscale features to model the scale variations of people. In graph convolution networks, the message passing mechanism is usually embedded to aggregate information from neighboring nodes [66, 67, 68, 69, 70]. Recently, this mechanism has also been adopted for cross-modal representation learning. For instance, Wang et al. [71] addressed the text-image retrieval problem by transferring multimodal features and computing their matching scores. Nevertheless, most of these previous methods propagated information in a local manner (e.g., at short range). Without capturing global information, these methods may fail to discover the occluded roads that meanwhile have very few trajectories, as shown in Fig. 1-(d). Therefore, more effective approaches are desired to fully exploit the complementary information of aerial images and crowdsourced trajectories for traffic road extraction.

Refer to caption
Figure 3: The architecture of the proposed Cross-Modal Message Propagation Network (CMMPNet) for multimodal road extraction. Specifically, our CMMPNet is composed of (i) two deep AutoEncoders that take an aerial image and a trajectory heat-map respectively to learn modality-specific features, and (ii) a Dual Enhancement Module (DEM) that dynamically propagates the non-local messages (NLM, i.e, local one and global one) of every modality with gated functions to enhance the representation of another modality. The final features of the image and trajectory heat-map are concatenated to generate a traffic road map.

III Preliminaries

In this section, we first introduce how to generate trajectory heat-maps from raw GPS data and then formally define the problem of image+trajectory based road extraction.

Raw Trajectory Samples: With the rapid growth of vehicle ownership, we can easily collect a mass of vehicles’ GPS trajectories to construct a large-scale trajectory database [72, 73, 74]. In this database, each trajectory sample can be represented as a tuple {vid,lon,lat,t,sp,si}\{{vid},{lon},{lat},{t},{sp},{si}\}, where vid{vid} is the ID of a vehicle, lonlon and latlat are the longitude and latitude at timestamp t{t}. Term spsp denotes the vehicle’s speed. sisi is the trajectory sampling interval and different vehicles have different sampling settings. We notice that some early works [18] manually generated some virtual samples on the line segment between two consecutive real samples to augment the trajectory quantity. Nevertheless, this would cause a lot of noise in complex scenarios, since the real-world vehicles may have large sampling intervals (such as si{si} is mainly set to 10, 60, 180, and 300 seconds in Beijing [27]) and it is difficult to accurately infer the virtual trajectories under these settings. Thanks to the crowdsourcing mechanism, adequate trajectories can be easily collected nowadays. Therefore, we only use the real trajectory samples in this work.

Trajectory Heat-map Generation: For deep neural networks, matrix or tensor is one of the most common formats of input. Thus we need to transform the raw GPS data into 2D trajectory heat-maps before feeding them into networks. The whole transformation process is shown in Fig. 2. Specifically, give an aerial image with a resolution H×WH{\times}W, we first search out all trajectory samples in its coordinate range [lonl,latl][{lon}_{l},{lat}_{l}]*[lonu,latu][{lon}_{u},{lat}_{u}], where the subscripts ll and uu denote the lower and upper bounds, respectively. These samples are then projected into a H×WH{\times}W greyscale map by counting the number of samples projected at every pixel. In this map, the pixels of road areas usually have high values, while the pixel values in non-road regions are very small, even zero. This would facilitate the discovery of traffic roads. However, we find that such a projected map has two minor defects. First, some infrequently-traveled roads are so pale that they are hard to be recognized. Second, this map is too coarse and sharp. For example, two adjacent pixels in road areas may have values of different scales, or even a road pixel does not match any projected samples. Inspired by Kernel Density Estimation frequently used for trajectory processing [18], we normalize the projected map with a logarithm function and apply a 3×33{\times}3 Gaussian kernel filter for smoothing. The involved Gaussian filter can also eliminate trajectory noises to a certain degree [75]. In this way, the final trajectory heat-map becomes smooth and traffic roads are more distinct from backgrounds.

Image+Trajectory based Road Extraction: Given a H×WH{\times}W aerial image II and the corresponding trajectory heat-map TT, our goal is to automatically predict a H×WH{\times}W binary road map

M=({I,T},θ),M=\mathcal{F}(\{I,T\},\theta), (1)

where ()\mathcal{F}(\cdot) is a mapping function with learnable parameters θ\theta. Specifically, the pixels within road areas are supposed to have high response value (i.e.. 1), while the response values of background pixels should be 0.

IV Methodology

IV-A Framework Overview

As mentioned above, aerial images and vehicle crowdsourced trajectories are complementary for traffic road extraction. To recognize unconstrained roads effectively and robustly, we propose a Cross-Modal Message Propagation Network (CMMPNet), which mutually enhances the hierarchical features of different modalities for better capturing their complementary information. As shown in Fig. 3, our CMMPNet is composed of (i) two deep AutoEncoders for modality-specific feature learning and (ii) a Dual Enhancement Module (DEM) for cross-modal feature refinement. In this subsection, we mainly introduce the architecture of CMMPNet, whose specific components are described in the following subsections.

Specifically, given an aerial imagery II and a trajectory heat-map TT with a resolution H×W{H{\times}W}, we first explicitly learn modality-specific representations by feeding them into different AutoEncoders, each of which consists of four encoding blocks and four decoding blocks. As shown in Fig. 3, the first AutoEncoder takes II as input and extracts a group of image features:

fI={fI1,fI2,fI3,fI4,fI5,fI6,fI7,fI8},f_{I}=\{f_{I}^{1},f_{I}^{2},f_{I}^{3},f_{I}^{4},f_{I}^{5},f_{I}^{6},f_{I}^{7},f_{I}^{8}\}, (2)

where the first four features are the outputs of encoding blocks and the remaining four features are the output of decoding blocks. With the same architecture, the second AutoEncoder extracts a group of trajectory features:

fT={fT1,fT2,fT3,fT4,fT5,fT6,fT7,fT8},f_{T}=\{f_{T}^{1},f_{T}^{2},f_{T}^{3},f_{T}^{4},f_{T}^{5},f_{T}^{6},f_{T}^{7},f_{T}^{8}\}, (3)

from the input trajectory heat-map TT.

Rather than directly fuse image and trajectory features with concatenation [28] or weighted addition [29], we fully capture the multimodal complementary information through enhancing their features mutually with a message passing mechanism. For each pair of multimodal feature {fIi,fTi}\{f_{I}^{i},f_{T}^{i}\}, we employ the proposed DEM to generate two enhanced features {f^Ii,f^Ti}\{\widehat{f}_{I}^{i},\widehat{f}_{T}^{i}\} with their complementary information. This process can be formulated as:

f^Ii,f^Ti=DEM(fIi,fTi),i=1,2,,8.\widehat{f}_{I}^{i},\widehat{f}_{T}^{i}=\text{DEM}(f_{I}^{i},f_{T}^{i}),{~{}~{}~{}}i=1,2,...,8. (4)

These enhanced features are then fed into the next block of individual AutoEncoder respectively for further representation learning. For convenience, the final outputs of image AutoEncoder and trajectory AutoEncoder are denoted as fIof_{I}^{o} and fTof_{T}^{o}, and they have the same resolution H×W{H{\times}W}. Here, fIof_{I}^{o} and fTof_{T}^{o} are jointly utilized to predict a probability map MRH×WM\in R^{H{\times}W} for traffic roads with the following formulation:

M=Conv(fIofTo,𝕎11),M=Conv(f_{I}^{o}{\oplus}f_{T}^{o},\mathbb{W}_{1*1}), (5)

where \oplus denotes feature concatenation and 𝕎11\mathbb{W}_{1*1} refers to the parameters of a 1*1 convolutional layer. For each position (x,y)(x,y), it can be regarded as a road region only when M(x,y)M(x,y) is greater than a given threshold.

It’s worth noting that our method is universal for multi-modal road extraction. Except for image+trajectory data, the proposed CMMPNet can also be directly employed to recognize traffic roads with image+lidar data. The university of our method would be verified in Section V and VI.

IV-B Modality-Specific Feature Learning

In the previous work [27], aerial images and trajectory heat-maps were directly concatenated to feed into the same network, which caused that their features were over mixed and their complementarities were missed to some extent. To address this problem, we feed the given aerial image and the corresponding trajectory heat-map into different networks to learn modality-specific features. Optimized with individual parameters, these features well preserve the specific information of each modality, thus can be further utilized for mutual refinement.

TABLE I: The configuration of our AutoEncoder. In the first convolutional layer, the input channel CiC_{i} is set to 3 for aerial images and 1 for trajectory heat-maps, and the stride is set to 2. In each block, DR denotes the downsampling ratio of resolution and CoC_{o} is the channel number of output. MP denotes a 2×22\times 2 max-pooling layer. Res, Up and Inter refer to the residual unit, upsampling unit and interim unit described in Fig. 4.
Block Configuration Output
Sign DR CoC_{o}
- Conv(7,CiC_{i},64,s=2) 1/2 64
encoding-1 MP\rightarrow3*Res(64, 64) f1f^{1} 1/4 64
encoding-2 MP\rightarrowConv(  64,128)\rightarrow3*Res(128,128) f2f^{2} 1/8 128
encoding-3 MP\rightarrowConv(128,256)\rightarrow5*Res(256,256) f3f^{3} 1/16 256
encoding-4 MP\rightarrowConv(256,512)\rightarrow2*Res(512,512) f4f^{4} 1/32 512
- Inter(512,512) 1/32 512
decoding 1 Up(512,256) + f3f_{3} f5f^{5} 1/16 256
decoding 2 Up(256,128) + f2f_{2} f6f^{6} 1/8 128
decoding 3 Up(128,  64) + f1f_{1} f7f^{7} 1/4 64
decoding 4 Up( 64,  64) f8f^{8} 1/2 64
- TConv(4,64,32,2)\rightarrowConv(3,32,32) fof^{o} 1 32

To maintain the high resolution of final outputs, two AutoEncoders are adopted intentionally to extract modality-specific features. Notice that various AutoEncoders (e.g, Res-UNet [43], LinkNet [45], and D-LinkNet [44]) are suitable to serve as the backbone network of our framework. Since these networks have similar architectures, we take D-LinkNet based AutoEncoder as an example to demonstrate the details of modality-specific feature learning. As shown in Table I, both the image AutoEncoder and trajectory AutoEncoder are mainly composed of four encoding blocks and four decoding blocks. Specifically, we first use a convolutional layer to extract initial features and then feed them into the following four encoding blocks, each of which consists of a 2×22\times 2 max-pooling layer and multiple residual units. As shown in Fig. 4-(a), each residual unit contains two 3×33\times 3 convolutional layers and a skip layer. After the encoding stage, an interim unit is adopted to capture more spatial context by expanding the receptive field with four dilated convolutional layers. At the decoding stage, four decoding blocks are utilized to enlarge the resolutions of features progressively. Specifically, each decoding block is developed as an upsampling unit, which consists of two convolutional layers for channel adjustment and a transposed convolutional layer for feature upsampling, as shown in Fig. 4-(c). To simultaneously exploit the lower-level information and high-level information, we incorporate the features of encoding blocks and decoding blocks with element-wise addition. Finally, we fully restore the resolution of feature to H×WH{\times}W with a transposed convolution and apply a 3×33\times 3 convolutional layer to generate the final modality-specific feature foRH×W×32f^{o}\in R^{H{\times}W{\times}32}. Note that our image AutoEncoder and trajectory AutoEncoder have individual parameters, thus they can effectively capture and preserve the specific information of each modality.

Refer to caption

Figure 4: The architecture of Residual Unit, Upsampling Unit, and Interim Unit. Conv(k,n,m)Conv(k,n,m) denotes a k×kk{\times}k standard convolution, whose input channel is nn and output channel is mm. DConv(k,n,m,r)DConv(k,n,m,r) refers to a dilated convolution with a dilated ratio rr and TConv(k,n,m,s)TConv(k,n,m,s) is a transposed convolution with a stride ss.

IV-C Cross-modal Feature Refinement

After modality-specific feature learning, we refine these features mutually with a Dual Enhancement Module (DEM) based on the message passing mechanism. In this module, a Non-Local Message Propagator and a Gated Message Propagator are integrated to dynamically transmit the local and global message of each unimodal feature to complement the feature of another modality. Absorbing the complementary information of other modalities, each unimodal feature becomes more reasonable and robust. In this subsection, we take the refinement of features fIif^{i}_{I} and fTiRh×w×cf^{i}_{T}\in R^{h{\times}w{\times}c} as an example to demonstrate the working mechanism of the tailor-designed DEM. Note that hh, ww, and cc are the height, width, and channel number of these features.

Refer to caption

Figure 5: The architecture of Dual Enhancement Module. This figure mainly illustrates how to enhance the image feature fIif^{i}_{I} with the information extracted from the trajectory feature fTif^{i}_{T}. The cross-modal information from fTif^{i}_{T} to fIif^{i}_{I} is obtained by dynamically fusing the local information LTiL^{i}_{T} and global information GTiG^{i}_{T} with the learnable fused weights θTLi{\theta}^{i}_{TL} and θTGi{\theta}^{i}_{TG}. This architecture can also be employed to enhance fTif^{i}_{T}. SPP and FC are the abbreviations of Spatial Pyramid Pooling [76] and Fully-Connected layer, respectively. + and \otimes denote the element-wise addition and multiplication, and \oplus refers to feature concatenation.

IV-C1 Non-Local Message Propagator

Unlike previous works [77, 65] that only used local cues, our method explores both the local and global information for feature enhancement. Here we mainly introduce how to utilize the information of trajectory feature fTif^{i}_{T} to enhance the image feature fIif^{i}_{I}. The refinement of fTif^{i}_{T} is performed with the same process.

As shown in Fig. 5, we first extract a local information map LTiRh×w×cL^{i}_{T}\in R^{h{\times}w{\times}c} by feeding fTif^{i}_{T} into a 3×33\times 3 convolutional layer. Then we aggregate the local information at different locations to generate a global information map. Rather than use the compute-intensive non-local module proposed in [78], we employ a lightweight NN-level Spatial Pyramid Pooling (SPP [76]) and a Fully-Connected (FC) layer for global information generation. Specifically, at the ii-th level (ii=1,2…,NN), LTiL^{i}_{T} is divided into 2i1×2i12^{i-1}\times 2^{i-1} regions, each of which has a dimension of h2i1×w2i1×c\frac{h}{2^{i-1}}{\times}\frac{w}{2^{i-1}}{\times}c and is fed into a h2i1×w2i1\frac{h}{2^{i-1}}{\times}\frac{w}{2^{i-1}} max-pooling layer to obtain a 1×1×c1{\times}1{\times}c information vector. Further, the information vectors at all levels are concatenated and fed into the FC layer with cc output neurons to generate a global information vector. This global vector is copied h×wh{\times}w times and reshaped to form the global information map GTiRh×w×cG^{i}_{T}\in R^{h{\times}w{\times}c}. After obtaining the local map LTiL^{i}_{T} and global map GTiG^{i}_{T}, we can easily propagate them to refine fIif_{I}^{i} with the following formulation:

f^Ii=fIi+LTi+GTi,\widehat{f}^{i}_{I}=f^{i}_{I}+L^{i}_{T}+G^{i}_{T}, (6)

where f^Ii\widehat{f}^{i}_{I} is the enhanced image feature and the operator “+” denotes a element-wise addition. In the same way, we can compute the enhanced trajectory feature f^Ti\widehat{f}^{i}_{T} as follow:

f^Ti=fTi+LIi+GIi,\widehat{f}^{i}_{T}=f^{i}_{T}+L^{i}_{I}+G^{i}_{I}, (7)

where LIiL^{i}_{I} and GIiG^{i}_{I} are the extracted local information map and global information map of image feature fIif^{i}_{I}.

IV-C2 Gated Message Propagator

In the previous propagator, the local and global information is transmitted statically, which isn’t optimal for cross-modal refinement and even has disturbing effects at some locations. To alleviate this issue, a Gated Message Propagator is introduced to adaptively determine and propagate the complementary information. With multiple learnable gate functions, the beneficial information is transmitted and the disturbing information (e.g., visual cues of train tracks and the noises of trajectories) is suppressed.

Specifically, we first introduce the computation of the gated weights of different information. As shown in Fig. 5, the trajectory information LTiL^{i}_{T} and GTiG^{i}_{T} is concatenated and fed into two 1×11\times 1 convolutional layers:

θTLi=Sigm(Conv(LTiGTi,𝕎TLi)),θTGi=Sigm(Conv(LTiGTi,𝕎TGi)),\begin{split}{\theta}^{i}_{TL}&=\textit{Sigm}(\textit{Conv}(L^{i}_{T}\oplus G^{i}_{T},\mathbb{W}^{i}_{TL})),\\ {\theta}^{i}_{TG}&=\textit{Sigm}(\textit{Conv}(L^{i}_{T}\oplus G^{i}_{T},\mathbb{W}^{i}_{TG})),\end{split} (8)

where θTLi,θTGiRh×w×c{\theta}^{i}_{TL},{\theta}^{i}_{TG}\in R^{h{\times}w{\times}c} are the gated weights of LTiL^{i}_{T} and GTiG^{i}_{T}, respectively. 𝕎TLi\mathbb{W}^{i}_{TL} and 𝕎TGi\mathbb{W}^{i}_{TG} are the parameters of convolutional layers, and Sigm()\textit{Sigm}() is an element-wise sigmoid function. In this same way, we can also compute the gated weights θILi,θIGiRh×w×c{\theta}^{i}_{IL},{\theta}^{i}_{IG}\in R^{h{\times}w{\times}c} for the information LIiL^{i}_{I} and GIiG^{i}_{I}. Finally, we re-weight each information with individual gated weight and then preform the dynamic message propagation. Therefore, Eq. 6 and 7 have become:

f^Ii=fIi+θTLiLTi+θTGiGTi,f^Ti=fTi+θILiLIi+θIGiGIi,\begin{split}\hat{f}_{I}^{i}&=f_{I}^{i}+{\theta}^{i}_{TL}\otimes L^{i}_{T}+{\theta}^{i}_{TG}\otimes G^{i}_{T},\\ \hat{f}_{T}^{i}&=f_{T}^{i}+{\theta}^{i}_{IL}\otimes L^{i}_{I}{~{}}+{\theta}^{i}_{IG}\otimes G^{i}_{I},\end{split} (9)

where \otimes denotes an element-wise multiplication.

IV-D Implementation Details

In this work, we implement the proposed CMMPNet on the representative deep learning platform PyTorch [79]. First, we perform data augmentation to alleviate the overfitting issue. Specifically, all training samples including the satellite images, trajectory heat-maps and ground-truth maps are i) flipped horizontally or vertically, ii) rotated by 90, 180, 270 degrees, iii) randomly cropped with a size range of [0.7, 0.9] and resized to the original resolution. After augmentation, the number of training samples is enlarged by 7 times. We then determine the hyper-parameters of our framework. The filter weights of all convolutional layers and Fully-Connected layers are uniformly initialized by Xavier [80]. The batch size is set to 4 and the learning rate is set to 0.0002. Finally, we apply the Adam [81] optimizer to train our CMMPNet for 30 epochs by minimizing the Binary Cross-Entropy Loss between the generated road maps and the corresponding ground-truth maps.

V Experiments

In this section, we first introduce the experiment settings of image+trajectory based road extraction. We then compare the proposed CMMPNet with existing state-of-the-art approaches and finally conduct extensive ablation studies to verify the effectiveness of each component in our network.

V-A Settings

Datasets: In this work, our experiments are mainly conducted on the BJRoad dataset [27], which is captured in Beijing, China. Specifically, this benchmark consists of 350 high-resolution aerial images that cover a large geographic area of about 100 square kilometers and around 50 million trajectory records of 28 thousand vehicles. The resolution of aerial images is 1024×10241024\times 1024 and each pixel denotes a 0.5m×\times0.5m region in the real world. For each aerial image, a 1024×10241024\times 1024 trajectory heat-map is generated with the preprocessing described in Section III, and the corresponding ground-truth (GT) map is manually created by masking out the pixel of traffic roads. Finally, this dataset is officially divided into three partitions: 70% samples are adopted for training, 10% for validation, and the rest 20% for testing.

Following the previous work [29], we also perform experiments on the Porto dataset, which covers a geographic area of about 209 square kilometers in Porto, Portugal. This dataset contains a mass of crowdsourced trajectories generated by 442 taxis from 2013 to 2014. On this dataset, we adopt a five-fold cross-validation setting, since the details of training/testing sets are not provided in [29]. Specifically, the aerial image of the whole area is first cut into 6,048 non-overlapping sub-images with a resolution of 512×512512\times 512. These sub-images are then randomly divided into five equal parts. For the ii-th validation, the ii-th part is used for testing, and the remaining parts are used for training. Finally, the mean and variance of five validations are reported.

Evaluation Details: Given a probability map MM, we need to determine an estimated road map MeRH×WM_{e}\in R^{H{\times}W} before evaluation. Same to [27], a pixel (x,y)(x,y) is predicted as a road region in our work, if the response value of M(x,y)M(x,y) is greater than 0.5. Following previous works [82, 44], we adopt Intersection over Union111https://en.wikipedia.org/wiki/Jaccard_index (IoU) to evaluate the performance for road extraction. Specifically, the IoU score between an estimated map MeM_{e} and its corresponding ground-truth map MgM_{g} is computed by:

IoU(Me,Mg)=|MeMg||MeMg|,IoU(M_{e},M_{g})={\frac{|M_{e}{\cap}M_{g}|}{|M_{e}{\cup}M_{g}|}}, (10)

where |MeMg||M_{e}{\cap}M_{g}| denotes the pixel number in the intersection set of MeM_{e} and MgM_{g}, and |MeMg||M_{e}{\cup}M_{g}| is the pixel number in their union set. There are two manners for computing the IoU of all testing samples. The first manner is to compute the IoU of each sample and then average the IoU of all samples. Such a metric is termed as average IoU (A_IoU). The second manner is to stitch the estimated maps of all samples into a global map and then compute an IoU score. This metric is termed as global IoU (G_IoU). Since different IoU metrics were used in previous works, we would report the results of both A_IoU and G_IoU in the following sections.

TABLE II: The performance of different methods on the testing set of BJRoad dataset. Our CMMPNet outperforms all existing approaches with large margins.
Method A_IoU (%) G_IoU (%)
  DeepLab (v3+) [83] 50.81 -
              UNet [61] 54.88 -
         Res-UNet [43] 54.24 -
           LinkNet [45] 57.89 -
        D-LinkNet [44] 57.96 -
         Sun et al. [27] 59.18 -
DeepDualMapper [29] 60.91 61.54
  Res-UNet+CMMPNet 62.58 63.03
    LinkNet+CMMPNet 63.09 63.46
D-LinkNet+CMMPNet 62.85 63.39

V-B Comparison with State-of-the-Art Methods

In this section, we compare our CMMPNet with seven deep learning-based approaches, including DeepLab (v3+) [83], UNet [61], Res-UNet [43], LinkNet [45], D-LinkNet [44], Sun et al. [27], and DeepDualMapper [29]. Specifically, these compared methods are reimplemented for multimodal road extraction. In particular, DeepDualMapper feeds aerial images and trajectory heat-maps into different backbone networks222In DeepDualMapper, the original backbone network is UNet. However, our reimplemented DeepDualMapper based on UNet performs poorly. Thus in this work, we adopt D-LinkNet as the backbone to reimplement DeepDualMapper and this model can obtain competitive performance on different datasets. and then fuses their features with a gated fusion module, while other methods directly take the concatenation of aerial images and trajectory heat-maps as input. Moreover, all the compared methods except DeepLab (v3+) and DeepDualMapper are equipped with 1D transpose convolution to better model traffic roads [27]. Notice that the first six compared methods were implemented by [27], and we utilize the official code of [27] to implement DeepDualMapper and our method with the same data partition. As mentioned above, our CMMPNet can be developed with various AutoEncoders. We hence evaluate multiple implementations of CMMPNet based on different AutoEncoders, such as Res-UNet, LinkNet, and D-LinkNet.

TABLE III: The performance of different methods on the PORTO dataset. Five-fold cross-validation is conducted on this dataset. The mean and variance of five validations are reported in this table.
Method A_IoU (%) G_IoU (%)
        D-LinkNet [44] 72.82±\pm0.47 72.92±\pm0.45
         Sun et al. [27] 72.94±\pm0.71 73.04±\pm0.63
DeepDualMapper [29] 73.67±\pm0.51 73.91±\pm0.51
D-LinkNet+CMMPNet 74.56±\pm0.46 74.66±\pm0.41

Refer to caption

Figure 6: Visualization of the feature maps and traffic road maps generated with/without global information on the testing set of BJRoad dataset. (a) is the input aerial image with trajectory points, and (e) is the ground-truth road map. (b) and (c) are the average maps of image feature and trajectory feature after the 1st/7th DEM without the global message, while (d) is the generated road network without the global message. (f), (g) and (h) are the generated feature maps and the road network using both the local message and global message. We can observe that our method can generate more discriminative features and recognize the occluded/unimpressive traffic roads effectively when performing road extraction with global information.

The performance of all methods on the BJRoad dataset is summarized in Table II. We can observe that DeepLab (v3+) obtains the worst A_IoU 50.81% probably because of parameter overfitting. Sun et al. [27] utilized various techniques (e.g., different sampling intervals and GPS augmentation) to obtain an improved A_IoU 59.18%. However, just directly feeding the concatenation of aerial images and trajectory heat-maps into networks, these methods have limited capabilities to capture the multimodal information, thus none of them can acquire an A_IoU above 60%. By fusing image and trajectory features with a gated fusion module, DeepDualMapper obtains a competitive A_IoU 60.91% and G_IoU 61.54%. Despite the progress, DeepDualMapper only use a fusion strategy rather than a mutual refinement strategy, thereby cannot address this task well. In contrast, when learning modality-specific features explicitly and enhancing cross-modal features mutually, our method can fully exploit the complementary information of aerial images and crowdsourced trajectories. For this reason, the proposed CMMPNet outperforms all previous methods with large margins. For instance, Res-UNet+CMMPNet achieves a competitive A_IoU 62.58% and obtains a relative improvement of 15.37%, compared with the original Res-UNet. By improving the A_IoU from 57.96% to 62.85%, our D-LinkNet+CMMPNet also obtains a substantial improvement of 8.4%, compared with the baseline D-LinkNet. Finally, with an impressive A_IoU 63.09% and a G_IoU 63.46%, our LinkNet+CMMPNet becomes the best-performing model. We notice that the performance of D-LinkNet+CMMPNet is slightly lower than that of LinkNet+CMMPNet. This is probably because D-LinkNet+CMMPNet contains two extra interim units and suffers from certain overfitting, although data augmentation has been performed.

Moreover, we compare the performance of our CMMPNet with three competitive models including D-LinkNet [44], Sun et al. [27], and DeepDualMapper [29] on the Porto dataset. As shown in Table III, all methods obtain much better results on this dataset, compared with their performance on the BJRoad dataset. The main reason is that the aerial images of Porto are clearer and the noises of trajectories are smaller [29]. Despite the existing benchmarks are high, our CMMPNet still can boost the IOU with substantial margins, ranking first in performance on the Porto dataset. In summary, these comparisons greatly demonstrate the effectiveness of the proposed CMMPNet for image+trajectory based road extraction.

V-C Component Analysis

After external comparison, we then perform extensive internal experiments to analyze the effectiveness of each module in the proposed CMMPNet. In this subsection, D-LinkNet is adopted as the backbone network and our implementation details have been described in Section IV-D.

TABLE IV: The influence of Non-Local Message Propagator and Gated Message Propagator on the testing set of BJRoad dataset.
Local Global Gate A_IoU (%) G_IoU (%)
61.62 62.09
62.32 62.78
61.98 62.43
62.85 63.39

The effect of global message: In previous works [84, 85], local information is widely adopted, but global information is neglected. In this section, we implement several variants of CMMPNet to verify the effectiveness of global information. As shown in Table IV, when propagating the global information extracted by SPP and FC layer, “Local+Global” model obtains an A_IoU 61.98% and a G_IoU 62.43%, and is better than “Local” model. With an A_IoU 62.85% and a G_IoU 63.39%, “Local+Global+Gate” model also outperforms “Local+Gate” model, whose A_IoU is 62.32% and G_IoU is 62.78%. Except for quantitative results, we also visualize some feature maps and traffic road maps generated by “Local+Gate” model and “Local+Global+Gate” model in Fig. 6. Note that those visualized feature maps are the channel-wise average of image features and trajectory features after DEM. We can observe that incorporating global information can generate more discriminative features and better recognize traffic roads, especially when the roads are occluded/unimpressive and the vehicle trajectories are rare in local regions. These quantitative and qualitative experiments show that global information is greatly effective for traffic road extraction.

The effect of Gated Message Propagator: In this propagator, multiple gate functions are employed to dynamically propagate the complementary information. In this subsection, we also implement several variants to verify the effectiveness of this mechanism. As shown in Table IV, after applying gate functions on “Local” model, A_IoU increases from 61.62% to 62.32% and G_IoU increases from 62.09% to 62.78%. Further, we can obtain a more substantial improvement (around 1% on both A_IoU and G_IoU), when performing gate functions on “Local+Global” model. These comparisons show that this proposed propagator can facilitate robust road extraction using multimodal information.

Refer to caption
Figure 7: Visualization of the traffic road networks generated by different methods on the testing set of BJRoad dataset. (a) and (b) are the input aerial images and trajectories heat-maps. (c) are the results that only aerial images are taken as input, while (d) are the results that the concatenation of images and heat-maps are taken as input. As shown in (e), the results of our CMMPNet are more accurate and are very similar to the ground-truth road networks.

Refer to caption

Figure 8: The influence of the level number NN of Spatial Pyramid Pooling layer in Dual Enhancement Module on the testing set of BJRoad dataset. Our method achieves the best performance when NN is set to 3.

The configuration of Spatial Pyramid Pooling: In Non-Local Message Propagator, we employ a NN-level SPP and an FC layer to extract global information. In this section, we explore the effect of the level number for road extraction using multimodal information. As shown in Fig. 8, when applying a global max-pooling (NN=1), our CMMPNet obtains an A_IoU 62.38% and is slightly better than the “Local+Gate” model in Table IV, since global pooling can only provide some coarse and limited information. As the level number increases, the performance also gradually increases, and our method achieves the best A_IoU 62.85% and G_IoU 63.39% when NN is equal to 3. When nn increases to 4, the performance slightly drops, probably because of overfitting, i.e., the amount of parameters of the FC layer in DEM increases sharply as the level number increase. Therefore, the level number NN of Spatial Pyramid Pooling is uniformly set to 3 for road extraction.

V-D More Discussion

Unimodal Data vs. Multi-modal Data: We first explore whether multimodal data is reliably useful for traffic road extraction. As shown in Table V, when only feeding trajectory heat-maps into a D-LinkNet, we obtain a poor performance (A_IoU 52.38%, G_IoU 52.90%) on the BJRoad dataset. When only utilizing aerial images, we obtain an A_IoU 59.79% and a G_IoU 60.24%, which indicates that image data is more crucial than trajectory data. In contrast, when using the aerial images and trajectory heat-maps simultaneously, our CMMPNet and the early/late fusion models described in the next paragraph outperform the unimodal models consistently with an improvement of at least 1% on IoU. This comparison demonstrates that multimodal data is more effective for traffic road extraction, because aerial images and vehicle crowdsourced trajectories have rich complementarities.

Which multimodal learning manner is better? We then explore the effects of different multimodal learning manners. Except for the proposed CMMPNet, we also implement another two commonly-used manners, i.e., early fusion model and late fusion model. Specifically, the former feeds the concatenation of aerial images and trajectory heat-maps into a D-LinkNet. In the latter, aerial images and trajectory heat-maps are respectively fed into individual D-LinkNet, and their final features are concatenated to estimate the road maps. As shown in Table V, the early fusion model obtains an A_IoU 61.11% and a G_IoU 61.53%, slightly outperforming the late fusion model (A_IoU 60.78%, G_IoU 61.24%). This is because the multimodal information is utilized at different layers in the former, but just utilized once in the latter. Compared with these two models, our CMMPNet is more reasonable to learn modality-specific features and propagate cross-modal information hierarchically. For this reason, our method achieves an impressive A_IoU 62.85% and G_IoU 63.39%, and outperforms early/late fusion models with a large margin. This comparison shows the effectiveness of our CMMPNet for multimodal representation learning.

TABLE V: The performance of different inputs and different representation learning manners on the testing set of BJRoad dataset.
Input Learning Manner A_IoU (%) G_IoU (%)
Trajectory - 52.38 52.90
Image - 59.79 60.24
Image+Trajectory   Late Fusion 60.78 61.24
Early Fusion 61.11 61.53
CMMPNet 62.85 63.39

Significance of crowdsourced trajectories: Although the IoU of aerial images is much better than that of trajectories, we argue that vehicle trajectories are crucial for the robustness of road extraction, especially when some cities (e.g, Chongqing and Chengdu, China) are greatly covered by fog and mist in aerial images. So here we explore to extract traffic roads from foggy images and crowdsourced trajectories. Since there are no foggy images in the BJRoad dataset, we need to generate some aerial images with heavy fog in advance. Specifically, for each cloudless image in BJRoad, we employ a fog effect renderer of Photoshop to generate a foggy image. After augmenting the training samples as described in Section IV-D, we reimplement the proposed CMMPNet and three other compared methods, including (1) two unimodal models which feed foggy images or trajectory heat-maps into D-LinkNet, and (2) an early fusion D-LinkNet model which takes the concatenation of foggy images and trajectory heat-maps as input.

Refer to caption

Figure 9: The first row shows some foggy images and mass vehicle trajectories on the testing set of the foggy BJRoad dataset. Although traffic roads are occluded extremely in these images, our CMMPNet can still generate high-quality road network maps by fully exploiting the complementary information of vehicle crowdsourced trajectories, as shown in the second row.
TABLE VI: The performance of traffic road extraction based on foggy images and vehicle trajectories on the testing set of the foggy BJRoad dataset.
Input Manner Way A_IoU (%) G_IoU (%)
Trajectory - 52.38 52.90
Fog_Img - 54.54 55.27
Fog_Img + Trajectory Early Fusion 57.98 58.49
CMMPNet 60.45 61.06

Refer to caption

Figure 10: The performance of different methods on the TLCGIS dataset. The proposed CMMPNet also outperforms all existing approaches for image+Lidar based road extraction.

The results of all methods are summarized in Table VI. We can observe that the unimodal D-LinkNet only obtains an A_IoU 54.54% and a G_IoU 55.27% when only using foggy images. Compared with the corresponding model using cloudless images, this model has a dramatic drop in performance, since traffic roads may be invisible in foggy images. When utilizing foggy images and trajectories simultaneously, the early fusion model obtains an A_IoU 57.98% and a G_IoU 58.49%. Based on the same D-LinkNet, our CMMPNet achieves a competitive A_IoU 60.45% and G_IoU 61.06%, having a performance improvement of at least 3% compared with other models. Moreover, the visualizations in Fig. 9 show that our CMMPNet can still generate high-quality road maps in foggy weather conditions. This is attributed to the fact that the vehicle trajectories can provide rich information to remedy the limitation of aerial images, and our method can fully capture their complementary information. In summary, crowdsourced trajectories are very crucial and beneficial for robust road extraction.

Refer to caption
Figure 11: Visualization of the generated traffic road maps on the TLCGIS dataset. (a) and (b) are the input aerial images and Lidar images. (c) are the results that only aerial images are taken as input. (d) are the results of our CMMPNet that utilizes both aerial images and Lidar data. (e) are the ground-truth road maps.

VI Apply to Image+Lidar based Extraction

As mentioned above, our method is general for road extraction by exploiting multimodal information. In this section, we employ the proposed CMMPNet to recognize traffic roads from aerial images and Lidar data. As shown in Fig. 11-(a,b), Lidar data can help to discover some occluded or inconspicuous roads in aerial images. Here we conduct extensive experiments on the TLCGIS [59] dataset, which consists of 5,860 pairs of aerial images and Lidar images rendered from raw Lidar point cloud data. The resolution of these images is 500×\times500 and the geographical length of each pixel is 0.5 feet. This dataset is officially divided into training, test, and validation sets with each having 2,640, 2,400, and 240 samples respectively. On this dataset, we also take D-LinkNet as the backbone to develop our CMMPNet and optimize this model with the process described in Section IV-D.

VI-A Comparison with State-of-the-Art Metods

In this subsection, we compare our CMMPNet with six state-of-the-art methods on the TLCGIS dataset. The details of these compared methods are described as follows. SegNet [60]: As a fully convolutional AutoEncoder, SegNet takes the concatenation of aerial images and Lidar images as input. PreConv [59]: The Lidar images are first fed into a depth convolution unit (DepthCNN) implemented with two convolutional layers. The Lidar features and aerial images are then concatenated and fed into SegNet. RFNorm [59]: Given aerial and Lidar images, some Random Forest classifiers [86] are first trained to estimate the road probability score at each location. The aerial-based and Lidar-based score maps are concatenated and fed into SegNet. ElePreConv [59]: In this model, Lidar images are first encoded with two convolutions (8 and 4 filters), while aerial images are extended with an extra zero-initialized channel. The element-wise addition of 4-channel images and Lidar features are fed into FuseNet [87]. TriSeg [59]: This model consists of three SegNets. The first two SegNets respectively take aerial or Lidar images to generate the road probability maps, which are concatenated and fed into the third SegNet for final estimation. DeepDualMapper [29]: This model has been described above and here we adopt D-LinkNet as the backbone network to reimplement this model.

The performance of all approaches is summarized in Fig. 10. We can observe that the previous best-performing methods are TriSeg and DeepDualMapper, whose G_IoU are 78.4% and 81.1%, respectively. Thanks to the cross-modal mutual refinement strategy, our CMMPNet achieves a new state-of-the-art G_IoU 83.1% on the TLCGIS dataset and greatly outperforms DeepDualMapper with an absolute improvement of 2%. Moreover, we also visualize some results in Fig. 11. As can be observed, the traffic road maps generated by our method are more accurate in complex scenarios. In summary, these quantitative and qualitative comparisons demonstrate that our CMMPNet is universal and effective to extract traffic roads from aerial images and Lidar data.

TABLE VII: The performance of different inputs and different representation learning manners on the testing set of TLCGIS dataset.
Input Learning Manner G_IoU (%)
Lidar - 69.12
Image - 80.96
Image+Lidar   Late Fusion 81.50
Early Fusion 81.62
CMMPNet 83.10
TABLE VIII: The influence of Non-Local Message Propagator and Gated Message Propagator on the testing set of TLCGIS dataset.
Local Global Gate G_IoU (%)
81.88
82.31
82.06
83.10

VI-B Internal Analysis

In this subsection, we verify the effectiveness of each component in the proposed CMMPNet for image+Lidar based road extraction. We first explore which manner can better exploit the information of these modalities. As shown in Table VII, we can obtain a G_IoU of 69.12%, when only feeding the rendered Lidar images into D-LinkNet. When only using aerial images, the G_IoU of D-LinkNet is 80.96%, which indicates that aerial images are more important. Incorporating the information of aerial and Lidar images simultaneously, the early fusion model obtains a G_IoU of 81.62%, while the late fusion model has a comparable G_IoU of 81.50%. When fully exploring their complementary information with cross-modal message propagators, our CMMPNet achieves an impressive G_IoU 83.10%, outperforming the early/late fusion models with an absolute improvement of 1.5%. This demonstrates that the proposed CMMPNet can also effectively capture the complementary information among aerial images and Lidar data.

We then explore the effect of global information and gate functions. Similar to Section V-C, we implement several variants of CMMPNet. As shown in Table VIII, when only propagating local information with gate functions, “Local+Gate” model obtains a G_IoU 82.31%. When incorporating global information, “Local+Global+Gate” model has a better G_IoU 83.10%, which indicates that the global information is also useful for image+Lidar based road extraction. Moreover, by comparing the performance of “Local+Global” model and “Local+Global+Gate” model, we can observe that the gate functions help to make an absolute improvement of 1.04% on G_IoU, which also demonstrates the effectiveness of Gated Message Propagator for image+Lidar based road extraction.

VII Conclusion

In this work, we investigate a challenging task for land remote sensing analysis, i.e., how to robustly extract traffic roads using the complementary information of aerial images and vehicle crowdsourced trajectories. To this end, we introduce a novel Cross-Modal Message Propagation Network (CMMPNet), which learns modality-specific features explicitly with two individual AutoEncoders and enhances these features mutually with a tailor-designed Dual Enhancement Module. Specifically, we comprehensively extract and dynamically propagate the complementary information of each modality to enhance the representation of another modality. Extensive experiments conducted on two real-world benchmarks show that the proposed CMMPNet is not only effective for image+trajectory based road extraction, but also suitable for image+Lidar based road extraction.

Nevertheless, there are still several issues worthy of further study. First, the connectivity of traffic roads has not been explicitly explored in conventional works. Intuitively, the temporal information of vehicle trajectories could be utilized to distinguish disconnected road regions (e.g., urban roads are usually separated by fences and green belts). However, existing image+trajectory datasets lack the road connectivity annotation. To facilitate the researches in this field, we will construct a large-scale multimodal road extraction with rich connectivity annotation and propose a multimodal spatial-temporal framework to explicitly estimate the road connectivity in future work. Second, some elevated roads at different heights are overlapped on aerial images. The height information accessed with GPS devices is relatively coarse. Thus in future work, we will also develop some advanced approaches to effectively recognize the roads at different heights with the coarse height information of crowdsourced trajectories.

References

  • [1] W. Von Engelhardt, J. Zimmermann, and J. Zimmerman, Theory of earth science.   CUP Archive, 1988.
  • [2] N. R. Council et al., Basic research opportunities in earth science.   National Academies Press, 2001.
  • [3] G. R. Keller and C. Baru, Geoinformatics: Cyberinfrastructure for the solid earth sciences.   Cambridge University Press, 2011.
  • [4] Y. Zhang, Y.-L. Hsueh, W.-C. Lee, and Y.-H. Jhang, “Efficient cache-supported path planning on roads,” IEEE Transactions on Knowledge and Data Engineering, vol. 28, no. 4, pp. 951–964, 2015.
  • [5] T. Wang, Y. Zhao, J. Wang, A. K. Somani, and C. Sun, “Attention-based road registration for gps-denied uas navigation,” IEEE Transactions on Neural Networks and Learning Systems, 2020.
  • [6] Q. Wang, T. Han, Z. Qin, J. Gao, and X. Li, “Multitask attention network for lane detection and fitting,” IEEE Transactions on Neural Networks and Learning Systems, 2020.
  • [7] M. Wang and C. Luo, “Extracting roads based on gauss markov random field texture model and support vector machine from high-resolution rs image,” IEEE Transaction on Geoscience and Remote Sensing, vol. 9, pp. 271–276, 2005.
  • [8] S. Movaghati, A. Moghaddamjoo, and A. Tavakoli, “Road extraction from satellite images using particle filtering and extended Kalman filtering,” IEEE Transactions on Geoscience and Remote Sensing, vol. 48, no. 7, pp. 2807–2817, 2010.
  • [9] W. Shi, Z. Miao, and J. Debayle, “An integrated method for urban main-road centerline extraction from optical remotely sensed imagery,” IEEE Transactions on Geoscience and Remote Sensing, vol. 52, no. 6, pp. 3359–3372, 2013.
  • [10] J. A. Suykens and J. Vandewalle, “Least squares support vector machine classifiers,” Neural Processing Letters, vol. 9, no. 3, pp. 293–300, 1999.
  • [11] S. Z. Li, Markov random field modeling in image analysis.   Springer Science and Business Media, 2009.
  • [12] G. Cheng, Y. Wang, S. Xu, H. Wang, S. Xiang, and C. Pan, “Automatic road detection and centerline extraction via cascaded end-to-end convolutional neural network,” IEEE Transactions on Geoscience and Remote Sensing, vol. 55, no. 6, pp. 3322–3337, 2017.
  • [13] A. Buslaev, S. S. Seferbekov, V. Iglovikov, and A. Shvets, “Fully convolutional network for automatic road extraction from satellite imagery.” in IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2018, pp. 207–210.
  • [14] X. Lu, Y. Zhong, Z. Zheng, Y. Liu, J. Zhao, A. Ma, and J. Yang, “Multi-scale and multi-task deep learning framework for automatic road extraction,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 11, pp. 9362–9377, 2019.
  • [15] S. Rogers, P. Langley, and C. Wilson, “Mining gps data to augment road models,” in ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 1999, pp. 104–113.
  • [16] S. Schroedl, K. Wagstaff, S. Rogers, P. Langley, and C. Wilson, “Mining gps traces for map refinement,” Data Mining and Knowledge Discovery, vol. 9, no. 1, pp. 59–87, 2004.
  • [17] J. J. Davies, A. R. Beresford, and A. Hopper, “Scalable, distributed, real-time map generation,” IEEE Pervasive Computing, vol. 5, no. 4, pp. 47–54, 2006.
  • [18] X. Liu, J. Biagioni, J. Eriksson, Y. Wang, G. Forman, and Y. Zhu, “Mining large-scale, sparse gps traces for map inference: comparison of approaches,” in ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2012, pp. 669–677.
  • [19] J. Biagioni and J. Eriksson, “Map inference in the face of noise and disparity,” in International Conference on Advances in Geographic Information Systems, 2012, pp. 79–88.
  • [20] Y. Wang, X. Liu, H. Wei, G. Forman, C. Chen, and Y. Zhu, “Crowdatlas: Self-updating maps for cloud and personal use,” in Annual International Conference on Mobile Systems, Applications, and Services, 2013, pp. 27–40.
  • [21] Z. Shan, H. Wu, W. Sun, and B. Zheng, “Cobweb: a robust map update system using gps trajectories,” in ACM International Joint Conference on Pervasive and Ubiquitous Computing, 2015, pp. 927–937.
  • [22] S. Karagiorgou, D. Pfoser, and D. Skoutas, “A layered approach for more robust generation of road network maps from vehicle tracking data,” ACM Transactions on Spatial Algorithms and Systems, vol. 3, no. 1, pp. 1–21, 2017.
  • [23] W. Shi, S. Shen, and Y. Liu, “Automatic generation of road network map from massive gps, vehicle trajectories,” in International IEEE Conference on Intelligent Transportation Systems, 2009, pp. 1–6.
  • [24] C. Chen, C. Lu, Q. Huang, Q. Yang, D. Gunopulos, and L. Guibas, “City-scale map creation and updating using gps collections,” in ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2016, pp. 1465–1474.
  • [25] W. Yang, T. Ai, and W. Lu, “A method for extracting road boundary information from crowdsourcing vehicle gps trajectories,” Sensors, vol. 18, no. 4, p. 1261, 2018.
  • [26] S. Ruan, C. Long, J. Bao, C. Li, Z. Yu, R. Li, Y. Liang, T. He, and Y. Zheng, “Learning to generate maps from trajectories,” in AAAI Conference on Artificial Intelligence, 2020.
  • [27] T. Sun, Z. Di, P. Che, C. Liu, and Y. Wang, “Leveraging crowdsourced gps data for road extraction from aerial imagery,” in IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 7509–7518.
  • [28] Y. Li, L. Xiang, C. Zhang, and H. Wu, “Fusing taxi trajectories and rs images to build road map via dcnn,” IEEE Access, vol. 7, pp. 161 487–161 498, 2019.
  • [29] H. Wu, H. Zhang, X. Zhang, W. Sun, B. Zheng, and Y. Jiang, “Deepdualmapper: A gated fusion network for automatic map extraction using aerial images and trajectories,” in AAAI Conference on Artificial Intelligence, 2020.
  • [30] https://en.wikipedia.org/wiki/Earth_science.
  • [31] G. A. Fine, Authors of the Storm: Meteorologists and the Culture of Prediction.   University of Chicago Press, 2009.
  • [32] M. Birylo, J. Nastula, and J. Kuczynska-Siehien, “The creation of flood risks model using a combination of satellite and meteorological models–the first step,” Acta Geodynamica et Geomaterialia, vol. 12, no. 2, pp. 151–156, 2015.
  • [33] Y. Y. Kagan and D. D. Jackson, “Probabilistic forecasting of earthquakes,” Geophysical Journal International, vol. 143, no. 2, pp. 438–453, 2000.
  • [34] R. N. Clark and A. N. Rencz, “Spectroscopy of rocks and minerals, and principles of spectroscopy,” Manual of remote sensing, vol. 3, pp. 3–58, 1999.
  • [35] W. Wang, N. Yang, Y. Zhang, F. Wang, T. Cao, and P. Eklund, “A review of road extraction from remote sensing images,” Journal of Traffic and Transportation Engineering, vol. 3, no. 3, pp. 271–282, 2016.
  • [36] H. Zhang, Y. Liao, H. Yang, G. Yang, and L. Zhang, “A local-global dual-stream network for building extraction from very-high-resolution remote sensing images,” IEEE Transactions on Neural Networks and Learning Systems, 2020.
  • [37] Q. Lin, J. Zhao, G. Fu, and Z. Yuan, “Crpn-sfnet: A high-performance object detector on large-scale remote sensing images,” IEEE Transactions on Neural Networks and Learning Systems, 2020.
  • [38] S. Hinz and A. Baumgartner, “Automatic extraction of urban road networks from multi-view aerial imagery,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 58, no. 1-2, pp. 83–98, 2003.
  • [39] P. Anil and S. Natarajan, “A novel approach using active contour model for semi-automatic road extraction from high resolution satellite imagery,” in IEEE International Conference on Machine Learning and Computing, 2010, pp. 263–266.
  • [40] D. Chaudhuri, N. Kushwaha, and A. Samal, “Semi-automated road detection from high resolution satellite images by directional morphological enhancement and segmentation techniques,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 5, no. 5, pp. 1538–1544, 2012.
  • [41] S. Leninisha and K. Vani, “Water flow based geometric active deformable model for road network,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 102, pp. 140–147, 2015.
  • [42] Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, 2015.
  • [43] Z. Zhang, Q. Liu, and Y. Wang, “Road extraction by deep residual u-net,” IEEE Geoscience and Remote Sensing Letters, vol. 15, no. 5, pp. 749–753, 2018.
  • [44] L. Zhou, C. Zhang, and M. Wu, “D-linknet: Linknet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction.” in IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2018, pp. 182–186.
  • [45] A. Chaurasia and E. Culurciello, “Linknet: Exploiting encoder representations for efficient semantic segmentation,” in IEEE Visual Communications and Image Processing, 2017, pp. 1–4.
  • [46] G. Fu, C. Liu, R. Zhou, T. Sun, and Q. Zhang, “Classification for high resolution remote sensing imagery using a fully convolutional network,” Remote Sensing, vol. 9, no. 5, p. 498, 2017.
  • [47] S. Edelkamp and S. Schrödl, “Route planning and map inference with global positioning traces,” in Computer Science in Perspective.   Springer, 2003, pp. 128–151.
  • [48] R. Stanojevic, S. Abbar, S. Thirumuruganathan, S. Chawla, F. Filali, and A. Aleimat, “Robust road map inference through network alignment of trajectories,” in SIAM International Conference on Data Mining, 2018, pp. 135–143.
  • [49] L. Cao and J. Krumm, “From gps traces to a routable road map,” in ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, 2009, pp. 3–12.
  • [50] B. Niehoefer, R. Burda, C. Wietfeld, F. Bauer, and O. Lueert, “Gps community map generation for enhanced routing methods based on trace-collection by mobile phones,” in IEEE International Conference on Advances in Satellite and Space Communications, 2009, pp. 156–161.
  • [51] S. Wang, Y. Wang, and Y. Li, “Efficient map reconstruction and augmentation via topological methods,” in SIGSPATIAL International Conference on Advances in Geographic Information Systems, 2015, pp. 1–10.
  • [52] G. R. Terrell and D. W. Scott, “Variable kernel density estimation,” The Annals of Statistics, pp. 1236–1265, 1992.
  • [53] S. Clode, “The automatic extraction of roads from lidar data,” International Archives of the Photogrammetry, Remote Sensing, and Spatial Information Sciences, vol. 3, pp. 231–236, 2004.
  • [54] Z. Hui, Y. Hu, S. Jin, and Y. Z. Yevenyo, “Road centerline extraction from airborne lidar point cloud based on hierarchical fusion and optimization,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 118, pp. 22–36, 2016.
  • [55] W. Zhang, “Lidar-based road and road-edge detection,” in IEEE Intelligent Vehicles Symposium, 2010, pp. 845–848.
  • [56] X. Hu, Y. Li, J. Shan, J. Zhang, and Y. Zhang, “Road centerline extraction in complex urban scenes from lidar data based on multiple features,” IEEE Transactions on Geoscience and Remote Sensing, vol. 52, no. 11, pp. 7448–7456, 2014.
  • [57] Y. Li, L. Ma, Z. Zhong, F. Liu, M. A. Chapman, D. Cao, and J. Li, “Deep learning for lidar point clouds in autonomous driving: a review,” IEEE Transactions on Neural Networks and Learning Systems, 2020.
  • [58] X. Hu, C. V. Tao, and Y. Hu, “Automatic road extraction from dense urban area by integrated processing of high resolution imagery and lidar data,” International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences, vol. 35, no. B3, pp. 288–292, 2004.
  • [59] B. Parajuli, P. Kumar, T. Mukherjee, E. Pasiliao, and S. Jambawalikar, “Fusion of aerial lidar and images for road segmentation with deep cnn,” in ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, 2018, pp. 548–551.
  • [60] V. Badrinarayanan, A. Kendall, and R. Cipolla, “Segnet: A deep convolutional encoder-decoder architecture for image segmentation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 12, pp. 2481–2495, 2017.
  • [61] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, 2015, pp. 234–241.
  • [62] S. Ross, D. Munoz, M. Hebert, and J. A. Bagnell, “Learning message-passing inference machines for structured prediction,” in IEEE Conference on Computer Vision and Pattern Recognition, 2011, pp. 2737–2744.
  • [63] J. Winn and C. M. Bishop, “Variational message passing,” Journal of Machine Learning Research, vol. 6, no. Apr, pp. 661–694, 2005.
  • [64] D. Wang, W. Ouyang, W. Li, and D. Xu, “Dividing and aggregating network for multi-view action recognition,” in European Conference on Computer Vision, 2018, pp. 451–467.
  • [65] L. Liu, Z. Qiu, G. Li, S. Liu, W. Ouyang, and L. Lin, “Crowd counting with deep structured scale integration network,” in IEEE International Conference on Computer Vision, 2019, pp. 1774–1783.
  • [66] L. Zhang, D. Xu, A. Arnab, and P. H. Torr, “Dynamic graph message passing networks,” in IEEE Conference on Computer Vision and Pattern Recognition, 2020, pp. 3726–3735.
  • [67] Z. Zhong, C.-T. Li, and J. Pang, “Hierarchical message-passing graph neural networks,” arXiv preprint arXiv:2009.03717, 2020.
  • [68] L. Liu, J. Chen, H. Wu, J. Zhen, G. Li, and L. Lin, “Physical-virtual collaboration modeling for intra-and inter-station metro ridership prediction,” IEEE Transactions on Intelligent Transportation Systems, 2020.
  • [69] I. Spinelli, S. Scardapane, and A. Uncini, “Adaptive propagation graph convolutional network,” IEEE Transactions on Neural Networks and Learning Systems, 2020.
  • [70] T. Chen, T. Pu, H. Wu, Y. Xie, L. Liu, and L. Lin, “Cross-domain facial expression recognition: A unified evaluation benchmark and adversarial graph learning,” IEEE transactions on pattern analysis and machine intelligence, 2021.
  • [71] Z. Wang, X. Liu, H. Li, L. Sheng, J. Yan, X. Wang, and J. Shao, “Camp: Cross-modal adaptive message passing for text-image retrieval,” in IEEE International Conference on Computer Vision, 2019, pp. 5764–5773.
  • [72] L. Liu, Z. Qiu, G. Li, Q. Wang, W. Ouyang, and L. Lin, “Contextualized spatial–temporal network for taxi origin-destination demand prediction,” IEEE Transactions on Intelligent Transportation Systems, vol. 20, no. 10, pp. 3875–3887, 2019.
  • [73] J. Lou, Y. Jiang, Q. Shen, R. Wang, and Z. Li, “Probabilistic regularized extreme learning for robust modeling of traffic flow forecasting,” IEEE Transactions on Neural Networks and Learning Systems, 2020.
  • [74] L. Liu, J. Zhen, G. Li, G. Zhan, Z. He, B. Du, and L. Lin, “Dynamic spatial-temporal representation learning for traffic flow prediction,” IEEE Transactions on Intelligent Transportation Systems, 2020.
  • [75] https://en.wikipedia.org/wiki/Gaussian_blur.
  • [76] K. He, X. Zhang, S. Ren, and J. Sun, “Spatial pyramid pooling in deep convolutional networks for visual recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 37, no. 9, pp. 1904–1916, 2015.
  • [77] L. Zhang, J. Dai, H. Lu, Y. He, and G. Wang, “A bi-directional message passing model for salient object detection,” in IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 1741–1750.
  • [78] X. Wang, R. Girshick, A. Gupta, and K. He, “Non-local neural networks,” in IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 7794–7803.
  • [79] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga et al., “Pytorch: An imperative style, high-performance deep learning library,” in Advances in Neural Information Processing Systems, 2019, pp. 8026–8037.
  • [80] X. Glorot and Y. Bengio, “Understanding the difficulty of training deep feedforward neural networks,” in International Conference on Artificial Intelligence and Statistics, 2010, pp. 249–256.
  • [81] D. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv:1412.6980, 2014.
  • [82] J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 3431–3440.
  • [83] L.-C. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in European Conference on Computer Vision, 2018, pp. 801–818.
  • [84] G. Lin, C. Shen, I. Reid, and A. van den Hengel, “Deeply learning the messages in message passing inference,” in Advances in Neural Information Processing Systems, 2015, pp. 361–369.
  • [85] M. T. Teichmann and R. Cipolla, “Convolutional crfs for semantic segmentation,” in British Machine Vision Conference, 2019.
  • [86] A. Liaw, M. Wiener et al., “Classification and regression by randomforest,” R news, vol. 2, no. 3, pp. 18–22, 2002.
  • [87] C. Hazirbas, L. Ma, C. Domokos, and D. Cremers, “Fusenet: Incorporating depth into semantic segmentation via fusion-based cnn architecture,” in Asian conference on computer vision, 2016, pp. 213–228.
[Uncaptioned image] Lingbo Liu received the Ph.D degree from the School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, China, in 2020. From March 2018 to May 2019, he was a research assistant at the University of Sydney, Australia. He is currently a postdoctoral fellow in the Department of Computing at the Hong Kong Polytechnic University. His current research interests include machine learning and urban computing. He has authorized and co-authorized on more than 15 papers in top-tier academic journals and conferences. He has been serving as a reviewer for numerous academic journals and conferences such as TPAMI, TKDE, TNNLS, TITS, CVPR, ICCV and IJCAJ.
[Uncaptioned image] Zewei Yang received the B.E. degree from the School of Mathematics, Sun Yat-sen University, Guangzhou, China, in 2019, where she is currently pursuing the Master’s degree in applied mathematics. His current research interests include machine learning and data mining.
[Uncaptioned image] Guanbin Li is currently an associate professor in School of Computer Science and Engineering, Sun Yat-sen University. He received his PhD degree from the University of Hong Kong in 2016. His current research interests include computer vision, image processing, and deep learning. He is a recipient of ICCV 2019 Best Paper Nomination Award. He has authorized and co-authorized on more than 60 papers in top-tier academic journals and conferences. He serves as an associate editor for journal of The Visual Computer, an area chair for the conference of VISAPP. He has been serving as a reviewer for numerous academic journals and conferences such as TPAMI, IJCV, TIP, TMM, TCyb, CVPR, ICCV, ECCV and NeurIPS.
[Uncaptioned image] Kuo Wang received the B.E. degree from the School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, China, in 2020, where he is currently pursuing the Ph.D degree in computer science. His current research interests include deep learning and recommended system.
[Uncaptioned image] Tianshui Chen received a Ph.D. degree in computer science at the School of Data and Computer Science Sun Yat-sen University, Guangzhou, China, in 2018. Before that, he received a B.E. degree from the School of Information and Science Technology. He is currently the lecturer in the Guangdong University of Technology. His current research interests include computer vision and machine learning. He has authored and coauthored approximately 20 papers published in top-tier academic journals and conferences. He has served as a reviewer for numerous academic journals and conferences, including TPAMI, IJCV, TIP, TMM, TNNLS, CVPR, ICCV, ECCV, AAAI, and IJCAI. He was the recipient of the Best Paper Diamond Award at IEEE ICME 2017.
[Uncaptioned image] Liang Lin is a full Professor of Sun Yat-sen University. He is the Excellent Young Scientist of the National Natural Science Foundation of China. From 2008 to 2010, he was a Post-Doctoral Fellow at University of California, Los Angeles. From 2014 to 2015, as a senior visiting scholar, he was with The Hong Kong Polytechnic University and The Chinese University of Hong Kong. From 2017 to 2018, he leaded the SenseTime R&D teams to develop cutting-edges and deliverable solutions on computer vision, data analysis and mining, and intelligent robotic systems. He has authorized and co-authorized on more than 100 papers in top-tier academic journals and conferences. He has been serving as an associate editor of IEEE Trans. on Neural Networks and Learning Systems, IEEE Trans. Human-Machine Systems, The Visual Computer and Neurocomputing. He served as Area/Session Chairs for numerous conferences such as ICME, ACCV, ICMR. He was the recipient of Best Paper Nomination Award in ICCV 2019, Best Paper Runners-Up Award in ACM NPAR 2010, Google Faculty Award in 2012, Best Paper Diamond Award in IEEE ICME 2017, and Hong Kong Scholars Award in 2014. He is a Fellow of IET.