Automated
Road Extraction from LiDAR Data
By
Abdullatif Alharthy and James Bethel
It
has been shown that dense LiDAR (Light Detection And Ranging)
data is very suitable for 3D reconstruction of urban features.
In this research, an algorithm has been developed to detect
roads in urban areas using the intensity return and the 3D
information of LiDAR data exclusively. The intensity return of
laser scanning was the initial key to detect candidate road
pixels by filtering out the undesired range of reflectance. As
has been demonstrated in prior research, reflectance or the
spectral response as the only resource for road extraction is
not sufficient in all cases, especially in urban areas, due to
the similarity in reflectance between objects. This similarity
leads to a high rate of misclassification. The resulting
classification from the first step was refined using the 3D
information of the LiDAR data. Then the road centerlines and
edges were obtained using an image matching technique. Extracted
roads from actual dense LiDAR collected over the Purdue campus
are presented in this article.
The
extensive demand for the rapid and high quality acquisition of
3D urban features has motivated much research effort towards
increased automation. Road network database development is
expensive and time consuming since most of the road network
extraction depends heavily on human labor. Creating,
maintaining, and updating the transportation network database is
an essential requirement for many applications such as automated
vehicle navigation, traffic management, safety, disaster
assessment and future planning. Moreover, many critical
decisions such as emergency response, evacuation, and incident
management could be based on such data bases (Xiong, 2001).
With
the availability of many sources of data such as conventional
imagery, Synthetic Aperture Radar (SAR) imaging, Interferometric
Synthetic Aperture Radar Digital Elevation Models (IFSAR DEMs),
and LiDAR DEMs, there are many avenues open to derive terrain
and feature data in urban areas. Through much research, it has
been shown that laser scanning data has the potential to support
3D feature extraction, especially if combined with other types
of data such as 2D GIS ground plans (Maas, 1999; Brenner and
Haala, 1999; Weidner and Förstner, 1995). Despite the fact that
LiDAR data is attractive in terms of cost per high quality data
point, the quantity of the data makes a challenge for storage
and display (Maas and Vosselman, 1999). Acquiring 3D object
descriptions from such data is a challenging problem and many
approaches have been tried to solve it. Several of them have
succeeded with some limitations. The principal idea of this
research is to extract the road network from laser data
exclusively, utilizing both laser intensity returns and height
information.
In
much previous work, road extraction research activities were
more concentrated on extracting roads from imagery in rural
areas than urban areas due to the complexity of urban scenes (Hinz
and Baumgartner, 2000). Moreover, road extraction from aerial
imagery has some limitations due to several factors. Reflectance
dependency on viewing conditions and illumination, variability
of the surface material, occlusion and resolution limitations
are some examples of those factors (Price, 1999).
Recently,
pattern recognition and a hierarchical road model have been
utilized to improve the extraction procedures and consequently
the results. A detailed and scaled road model also was used to
update road maps (Vosselman and Gunst, 1997). This method works
very well in open rural areas but it is sensitive to shadows and
occlusions.
The
road detection procedure described here includes first
segmenting roads from other natural and man-made features such
as trees and buildings. The aim is to develop a fast and
accurate method to detect the road network in urban areas from
LiDAR data. Filtering to identify candidate roads from other
urban features was the first step in this work. Many
segmentation techniques are feasible with LiDAR data such as
direct thresholding determined by histogram analysis, the use of
2D GIS data, and multispectral and hyperspectral inferences
(Maas, 1999; Brunn and Weidner, 1997). Direct thresholding works
by extracting the terrain surface using a filtering technique
such as the morphological filter. The segmentation of roads from
other urban features in this research was based on the resulting
intensity from the laser scanning survey. However, due to some
similarity in the reflections between some features, LiDAR
heights were also necessary to support the segmentation
procedure and to refine it.
In
this work, the intensity was used to segment roads from other
objects in the data set. However, as in many image
interpretation tasks, the result of segmentation is not always
complete and this problem is more obvious in road extraction in
particular. So, more steps were needed to refine this
classification. Besides the intensity analysis, the LiDAR
heights and their inferences was used as a key to discriminate
between ground and non-ground features since roads are known to
be ground features with few exceptions such as in bridges. The
DTM (Digital Terrain Model) was extracted by applying a modified
mathematical morphological filter on the last return heights of
the digital surface model (LDSM). Consequently, non-ground
features were detected using the normalized LDSM (the difference
between LDSM and DTM), and replaced by the corresponding
terrain. Then, this result was utilized to filter out
misclassified pixels from the previous step. This approach was
tested on the data set that has been collected over the Purdue
University campus in spring 2001 with an approximate density of
one data point per each square meter.
Data
Classification Procedure
Laser
point clouds must be segmented in order to distinguish between
building, roads, trees and other urban features. It is a
mandatory process to differentiate among diverse objects in the
scene. Extraneous objects such as trees that do not belong to
the road category should be detected and removed from the scene.
The filtering process, which was used in this project to segment
roads from other objects, consists of two main steps. The first
one is to utilize the return intensity image of the LiDAR data
in order to detect roads based on their surface reflectivity.
But due to similarity in feature reflectance another step was
needed to filter out the undesired features. The second step is
making use of "first minus last" return analysis and
utilizing the 3D LiDAR information inferences to filter out
misclassified pixels that do not belong to road category.
Intensity
Return Analysis
Most
laser scanning systems today have the ability to record the
intensity of the reflected laser pulse. The data that we have
tested was collected using an Optech ALTM1210 LiDAR system,
which was operated by Woolpert Consultants, and it has the
ability to record the laser reflection intensity. This
particular system operates at wavelength of 1.064 µm, i.e., in
the near infrared region of the spectrum. In current ranging
systems, the laser wavelength can occupy any segment of the
optical band between 0.8µm and 1.6µm. However, the selection
of the optical wavelength of the laser depends on the design
aspects of the laser system and the purpose of the survey. The
way to measure the intensity of a reflected pulse is by
quantizing the amplitude when receiving the echo. This technique
is used in most current laser sensors. It produces a noisy image
since it does not reflect the intensity of the whole span of the
received echo, rather only an instantaneous intensity. However,
the intensity image can be used to improve the laser scanning
data classification and filtering. Beside other factors such as
wavelength and flight height, the reflectance varies from one
scanned surface to another based on the surface properties. An
object surface like sand has a reflectance range between 10% and
30%, vegetation between 30% and 50%, while snow and ice can have
up to 80% reflectance (Wever and Lindenberger, 1999). Calm water
surfaces introduce strong dependence on view angle since they
become specular rather than diffuse surfaces. Dark surfaces such
as asphalt or slate roofs absorb the laser beam and consequently
no reflection will be recorded which results in a gap in LiDAR
data.
The
principle of this approach is to start with the segmentation of
the intensity image of the LiDAR data. First, the histogram of
the digital numbers from the quantized intensity over the tested
area was analyzed. It was found that the quantized intensity
values range from zero to over 8000 with a mean of 330 and a
standard deviation of 230. However, more than 95% of the values
range from zero to 800 as shown in Figure 1. Those response
values are based on a scale starting at zero which represents
the weakest return in the data set. By analyzing the data we
found that high intensity (larger than 800) are small in number
and they are mostly positioned in parking lots. After inspecting
the intensity image closely it has been found that those high
intensity spots belong to specular effects on glass and metal
parts of cars, and occasionally, buildings. Those high responses
were then filtered out based on the histogram interpretations of
the desired range and replaced by the maximum acceptable range,
which is 800 in our case. The histogram of the filtered
intensity responses is shown Figure 2.
Since
the raw LiDAR data is in an irregular form, the intensity
returns were interpolated and resampled into a regular structure
in order to view and process the intensity image. The resampling
was done with a spacing of one meter since the irregular spacing
was approximately one meter. Figure 3 shows the filtered and
scaled (0-250) intensity image. The following step was to define
the intensity range that road pavement materials might have in
the data set in order to supervise the classification procedure.
To accomplish that, training samples were chosen based on prior
knowledge of road locations to estimate the reflectivity range
of roads in the tested area. The training samples were chosen
carefully from different locations among the data set (as shown
in Figure 3) to cover the whole range of pavement materials.
Consequently, the decision boundaries were obtained for the road
class. It was obvious from the intensity image (Figure 3) that
road pavement materials have low reflectance. According to the
training samples the intensity values of roads range between 18
and 60. Accordingly, the intensity image was classified into two
classes, roads and non-roads, based on the intensity returns for
each resampled pixel. The classification result is shown in
Figure 4.
However,
due to the similarity in reflectance between objects such as
asphalt road and building roofs, which maybe be coated with
similar materials, there was a significant misclassification
error in this procedure. Also sidewalks and some areas of
vegetation have similar returns to roads. As a result of that, a
refinement step was needed. This step utilizes the LiDAR 3D
information, which is the main part of the LiDAR data, as
discussed in the following section.
Classification
Refinement
The
classification refinement step was designed to address the
misclassified areas and improve the classification accuracy. The
misclassification was mostly a positive error; i.e., areas were
classified as roads while they are not in reality. Most of the
misclassified areas are buildings roofs, sidewalks, driveways,
and grass. Therefore it was mandatory to detect those objects in
order to improve the classification result. In order to detect
grass and other ground vegetation, this step utilizes the
capability of the LiDAR system of being able to measure and
record two returns (first and last) from each pulse (Alharthy
and Bethel, 2002). Also the interpretations of laser heights and
extracted DTM provide good indication of the position of an
object, i.e., whether it is a terrain or above terrain object.
Those two algorithms and the way they were utilized are
discussed in brief below.
First-Last
return Analysis
The
LiDAR system that was used to collect this data was an Optech
ALTM1210 and it has the ability to capture two returns (first
and last) per each height point. This is due to the fact that
the laser pulse is not a single ray but occupies an extended
solid angle. It has an angular beamwidth and its footprint will
a take circular shape when it reaches the ground. Based on both
laser and scene characteristics and physics, the laser beam
could penetrate some objects. Therefore some of its energy will
be reflected back from the object top surface and others might
penetrate to different depths before they are reflected.
Generally, this produces an extended return signal. The first
and last will be recorded and their corresponding heights will
be calculated separately. As a result of that, two heights were
assigned to each data point. The first return heights contain
more noise since they will reflect every object on the ground
such as trees, bushes, cars, etc. On the other hand, the
computed heights based on the last received return represent
only the non-penetrable objects such as the ground, and
buildings. Two different heights of one point give an indication
of the presence of a penetrable object such as trees and
vegetation or possibly "straddling" a multi-height
object. On the other hand, if a data point has the same height
for first and last returns, then this point belongs to a
non-penetrable 'solid' object. This step is just to locate the
penetrable objects, which are categorically considered non-road
regions.
A
high level of discrepancy between first and last return heights
takes place only on penetrable objects, which is considered as
noise, as stated above, according to the objective of this work.
The discrepancy map was then employed to improve the
classification. If a pixel among those that were classified as
road from the first step has a discrepancy larger than a certain
threshold, this pixel will be considered as a penetrable object
and consequently a non-road pixel. Accordingly, it will be
filtered out. The result of this refinement step is shown in
Figure 5. We can see how the trees on roadsides and grass in
open areas are filtered out. During this step the classification
was improved as shown in Table 1. However, there are some parts
of the misclassified areas still not filtered out. Those parts
are mostly buildings roofs that are coated with materials
similar to road pavements materials.
3D
Information and Terrain
Non-terrain Point Analysis
As
stated above, some of the roof buildings were covered with
similar materials to those that were used in road pavement. This
similarity causes both surfaces to have comparable intensity
returns when they are illuminated with the laser pulse.
Consequently such roofs were classified as roads during the
intensity segmentation procedure. Moreover, they are not
filtered out in the first-last return analysis since roofs are
non-penetrable objects. Accordingly, in order to reclassify
those roofs "non-ground objects," they need to be
detected first. The direct route to distinguish between terrain
and above terrain objects is to extract the terrain model first.
Lindenberger (1993) introduces and shows how mathematical
morphology can be used for filtering data recorded by a laser
profiler and consequently for DTM generation. (Vosselman 2000)
introduces an approach called slope based filtering in which he
established a modified version of the mathematical morphology.
In this work, a similar mathematical morphology filter in an
autoregressive mode was used to extract the DTM as shown in
equation 1. I represents a pixel in the given intensity image, G
represents the computed gradient at that pixel using the last
return heights, DSM is the raw resampled digital surface model,
DTM is the extracted digital terrain model which represents the
bare ground, and tG is the maximum allowed gradient and can be
computed based on surface nature and scene characteristics.
Then, each classified road pixel will be tested to determine
weather it belongs to the bare ground model (DTM) or not by
comparing its last LiDAR height with the extracted DTM.
Accordingly, if it is not a ground pixel then it will be
reclassified as a non-road one. The result of this refinement
step was significant as shown in Figure 6 and Table 1.
DTM
= {I (i,j) ? DSM ? ;I (i,j) :G(i,j) < tG}
Connected
Component and
Final Result
Connectivity
between pixels is a useful concept used in obtaining homogenous
regions in an image. So, to decide that two pixels belong to one
region they should be adjacent and homogenous (have similar gray
level for example). This concept was utilized here to establish
road regions and consequently construct the road network. Also,
the size of those regions was used to filter out small regions
like side walks and other scattered noise. The result of this
step is shown in Figure 7. Table 1 shows the summary of this
procedure's performance.
Discussion
The
aim in this work is to design a simple and fast method to detect
road networks in urban areas using LiDAR data exclusively. We
restricted the procedure to not require any other source of data
other than LiDAR. This was done intentionally to avoid the
limitation of availability of other sources of information in
some areas. Sources such as ground plans, imagery and
multispectral data are not available for every desired site and
might be not up to date. This method starts with filtering the
raw LiDAR data to remove "noise" unrelated to the road
class. Refinement steps followed to improve the heuristic
classification of the data. The performance of those steps is
good, however, accuracy of the extracted road network needs to
be investigated more. The presented algorithm works very well
with the tested data and is expected to do the same with others.
Yet, it might fail in some other areas where road materials vary
within small areas and consequently their reflection will cover
a wide range of the intensity scale. This surely will affect the
classification procedure and increase the misclassification
error. Practically, in general, road detection algorithms are
able to work only with a limited set of characteristics. And
those characteristics can be modified based on the data nature
and specification. When these characteristics change beyond
expected limits the algorithm may fail.
The
quality of the result depends on both the algorithm and the
data. In general, LiDAR data is not continuous and it does not
represent the entire scene since there are some gaps between
pulse footprints on the ground. Thus we need to treat the data
as sampling of a continuous surface with an aperture
corresponding to the ground footprint of the laser. With a
sampling interval of one meter, ideally to avoid aliasing, the
aperture would also be one meter. In this case, the aperture
(laser footprint) was about 30cm so the data was undersampled to
that degree.
Therefore,
as shown in Figure 7, the resulting road network is not totally
complete due to some classification errors as stated above.
However, these results seem to be promising compared with other
data sources and approaches. These results are preliminary and
more development of the algorithm needs to be performed.
Acknowledgments
The
author would like to acknowledge the support for this research
from the following organizations: the Army Research Office and
the Topographic Engineering Center. Also, the author would like
to acknowledge the helpful discussion from Dr. Edward Mikhail
and Dr. James Bethel of the Geomatics Area, School of Civil
Engineering, Purdue University.
About
the Authors
Abdullatif
Alharthy is a Ph.D
Candidate in the Geomatics Engineering School of Civil
Engineering at Purdue University.
James
Bethel is Associate Professor, Purdue University School of Civil
Engineering.
References
Ackermann,
F. (1999). Airborne laser scanning-present status and future
expectations. ISPRS Journal of Photogrammetry & Remote
Sensing, 54:64-67.
Alharthy,
A., Bethel, J. (2002). Heuristic filtering and 3d feature
extraction from LiDAR data. ISPRS Commission III, September
9-13, 2002, Graz, Austria.
Brenner,
C. (2000). Towards fully automated generation of city models.
ISPRS, vol. XXXIII, Amsterdam 2000.
Brenner,
C., Haala, N. (1999). Extracting of building and trees in urban
environments. ISPRS Journal of photogrammetry & Remote
Sensing, 54:130-137.
Brenner,
C., Haala, N. (1999). Rapid production of virtual reality city
models. Geoinformationssysteme, (2):22-28.
Brunn,
A. (2001). Statistical interpretation of dem and image data for
building extraction. In A. Grün, editor, ASCONA-Workshop 2001.
Balkema-Verlag.
Brunn,
A., Weidner, U. (1997). Extracting buildings from digital
surface models. IAPRS Vol. 32, part 3-4w2, Stuttgart.
Hinz,
S., Baumgartner, A. (2000). Road Extraction in Urban Areas
supported by Context Objects. In: International Archives of
Photogrammetry and Remote Sensing, Vol. 33, part B3.
Hug,
Ch., Wehr, A. (1997). Detecting and identifying topographic
objects in imaging laser altimetry data. IAPRS 32, 19-26, part
3-2w3.
Jelalian,
A. (1992). Laser Radar System, Artech House, Boston.
Lindenberger,
J., (1993). Laser-Profilmessungen zur topographischen Geländeaufnahme.
Deutsche Geodätische Kommission, Series C, No. 400, Munich.
Maas,
G-H. (1999). The potential of height texture measures for the
segmentation of airborne laserscanner data. Presented at the 4th
Airborne Remote Sensing Conference and Exhibition, Ottawa,
Ontario, Canada, 21-24 June 1999.
Maas,
H., Vosselman, G. (1999). Two algorithms for extracting building
model from raw laser altimetry data. ISPRS Journal of
Photogrammetry and Remote Sensing, 54(2-3):153-163.
Mikhail,
Edward M., Ackermann, F. (1976). Observation and Least Squares.
Price,
K., (1999). Road grid extraction and verification. In:
International Archives of Photogrammetry and Remote Sensing,
Vol. 32, part 3-2W5, pp. 101-106.
Song,
J., Han, S., Yu, K., Kim, Y., (2002). Assessing the Possibility
of Land-Cover Classification Using LiDAR Intensity Data. ISPRS
Commission III, September 9-13, 2002, Graz, Austria.
Uwe
Weidner, Wolfgang Förstner (1995). Towards Automatic Building
Reconstruction from High Resolution Digital Elevation Models.
ISPRS-Journal, 50(4), pp.38-49, 1995.
Vosselman,
G., Gunst, M., (1997). Updating road maps by contextual
reasoning. Proceedings 2nd Workshop on Automatic Extraction of
Man-Made Objects from Aerial and Space Images, Ascona, May 5-9.
Vosselman,
G., (2000). Slope based filtering of laser altimetry data. IAPRS,
Vol XXIII, Part B3, pages 935-942. Amsterdam, The Netherlands.
Xiong,
Demin. (2001). Automated road network extraction from high
resolution images. Technical notes, National consortium for
safety, issue 3, may 2001.
Wever,
C. and Lindenberger, J., (1999). Experiences of 10 years laser
scanning, Photogrammetry Week, pp. 125-132.
|