Download as pdf or txt
Download as pdf or txt
You are on page 1of 11

322 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 9, NO.

2, JUNE 2008

Real-Time Speed Sign Detection Using the


Radial Symmetry Detector
Nick Barnes, Member, IEEE, Alexander Zelinsky, Fellow, IEEE, and
Luke S. Fletcher, Member, IEEE

Abstract—Algorithms for classifying road signs have a high


computational cost per pixel processed. A detection stage that
has a lower computational cost can facilitate real-time process-
ing. Various authors have used shape and color-based detectors.
Shape-based detectors have an advantage under variable lighting
conditions and sign deterioration that, although the apparent color
may change, the shape is preserved. In this paper, we present the
radial symmetry detector for detecting speed signs. We evaluate
the detector itself in a system that is mounted within a road vehicle.
We also evaluate its performance that is integrated with classi-
fication over a series of sequences from roads around Canberra
and demonstrate it while running online in our road vehicle. We
show that it can detect signs with high reliability in real time. We
examine the internal parameters of the algorithm to adapt it to
road sign detection. We demonstrate the stability of the system
under the variation of these parameters and show computational
speed gains through their tuning. The detector is demonstrated to
Fig. 1. ANU/NICTA/CSIRO intelligent vehicle. The cameras are mounted
work under a wide variety of visual conditions. where the rearview mirror would be.
Index Terms—Detection, fast radial symmetry, real time,
road sign.

limit [1]. There are many possible reasons that a driver may
I. I NTRODUCTION be speeding; one of these is failure to notice a speed sign.
In laboratory experiments, it has been shown that distractions
I MPROVING safety is a key goal in road vehicle technology.
Driver support systems aim to improve safety by helping
drivers react to changing road conditions. Although the full
result in a twofold increase in the failure to detect simulated
traffic signals [2].
automation of road vehicles remains a motivating goal, our Technology can assist by alerting drivers to speed limit
research focuses on systems that can immediately assist drivers. changes to which they do not appear to have reacted and
Rather than replacing the drivers, we aim to keep the drivers by maintaining a monitor of the current speed limit. Road
in the loop while supporting them in controlling the car. This sign recognition also has other possible applications, such as
is the approach that has been adopted by driver assistance building and maintaining maps of signs within an urban envi-
systems (DASs). ronment and automated inspection of signs for deterioration.
The risk of a crash causing death or injury rapidly increases The primary focus of the research program at the National
with increases in the speed over an appropriately set speed ICT Australia (NICTA) on sign detection is on road safety,
particularly supporting drivers in their tasks.
Presently, available techniques for the classification of can-
Manuscript received April 10, 2007; revised August 5, 2007, October 28,
didate signs show strong performance. The major shortcoming
2008, December 11, 2007, and December 24, 2007. The National ICT Australia of many systems is with high-speed detection, i.e., quickly
is funded by the Australian Government’s Backing Australia’s Ability initiative identifying likely candidates for classification in a way that
in part through the Australian Research Council. The Associate Editor for this
paper was S. Nedevschi.
is robust to the vast range of lighting conditions that occur
N. Barnes is with the National ICT Australia, Canberra, A.C.T. 2600, in road scenes. In this paper, we present the application of a
Australia, and also with the Department of Information Engineering, The shape detection algorithm, i.e., the radial symmetry detector, to
Australian National University, Canberra, A.C.T. 0200, Australia (e-mail:
nick.barnes@nicta.com.au).
speed sign detection. We present its integration into a system
A. Zelinsky is with the Commonwealth Scientific and Industrial Re- for speed sign recognition, showing real-time performance and
search Organization ICT Centre, Canberra, A.C.T. 0200, Australia (e-mail: robustness to a broad array of lighting conditions over image
alex.zelinsky@csiro.au).
L. S. Fletcher is with the Computer Science and Artificial Intelligence Labo- sequences. The detector system is also demonstrated running
ratory, Massachusetts Institute of Technology, Cambridge, MA 02139 USA real time in on-road trials at the Australian National Univer-
(e-mail: lukesf@mit.edu). sity (ANU)/NICTA/Commonwealth Scientific and Industrial
Color versions of one or more of the figures in this paper are available online
at http://ieeexplore.ieee.org. Research Organization (CSIRO) intelligent vehicle, as shown
Digital Object Identifier 10.1109/TITS.2008.922935 in Fig. 1.
1524-9050/$25.00 © 2008 IEEE
BARNES et al.: REAL-TIME SPEED SIGN DETECTION USING THE RADIAL SYMMETRY DETECTOR 323

II. B ACKGROUND GPS or a gyro and a speedometer may be useful to note things
such as T-intersections. In this paper, we focus on vision-based
Speed sign regulations vary worldwide. We consider the
detection.
Australian regulatory framework that offers a typical exam-
ple of conditions.1 Speed signs are one of several ways in
B. Vision-Based Sign Recognition Systems
which the current speed limit may be specified. A single type
of regulation sign indicates a change in the speed limit; the Research into vision-based road sign recognition began in
signs are rectangular with a red circle and black numbers or earnest in the mid-1980s. Brute force approaches may be
illuminated white numbers indicating the prescribed speed. used where classifiers, such as normalized correlation, are
A speed zone can end with another speed sign, a dead end, a applied at every image pixel. Such an approach is computa-
T-intersection, or a derestriction sign. Derestriction signs also tionally prohibitive; however, this can be eased somewhat by
have a circle on a rectangular background. Furthermore, some approaches such as simulated annealing [4] or coarse-to-fine
areas have geographic limits, such as a limit of 50 km/h on techniques in applying a hierarchy of templates [5]. However,
urban local streets, unless a sign indicates otherwise. However, these methods are still computationally expensive. To manage
noting a reduction of speed limit is clearly a critical aspect for computation while still maintaining classification quality, many
any complete system and would form a useful function on its approaches have introduced separate stages for the detection
own. For the purposes of this detector, Australian signs are and the classification of the sign type (e.g., [6]–[8]). A low-
largely consistent with many international road signs such as expense algorithm that produces few false negatives is used, and
the European Union, and therefore, the technology that has then, only a small fraction of the incoming image stream needs
been developed would be applicable more broadly. to be explored with more expensive and accurate classification
algorithms.

A. Sign Recognition Approaches C. Detection Techniques


Generally, proposals for automatically informing the driver Color-based segmentation is commonly used for detection
of the current speed limit are based on three major either to identify likely signs or eliminate regions that are
technologies—the GPS, the RF transmitters, and vision. The unlikely to contain signs. The standard assumption behind these
road angel system (http://www.roadangel.co.uk/) employs the techniques is that the wavelength arriving at the camera from
GPS and a map of speed zones. Installing an RF transmitter on a sign is invariant to the intensity of the incident light and
every speed sign to broadcast the speed to in-car receivers has that the transformation into an image preserves this relation.
been proposed. Other sensors that may be usefully employed This assumption is typically stated in the form where the hue,
are the speedometer and inertial sensing. saturation, and value or hue, saturation, and intensity (HSI)
The GPS and the RF offer viable solutions to suburban spaces or the ratios over the red, green, and blue color space are
geography-based speed limits. However, difficulties arise with invariant to lighting conditions [8], [9]. Much of the color-based
frequent updates to speed signs, including regulatory changes, detection research literature exploits this assumption (e.g., [6]
temporary signs such as roadwork, and automatically changing and [8]–[14]). Researchers have also attempted more complex
speed signs such as some bridges with high winds. To track all analysis of hue and saturation [15] and other schemes such as
these changes online, a database must be maintained, and this CIECAM97 [16], which is a color appearance standard based
information somehow has to be made available to the driver in on human perception. Systems often use the color first and then
real time for all signs. For the RF, installing the receiver in a combine other methods, such as Estevez and Kehtarnavaz [17],
vehicle is low cost and simple; however, it is unclear who bears who developed real-time implementation to recognize various
the cost for a large number of transmitters. Government funding traffic signs by combining the color, the edges, and a local
may be politically difficult, and it is unlikely that any single histogram.
industry player would pay such an infrastructure cost. Other Alternatively, a priori assumptions about the scene structure
problems must also be overcome: Single sign posts on narrow are often used in association with color-based segmentation.
roads indicate different speeds to different directions of traffic, A simple example is to assume that the road is approximately
and on multilane highways, signs may be far from the vehicle, straight, so that, using color segmentation, much of the image
whereas exit ramps may have a sign just off the highway. can be ignored [13]. More complex approaches use some de-
Thus, although we consider that the GPS and the RF trans- tection to facilitate scene understanding and eliminate regions
mitters may be plausible parts of an automated sign detection where signs are unlikely to appear, e.g., detecting large uniform
system, vision-based sign recognition is a necessary compo- regions corresponding to the road and the sky and assuming that
nent. Certainly, combining the GPS and/or the RF transmitters signs are only likely to appear in the region alongside the road
with visual sign recognition into a single system could improve and below the sky [10].
the performance over any individual component. Lindner et al. Some work has tackled shape-based detection, often in as-
[3] demonstrated a very low rate of false positives by combining sociation with color. De la Escalera et al. [9] applied a corner
visual traffic signal recognition with the GPS. Furthermore, the detector with the color to find signs. However, they only used a
few points at the border of the circle, rather than all pixels, and
therefore, the method is prone to problems of occlusion. The
1 See http://www.rta.nsw.gov.au/rulesregulations/downloads/pts1-21.pdf. same group performed detection by combining the color and
324 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 9, NO. 2, JUNE 2008

restriction is difficult. Furthermore, a simple road scene under-


standing, such as a single-colored sky and a road, is inadequate
on road scenes where trees overhead break up the sky and
cast strong shadows on the road [see Fig. 14(a)–(c)]. These
exceptions cannot be ignored, as the false-negative rate for a
commercial sign recognition system must be very low, and slow
processing in difficult situations is not acceptable. By detecting
and modeling the road [22], it may be possible to dictate parts
of the image where a sign cannot appear; however, robust road
modeling has its own computational expense.
Fig. 2. (a) Systems that are overly dependent on image structure may miss
one of the two signs here, missing a chance for correct classification. (b) Radial However, the roadway is well structured. Signs are placed to
symmetry detector correctly detected this sign in heavy rain. be easily visible so that the driver can easily see them without
having to look away from the road. Unless damaged, critical
signs, including speed signs, give way signs, and stop signs,
have a highly regulated appearance; they appear upright and
approximately orthogonal to the road. Even on rapidly curving
roads, there will be a short period where the sign appears
parallel to the image plane, and this will be when the vehicle
is close to the sign. Thus, the problem is solvable through a
system that can recognize a sign from a small number of frames
and processes frames sufficiently quickly.

Fig. 3. Typical detected candidate sign (a) at input image size and (b) close D. Sign Recognition
up. The circle does not appear as a single color, the edges fall within pixels, and
the right edge touches the edge of the sign. As early as 1996, correct classification rate percentages of
more than 90% were reported using normalized cross cor-
the shape using edges [18] in an interesting method that applies relation [10]. More recently, Escalera and Radeva [18] have
a deformable model using genetic algorithms and simulated applied the normalized k nearest neighbor classification after
annealing. Real-time detection performance was not possible initial detection using Adaboost and orientation correction
at the time of publication. Recent work has combined the using algorithms such as the radial symmetry. They have re-
color with support vector machines (SVMs) for shape-based ported 98% correct classification. Another approach [23] also
detection as well as SVM-based classification [19]. The system applied Adaboost for detection but used Bayesian generative
is able to classify a broad array of signs at 1.7 frames/s using modeling for classification, reporting error rates of 6% over a
large images with high accuracy. Early work by Besserer et al. large data set. SVMs have shown impressive results for sign
[20] used evidential reasoning for shape-based detection of classification, with the correct class selected out of 35 general
traffic signs, finding circles, triangles, and polygons from sign types in over 97% of the cases and substantially improving
chain codes. for smaller sets of classes [24]. Recent work by Paclik et al.
The complexity of sign detection. Lighting conditions change [25] improves the performance over standard cross correlation
between direct sunlight, headlights at night, heavy cloud cover, with a trainable similarity measure. By incorporating the most
and smog, and a sign may be in complete shadow on a bright discriminative regions, the classifier is not sensitive to uninfor-
day.2 Such changes have an impact on lighting chromaticity mative pixels such as occlusions. Computational demands are
and, therefore, on the wavelength of the reflected light. Also, also reduced.
heavy rain may blur the sign appearance [see Fig. 2(b)]. Fur- Although commercial systems will require results of close to
thermore, for most cameras, the varying incident intensity and 100% recognition with few false positives, with the incorpora-
wavelength produce both nonlinear and interdependent effects tion of additional cues and further research, viable commercial
on the HSI in the resultant image [21]. Ideally, signs have clear classification may be possible in the near future. For example,
color contrast; however, over time, they may fade but are still the application of superresolution over multiple images [26] can
clear to drivers [see Fig. 14(d)]. improve classification performance. Alternatively, better classi-
Last, compared to ideogram-based signs, speed signs only fication may be achieved by the fusion of multiple classifiers,
have narrow bands of color, which are a few pixels wide unless such as Adaboost, if their problems are complementary.
some form of high-resolution active vision is employed. If more In general, sign classification is closer to commercial appli-
than one color falls on a charge-coupled device element, the cation than detection. Thus, in this paper, we focus on detection.
resulting color is not easily predicted (e.g., Fig. 3). We combine our detection algorithm with basic classification
In practice, as signs may appear in many image positions based on normalized cross correlation. Our aim here is not to
on highways, or over uneven and winding roads, a priori propose an ideal method for classification but to demonstrate
that the detection method can be effectively combined with
2 See Fig. 14 and the supporting video at http://cecs.anu.edu.au/~nmb/ classification to produce an effective complete sign-recognition
speedsigns/. system.
BARNES et al.: REAL-TIME SPEED SIGN DETECTION USING THE RADIAL SYMMETRY DETECTOR 325

E. Driver Feedback
Once a sign is detected, this information must be presented to
the driver. The most passive approach that is likely to have high
acceptance is to add a simple readout of the current speed limit,
perhaps by dynamically changing the speedometer to show the
current speed limit. However, a driver may also wish to be
alerted in the case that he or she has failed to notice the speed
sign. Speed monitors that sound an alert when a driver-specified
Fig. 4. Radial symmetry voting. (a) For a given edge point as shown, the
warning limit is exceeded are now common features in vehicles. centroids of a 1-D family of circles lie along the line that is orthogonal to
These could be set based on the currently detected speed, along the direction of the edge point. Note that if we have no prior expectation
with a driver-specified tolerance if they so choose. Alerting the of light–dark versus dark–light transitions on the circle, the family of circles
extends in both directions from the edge point. (b) From a continuous circle,
driver to a sign only when an adjustment is required, and he or every edge point votes along a line of possible radii; these lines will all intersect
she has not seen it, is easily implemented within a DAS, such at the center of the circle, resulting in a peak. A number of vote lines are
as monitoring if the driver has looked at the sign [26]. Heads- shown here.
up displays, e.g., from Volkswagen, could be used to draw the
driver’s attention when required.
point. The fast radial symmetry detector eliminates factor b by
directly taking the gradient of the edge point from the output
III. A PPROACH of the Sobel edge detector. Instead of voting in all directions, it
only votes in the direction that is orthogonal to the edge, which
We examine an efficient method for speed sign detection—
is where the center must lie. This way, the computation of the
the fast radial symmetry detector [27]. Many shape detectors
radial symmetry detector is reduced by a factor of b, and the
are nonrobust because they require closed shapes. Robust tech-
resulting circle map is simplified by a dimension. This makes it
niques such as Hough circle detection [28] are slow to compute
suitable for real-time use.
over large images. Fast radial symmetry can be run as a detector
The operation of the fast radial symmetry detector can be
at the frame rate. All Australian speed signs have a red circle
understood simply by considering all the possible circles of
with a contrasting background. We are able to eliminate the
which an image edge pixel could be part. If we have the
vast majority of false positives by considering only radially
direction of the edge pixel, then a family of circles lies along a
symmetric regions that are stable across several images and
line in the direction that is orthogonal to the edge, with varying
have a high count of pixels in proportion to the radius. This
radius [see Fig. 4(a)]. A single edge point such as this would
reduces the number of locations where signs can possibly occur.
vote for all circle centroids along this line.
Fast-radial-symmetry-based detection is compatible with all
In considering an edge image, for each edge pixel, we form
major forms of classification, as it identifies the center of the
a set of radial symmetry transform images for different radii.
sign according to maximum-likelihood criteria. The maximum-
Thus, when many edge points align around the boundary of
likelihood formulation follows from the maximum-likelihood
a circle, each of them will vote along their orthogonal lines.
basis of the regular polygon algorithm (see [29]). With the
This set of votes will peak at the circle center. Fig. 4(b) shows
center of the sign identified with the center of the circle,
sample vote lines that would emerge from a continuous set of
recognition only needs to be applied to a small number of
edge points corresponding to the circle shown. All of these lines
pixels. To demonstrate the compatibility of fast radial symmetry
intersect at the centroid, leading to a peak in the vote space.
detection with classification, we apply normalized cross cor-
In practice, this is performed on a discrete image, and the
relation to detected candidates. Typically, for recognition after
radius space is sampled with separate vote spaces for each
detection, an unknown scale requires further processing or, for
radius to simplify the interpretation. An image and samples of
cross correlation, multiple templates at all possible resolutions.
its radial symmetric vote space are shown in Fig. 5. The sign
However, from the radius that is returned by the fast radial
vote peaks in Fig. 5(c). It appears blurred in Fig. 5(b) due to
symmetry detector, we know the approximate scale of the
the detector radius being too small and in Fig. 5(d) due to the
candidate.
detector radius being too large.
For the sake of self-contained presentation, we include a
A. Sign Detection: Fast Radial Symmetry formal description of the detector largely taken from [27]. For a
given pixel p, the gradient g is calculated using an edge operator
The fast radial symmetry detector [27] is a variant on the that yields orientation. If this pixel lies on the arc of a circle,
circular Hough transform that executes in the order kp, where then its center would be in the direction of the gradient at the
p is the number of pixels, and k is the number of discrete radii size of the radius. To achieve lighting invariance, we apply the
that are searched. This is as opposed to the traditional circular discrete form of the detector. The location of a pixel that will
Hough transform that executes in the order kbp. For the circular gain a vote as a potential center is defined as follows:
Hough transform, each edge pixel votes on all circles over a
discrete set of radii k that could pass through that edge pixel.  
Factor b comes from the discretization into a number of bins g(p)
p+ve = p + round n (1)
on the angle of circular tangents that could pass through this g(p)
326 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 9, NO. 2, JUNE 2008

Fig. 6. Templates that are used for matching. From the largest shown [(a) and
(c)], we scale all sizes down to (b) and (d).

within the capture zone for the centroid. Thus, the sign should
have its full vote. As all edge points vote, the degradation is
graceful with larger angles.

B. Sign Classification
To show the efficacy of a detector, we must demonstrate that
it can be effectively combined with classification. We reviewed
the effective methods for classification above; however, for the
purposes of this paper, we combined the detector with simple
Fig. 5. Image including a road sign and sample radial symmetry images for standard classification. We applied normalized cross correlation
the three largest radii. Detection peaks in (c).
from the Intel Integrated Performance Primitives library. Cross
correlation is run based on the templates that come from the text
where n ∈ N is the radius, and N is the set of possible radii. of a single sign from close by. We scaled this down by linear
In application to sign detection, this is defined by expectations interpolation across all color channels to all expected radii N
about the apparent sign size. (A negative image is similarly to form a total of eight templates for each sign. Prescaling of
defined, facilitating constraining the operator to find only dark templates saves online computation. The base templates used
light circles or only light dark circles.) A vote image is incre- for 40 and 60 signs are shown in Fig. 6. For each candidate
mented for each pixel based on the orientation votes, i.e., found, we are able to run cross correlation on only a 5 × 5
pixel window around the center of the candidate. The algorithm
On (p+ve ) = On (p+ve ) + 1. (2) provides a good estimate of the radius, and therefore, we are
able to apply a narrow band of sizes of templates. Note that the
In our implementation, the vote image is defined as follows: radius is only as precise as the sampling of the radial symmetry
transform in the radius dimension. Furthermore, the circle on
Õn (p) the sign has a finite width. The radius returned is the sample
F̂n (p) = (3)
kn that has the most support. As the sign is assumed to be vertical
and parallel to the image plane, no scaling or rotation of the
where kn is a scaling factor that normalizes On across different template is necessary. The detector itself, however, is unaffected
radii. We do not use the other parameters from the original by text orientation within the plane; therefore, recognition could
paper, as we are only interested in the maxima of the radial be extended to recognize damaged signs that are rotated in the
symmetry. image plane. Given the simplified nature of the recognition
To obtain the radial symmetry image, F̂n is convolved with a problem as described, it takes less than 1 ms to classify a
Gaussian function. There are several images that are produced single candidate as 40 or 60. Note that this classification can
by the transform. Each radius of N votes into a separate image. be performed on a separate processor and does not have to
There is also a full transform image that averages all the be synchronized with detection; overall, it must keep up to
symmetry contributions over all the radii considered, i.e., minimize the delays in the display to the driver.

1 
S= Sn . (4) IV. A DAPTATION OF THE R ADIAL S YMMETRY T RANSFORM
|N |
n∈N
TO S IGN D ETECTION

See [27] for full details. Detecting speed signs using the radial symmetry transform
Sensitivity to sign alignment with the image plane. The assumes that the sign is close to parallel to the image plane.
radial symmetry detector is derived for circles; however, in our However, road signs are designed to be visible to the driver. On
discrete implementation, we accept votes for a single circle over a straight road, the sign will always be oriented perpendicular
a capture zone of four pixels. Thus, a vote from an edge point to the motion of the car. On highly curved roads, motion will be
must miss the centroid by at least 1.5 pixels to not count. The slower, and the sign will appear perpendicular to the car motion
largest radius that is applied in this paper is 13. Thus, if we turn when it is most visible to the driver. This means that when the
a sign off the image plane by up to 25◦ in either direction, then, sign is most easily detectable, unless it is damaged, the circle
without noise, the votes of the outer edge pixels should still fall will appear approximately as a circle in the image.
BARNES et al.: REAL-TIME SPEED SIGN DETECTION USING THE RADIAL SYMMETRY DETECTOR 327

To detect speed signs, we search the full transform image


[see (4)] for local maxima. This will result in some false
positives, many of which can be quickly eliminated. For each
of the maxima, we search all individual radius images [see (3)]
to find the actual candidates, as a high vote in (4) often results Fig. 7. x and y 3 × 3 Sobel masks.
from a collection of accidental votes from several contributing
radii.
Accidental alignment of objects or image noise may result
in false positives; we found that most instances only last a
single frame. At the same time, on corners, signs may only
appear for a few frames. We found that simple matching over
Fig. 8. Horizontal edge without noise and with 30% noise.
a small number of frames was sufficient to eliminate many
false positives. The sign must not move far (we used ten pixels
per frame) nor have large changes in the estimated radius.
When integrated with classification, the classification result
must be consistent. If the vehicle velocity and the steering angle
were available, tracking using Bayesian filters would reduce
false positives. However, this would increase complexity by
adding other components. In the absence of vehicle dynamic
information, methods such as Kalman filtering are not well
suited to tracking for very short sequences.

A. Radial Symmetry Detection Parameters for Sign Detection


We found the fast radial symmetry algorithm to be highly
suitable and robust for road sign detection. Good results can
be directly achieved with this algorithm; however, to reduce
the computational demands, we identified and tuned some
parameters. This analysis is largely independent of the platform
(the vehicle and the camera) and scales in the obvious way with
the camera resolution. The parameters are as follows:
1) the minimum radius considered; Fig. 9. Four images showing speed signs. (a)–(c) Same road scene from a
2) the maximum radius considered; range of distances, with the sign appearing at larger radii. (d) Different road
3) the increments between radii; scene where it is raining heavily, and visual conditions vary quite substantially
from the other images.
4) the gradient threshold for voting;
5) the number of votes required to be a candidate or the
number of candidates taken for each image; by eliminating some, the computation can be greatly reduced
6) the number of frames we require a candidate to be present with a little impact on performance.
before we declare it to be valid.
The minimum and maximum radii are the embodied param-
B. Sensitivity With Respect to the Gradient Threshold
eters of the system [30] and are set by constraints of sign
appearance. Very small circles are ignored, as the signs are too Consider the 3 × 3 Sobel masks (Fig. 7) convolved with a
far to be recognized. Furthermore, in normal driving conditions, local area that has noise (specifically a straight line in x), where
a sign will not appear very close to the camera; therefore, for a the third column is all increased by 30% (Fig. 8). Without noise,
known focal length, a maximum can be set. the total y gradient is 36k, and the total x gradient is zero.
A simple empirical exercise minimizes the number of radii to With noise, the y gradient is 39.7k, and the x gradient is 9.3k,
be detected. From in-vehicle sequences, we sample the radius i.e., shifted by 13◦ . Thus, at radius 9 (a typical radius for sign
space densely enough that at least one radius strongly responds voting), the error is greater than two pixels and, therefore, is
for every visible sign. As sign borders have a finite width, unlikely to have a great impact on the sign vote.
and we take the dark and light votes from (1), the voting In practice, most systems take the edge magnitude as the
naturally blurs in radius space, allowing a sparse sample. This threshold. Radial symmetry has different demands, however, as
also enables elliptical features to be detected if signs are not we are considering direction and voting over a distance. Thus,
quite frontoparallel. given a constant amount of noise in an image, we can expect
To contain the number of false positives to be classified, that regions of a low-intensity gradient will not lead to accurate
we only consider a fixed number of peaks from each radius votes in radial-symmetry-based algorithms.
per image. We examine the relation between this parameter Fig. 9 shows images with different sized signs and visibility
and the number of frames. Last, if we take all points with a conditions. For these images, we consider the pixels that vote
nonzero gradient, we will not miss additional signs; however, for the signs shown and their gradient magnitude. Fig. 10(a)
328 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 9, NO. 2, JUNE 2008

Fig. 11. Radial symmetry in the rain [see Fig. 9(d)]. (a) Detector vote image
with a threshold 10. (b) Detector image with a threshold 11 299.

the threshold applied. The signal-to-noise ratio is substantially


improved, with little loss of votes for the sign. In practice, this
Fig. 10. Cumulative histogram of the gradient magnitude of (a) pixels that fell
within two pixels of a sign center for all the images of Fig. 9 and (b) all pixels
yields a decrease of a factor of four to five in the execution time
(excluding invalid edges at the image boundary). of voting, which is the largest component of the computation.
A dynamic threshold can be efficiently implemented using
the “getkth element” algorithm [31] (order n average). Further-
shows a graph of the number of pixels that would vote for the more, with slow changes in illumination, it is adequate to adjust
signs versus the log of squared magnitude of the pixel gradient the threshold in the background every few seconds. See [32] for
summed across all four images. This graph shows that only a further empirical analysis.
small percentage of pixels actually have a low-gradient thresh- We tested this threshold over many images with highly
old. Particularly, the curve flattens out at about 9 or, actually, shadowed scenes and degraded signs (e.g., Fig. 14). Of these,
11 299 in squared magnitude. Setting the threshold on the there were no cases where having a low-gradient threshold
gradient squared magnitude to this value would remove 28% would have increased the ranking of a candidate and only very
of sign votes. However, many of these pixel votes are actually few where a threshold of less than 80% would do so. We
erroneous and have fallen close to the center by accident. The conclude from the trials over many images that there is little
surrounding structure also contributes; for example, the sign to be gained by considering low-contrast edges.
circle falls close to the left-hand edge of the sign in Fig. 9(b).
This edge has low contrast and votes to the center but is not
V. E MPIRICAL A SSESSMENT
actually part of the circle. This 28% appears to include much of
the noise. We empirically assessed detector performance, including its
Fig. 10(b) shows a cumulative histogram of the gradient parameters and its incorporation into a system for sign recog-
magnitude (greater than zero) for all pixels over all images nition. All equipment was on board the ANU/NICTA/CSIRO
of Fig. 9. Eighty percent of pixels lie below 9—where we intelligent vehicle (see Fig. 1) with cameras of approximately
propose the threshold; much of the signal is in the top 20% 380-pixel focal length at approximately the position of the
gradient magnitudes. Fig. 11 shows an image comparison with rearview mirror, with a horizontal field of view of around 45◦ .
BARNES et al.: REAL-TIME SPEED SIGN DETECTION USING THE RADIAL SYMMETRY DETECTOR 329

Fig. 12. Images from the second sequence taken around the university.

Normal sized signs were detectable from around 20 m away,


whereas highway signs are about 50% larger and, therefore,
were detectable from around 30 m. Our assessment includes
the system of detector plus classifier running online trials in the
vehicle and many recorded image sequences from the camera.
Two videos of this assessment appear on the accompanying
Web site.3 All sequences that are used in the experiments were
taken from signs on public roads around Canberra, including
in the campus at The Australian National University. Some
sequences were taken while driving at approximately the speed
limit, whereas others were taken during slower driving. As can
be seen from the sample signs that are shown throughout this
paper, there is a great variation in sign appearance. The vari-
ation includes scale, lighting, deterioration due to weathering,
and the weather conditions under which they were taken. Night
performance is also demonstrated.

A. Detector Stability and ROC Curves


Fig. 13(a) shows the receiver operating characteristic (ROC)
curve for detection under the variation of the number of frames
where detection is present, and the data points are the number of
candidates that are considered per radius. This analysis is based
on 132 frames composed of two sequences—the highway off-
ramp and another around the university (see Fig. 12). In these Fig. 13. ROC curves with respect to the number of frames for which a
sequences, we begin at the point where the sign is first detected candidate is required to be present. (a) Number of frames is varied from 0 to
and end when it leaves the frame. The line of best fit (cubic 9, while data is shown for accepting the single best to the ten best candidates.
The fit curve shows the best ten candidates. (b) For the top four candidates per
splines) is for the ten candidates per set of radius data. radius.
The shape of the curve does not vary greatly with the number
of candidates. Furthermore, the data tightly cluster around this
curve, with a slight performance improvement for less than ten
B. System Considerations
candidates. This shows that the detector is stable and reliable.
If we take the reasonable assumption that a sign will be visible Within the current system, we classify six candidate signs per
for at least a significant fraction of a second (otherwise, a driver frame. As classification processing is asynchronous with frame
would not see it) and, furthermore, that a reliable slow classi- processing, the system can tolerate five false positives per frame
fication method is available, then, over the visible frames, this on average. Given this, we seek a point where false positives are
shows the possibility of this detector being a component of a less than five, and we maximize the number of correct classifi-
highly reliable sign warning system. Under these assumptions, cations. We take the four candidates with maximum votes at
the detector will have a little tolerance for false negatives and a each radius and keep those that have a similar size and position
reasonable tolerance for false positives per frame, as shown in for two frames. In this case, the average false-positive rate is
Fig. 13(a). The curve also points to the ideal balance of each, 4.23, and over a number of processed frames, the system has a
given the available processing and system requirements. We low false-negative rate. To statistically consider the error rate,
now consider a configuration for a particular example system note that the total number of false negatives per frame for the
that we use to demonstrate this performance. top four candidates is 0.1894. For any given sequence of images
where the sign is visible, there are two failure cases when
every odd or even image is not recognized. Assuming that the
3 http://cecs.anu.edu.au/~nmb/speedsigns/. failure probability arises from a uniform distribution of failures
330 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 9, NO. 2, JUNE 2008

Fig. 15. These sign images were too small to be detected but were detected in
subsequent images in Fig. 14(a) and (b), respectively.

Fig. 14. Signs in difficult images were generally correctly detected with a Fig. 16. Parameters used in the integrated system.
high threshold, even when difficult with the human eye. All these signs were
detected with the standard threshold. (a)–(c) Heavily shadowed scenes and
(d) a weathered sign.
D. Real-Time Performance
The detector returning candidates comfortably runs at the
across the sequence, we may estimate the total probability of
given frame rate (30 Hz) with four radii and a 320 × 240
not detecting a sign as 0.1894n/2 for missing each of the odd
image. Classification only slows this down when synchronously
and even frames based on this data set. If we assume that a
run. The implementation is in C++ with some mmx operations
sign is visible to the camera for half a second with a processing
running on an Intel Xeon 3.40-GHz machine mounted within
rate of 30 Hz (n = 15 frames, (n/2) = 7 and 8) within the
the car. A slightly reduced version of the detector and the
possible range of radii, the probability of a false negative is
recognition system comfortably runs on a standard Pentium 4
9 per 1 million. This is actually a poor case of an assumed
1.6-GHz laptop.
distribution. Most failed recognition cases are clustered toward
the start of the sequences, as when the sign is closer, it is easier
to detect; therefore, the true failure rate for these sequences E. Integration With Classification
is lower. These statistics are only taken across two sequences.
We combined the detector and the classifier to demonstrate
Clearly, for a thorough examination, a very large number of
the detection ability in real time over a large number of video
sequences are necessary; however, this would be a significant
sequences, and running online in the vehicle on public roads, as
exercise.
what appears on the Web site. The parameters used are shown
in Fig. 16. The sequence video screen classification display is
updated every time the system detects signs, and it has high
C. Difficult Cases
correlation for two successive frames. For most sequences, the
The system, including the detector, was run over sequences sign is classified once or more than once, and often, multiple
showing varying lighting conditions and sign quality. In the correct classifications occur, although, in a few instances, the
case of Fig. 14, the heavily weathered sign was not well best correlation is with the wrong speed. The last classification
detected until it was a little closer than is generally the case but remains on the screen, and, although a sign is visible, this may
was detected in the image shown. The ANU campus has many rapidly change. Stepping through the video frame by frame, it
shaded roads. Fig. 14 shows a number of images where the is possible to see the transitions.
scene is difficult for color-based methods, and the substantial Fig. 17 shows false positives versus false negatives for all of
lighting variation means that the sign has poor contrast. All the sequences of signs using the detector classifier combination.
signs in these images were correctly detected. Four candidates are taken for each radius of the detector per
Last, Fig. 15 shows images with signs that are too small frame, and zero to five repeated appearances are required.
to be recognized but were recognized in subsequent images. Fig. 17 shows that by requiring classification for two frames,
Note that the detector may not reliably detect small signs as we can have 93% successful detection with around 0.5 false
the number of possible pixels becomes small, and the high positive per sequence or 86% successful detection with a false
curvature of the circle may reduce the accuracy of the gradient positive for every five signs that appear.
direction estimation. With a faster processor, higher resolution Note that this classifier may report multiple different sign
images could be used. Multiple processors could also process classifications in a sequence rather than taking the most likely
parts of the image. over the whole sequence. As noted earlier, better methods
BARNES et al.: REAL-TIME SPEED SIGN DETECTION USING THE RADIAL SYMMETRY DETECTOR 331

[8] L. Priese, J. Klieber, R. Lakmann, V. Rehrmann, and R. Schian, “New


results on traffic sign recognition,” in Proc. Intell. Vehicles Symp.,
Aug. 1994, pp. 249–254.
[9] A. de la Escalera, A. Moreno, L. E. Salichs, and J. M. Armingol,
“Road traffic sign detection and classification,” IEEE Trans. Ind. Electon.,
vol. 44, no. 6, pp. 848–859, Dec. 1997.
[10] G. Piccioli, E. D. Micheli, P. Parodi, and M. Campani, “Robust method for
road sign detection and recognition,” Image Vis. Comput., vol. 14, no. 3,
pp. 209–223, Apr. 1996.
[11] C. Y. Fang, C. S. Fuh, S. W. Chen, and P. S. Yen, “A road sign recognition
system based on dynamic visual model,” in Proc. IEEE Conf. Comput.
Vis. Pattern Recog., 2003, vol. 1, pp. 750–755.
[12] D. G. Shaposhnikov, L. N. Podladchikova, A. V. Golovan, and
N. A. Shevtsova, “A road sign recognition system based on dynamic visual
model,” in Proc. Vis. Interface, Calgary, AB, Canada, 2002.
[13] S. H. Hsu and C. L. Huang, “Road sign detection and recognition
using matching pursuit method,” Image Vis. Comput., vol. 19, no. 3,
pp. 119–129, 2001.
[14] L. Pacheco, J. Batlle, and X. Cufi, “A new approach to real time traffic sign
recognition based on colour information,” in Proc. IEEE Intell. Vehicles
Symp., Paris, France, 1994, pp. 339–344.
[15] A. de la Escalera, J. M. Armingol, J. M. Pastor, and F. J. Rodriguez,
Fig. 17. ROC curves with respect to the correlation value that is required for “Visual sign information extraction and identification by deformable mod-
different numbers of frames for which the sign is required to be present. els for intelligent vehicles,” IEEE Trans. Intell. Transp. Syst., vol. 5, no. 2,
pp. 57–68, Jun. 2004.
[16] X. W. Gao, L. Podladchikova, D. Shaposhnikov, K. Hong, and
N. Shevtsova, “Recognition of traffic signs based on their colour and shape
for classification are available, and clearly, there are methods features extracted using human vision models,” J. Vis. Commun. Image
Represent., vol. 17, no. 4, pp. 675–685, Aug. 2006.
that are available to improve the classification results shown. [17] L. Estevez and N. Kehtarnavaz, “A real-time histographic approach to
Some failures were also noticed on multilane roads, where a road sign recognition,” in Proc. IEEE Southwest Symp. Image Anal.
wider field-of-view camera would help. Possibly, high-quality Interpretation, San Antonio, TX, 1996, pp. 95–100.
[18] S. Escalera and P. Radeva, Recent Advances in Artificial Intelli-
imaging devices could also assist. However, video and on- gence Research and Development. Amsterdam, The Netherlands: IOS,
road trials illustrate that the detector is able to produce reliable 2004, ch. Fast Greyscale Road Sign Model Matching and Recognition,
results within a sign recognition system and run at the frame pp. 69–76.
[19] S. Maldonado-Bascon, S. Laufente-Arroyo, P. Gil-Jimenez, H. Gomez-
rate on off-the-shelf PCs. Moreno, and F. Lopez-Ferreras, “Road-sign detection and recognition
based on support vector machines,” IEEE Trans. Intell. Transp. Syst.,
vol. 8, no. 2, pp. 264–278, Jun. 2007.
VI. C ONCLUSION [20] B. Besserer, S. Estable, and B. Ulmer, “Multiple knowledge sources and
evidential reasoning for shape recognition,” in Proc. Int. Conf. Comput.
The fast radial symmetry detector has been demonstrated Vis., May 1993, pp. 624–631.
[21] M. Thomson and S. Westland, “Colour-imager characterization by
to be highly effective for sign detection, combining well with parametric fitting of sensor responses,” Color Res. Appl., vol. 26, no. 6,
recognition. We have shown that it has a low false-negative rate pp. 442–449, Dec. 2001.
and a sufficiently low false-positive rate ans that, when com- [22] R. Labayrade, D. Aubert, and J.-P. Tarel, “Real time obstacle detection
in stereovision on non flat road geometry through v-disparity representa-
bined with reliable classification, effective online sign recogni- tion,” in Proc. IEEE Intell. Vehicles Symp., 2002, pp. 646–651.
tion can be demonstrated. The fast radial symmetry detector [23] C. Bahlmann, Y. Zhu, V. Ramesh, M. Pellkofer, and T. Koehler, “A system
runs in real time and, being a pixel-based shape detector, is for traffic sign detection, tracking, and recognition using color, shape, and
motion information,” in Proc. IEEE Intell. Vehicles Symp., Las Vegas, NV,
robust under changing lighting conditions that cause problems 2005, pp. 255–260.
for color-based detectors. [24] P. Silapachote, J. Weinman, A. Hanson, R. Weiss, and M. A. Mattar,
“Automatic sign detection and recognition in natural scenes,” in Proc.
IEEE Workshop Comput. Vis. Appl. Visually Impaired, Jun. 2005, p. 27.
[25] P. Paclik, J. Novovicova, and R. P. W. Duin, “Building road-sign classifiers
R EFERENCES using a trainable similarity measure,” IEEE Trans. Intell. Transp. Syst.,
[1] C. N. Kloeden, A. J. McLean, V. M. Moore, and G. Ponte, “Travelling vol. 7, no. 3, pp. 309–321, Sep. 2006.
speed and the risk of crash involvement,” NHMRC Road Accident Res. [26] L. Fletcher, G. Loy, N. Barnes, and A. Zelinsky, “Correlating driver gaze
Unit, Univ. Adelaide, Adelaide, Australia, Tech. Rep. CR 204, 1997. with the road scene for driver assistance,” Robot. Auton. Syst., vol. 52,
[2] D. L. Strayer and W. A. Johnson, “Driven to distraction: Dual-task studies no. 1, pp. 71–84, Jul. 2005.
of simulated driving and conversing on a cellular telephone,” Psychol. [27] G. Loy and A. Zelinsky, “Fast radial symmetry for detecting points of
Sci., vol. 12, no. 6, pp. 462–466, Nov. 2001. interest,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 25, no. 8, pp. 959–
[3] F. Lindner, U. Kressel, and S. Kaelberer, “Robust recognition of traffic 973, Aug. 2003.
signals,” in Proc. IEEE IVS, Parma, Italy, Jun. 2004, pp. 49–53. [28] L. G. Minor and J. Sklansky, “The detection and segmentation of blobs in
[4] M. Betke and N. Makris, “Fast object recognition in noisy images using infrared images,” IEEE Trans. Syst., Man, Cybern., vol. SMC-11, no. 3,
simulated annealing,” A.I. Lab., Mass. Inst. Technol., Cambridge, MA, pp. 194–201, Mar. 1981.
Tech. Rep. AIM-1510, 1994. [29] N. Barnes, G. Loy, D. Shaw, and A. Robles-Kelly, “Regular polygon
[5] D. M. Gavrila, “A road sign recognition system based on dynamic visual detection,” in Proc. Int. Conf. Comput. Vis., 2005, pp. 778–785.
model,” in Proc. Int. Conf. Pattern Recog., Aug. 1998, pp. 16–20. [30] N. M. Barnes and Z. Q. Liu, “Embodied categorisation for vision-guided
[6] P. Paclik, J. Novovicova, P. Somol, and P. Pudil, “Road sign classification mobile robots,” Pattern Recognit., vol. 37, no. 2, pp. 299–312, Feb. 2004.
using Laplace kernel classifier,” Pattern Recognit. Lett., vol. 21, no. 13/14, [31] J. Bentley, More Programming Pearls: Confessions of a Coder. Reading,
pp. 1165–1173, Dec. 2000. MA: Addison-Wesley, 1988.
[7] J. Miura, T. Kanda, and Y. Shirai, “An active vision system for real-time [32] N. Barnes, “Improved signal to noise ratio and computational speed for
traffic sign recogntition,” in Proc. IEEE Int. Vehicles Symp., Oct. 2000, gradient-based detection algorithms,” in Proc. IEEE ICRA, Barcelona,
pp. 52–57. Spain, Apr. 2005, pp. 4661–4666.
332 IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. 9, NO. 2, JUNE 2008

Nick Barnes (M’92) received the B.Sc. degree Luke S. Fletcher (M’06) received the B.E. (Elec.)
(with Honors) and the Ph.D. degree in computer (Hons.) and B.Sc. (Comp. Sci.) degrees from The
vision for robot guidance from The University of University of Melbourne, Melbourne, Australia, in
Melbourne, Melbourne, Australia, in 1992 and 1999, 1996. He is currently working toward the Ph.D.
respectively. degree in engineering at The Australian National
From 1992 to 1994, he was a Consultant in the IT University, Canberra, Australia, specializing in
industry. In 1999, he was a Visiting Research Fellow driver monitoring integrated with environmental
with the LIRA-Lab, University of Genoa, Genoa, perception.
Italy, where he was supported by an Achiever Award He is currently a Postdoctoral Fellow with the
from the Queens’ Trust for Young Australians. From Computer Science and Artificial Intelligence Labo-
2000 to 2003, he lectured in computer science and ratory, Massachusetts Institute of Technology (MIT),
software engineering with The University of Melbourne. Since 2003, he has Cambridge. He recently competed in the Defense Advanced Research Projects
been with the Canberra Research Laboratory, National ICT Australia, Canberra, Agency Urban Challenge on the MIT team and is now working on related
Australia, where he is currently the Principle Researcher and Research Group technologies for driver support.
Manager in computer vision. His research interests include visual dynamic
scene analysis, computational models of biological vision, feature detection,
vision for vehicle guidance, and medical image analysis.

Alexander Zelinsky (M’93–SM’07–F’08) lectured


in computer science with the University of
Wollongong, Wollongong, Australia, from 1984
to 1991. From 1992 to 1995, he was a Research
Scientist with the Electrotechnical Laboratory,
National Institute of Advanced Industrial Science
and Technology, Japan. He was the Chief Executive
Officer and the Cofounder of Seeing Machines: a
computer vision systems company. Seeing Machines
is publicly listed in the London Stock Exchange.
Seeing Machines technology was developed at
The Australian National University, Canberra, Australia, where he was a
Professor of systems engineering from 1996 to 2004. In July 2004, he joined
the Commonwealth Scientific and Industrial Research Organization (CSIRO),
Canberra, where he is currently the Group Executive of Information and
Communication Sciences and Technology. He has over 100 publications in
robotics and is an active member of the robotics community. Since 2000, he
has served on the editorial board of the International Journal of Robotics
Research. His research interests include mobile robotics, computer vision, and
human–machine interaction.
Dr. Zelinsky became a Fellow of the Australian Academy of Technological
Sciences and Engineering in 2002. His contributions have been recognized
by awards given in Australia and internationally. These include the Australian
Engineering Excellence Awards in 1999 and 2002, the US R&D magazine Top
100 Award in 2001, the Technology Pioneer at the World Economic Forum
from 2002–2004, and the Clunies Ross Award in 2005. In 1997, he founded the
Field & Services Robotics conference series. From 2005 to 2007, he served with
the Administrative Committee of the IEEE Robotics and Automation Society,
where he is currently the Vice President of the Industrial Activities Board. He
served on the editorial board of IEEE Robotics and Automation Magazine from
2002 to 2006. He is currently an Associate Editor of the IEEE TRANSACTIONS
ON I NTELLIGENT T RANSPORTATION S YSTEMS .

You might also like