Академический Документы
Профессиональный Документы
Культура Документы
2. RELATED WORK
Object maps are mostly oriented towards mobile robot
navigation. Vasudevan et al. [23] use SIFT as a recognition
tool and develop a hierarchical probabilistic representation
of space that is based on objects. A global topological repre-
sentation of places with object graphs serving as local maps
is suggested.
Anguelov et al. describe a probabilistic approach for de-
tecting and modeling doors in a corridor environment [1].
They use features based on shape, color, and motion proper-
ties of door and wall objects. Modayil and Kuipers describe
in [11] an approach for object localization and recognition
based on object shape models.
Brezetz et al [3] use range scans to segment objects and
represent them in a topological framework. Limeketkai et
al [8] describe Relational Object maps of walls and doors.
Mozos et al [14] used the object map for interpretation of
the environment and Ranganathan et al [18] develop object
map for giving semantic interpretation of places.
Tomono describe in [22] 3D map building algorithms using
vision data. Moro et al. in [13] describe an approach for
object classification based on edge features.
3. THE PROPOSED 3D OBJECTS MAP Figure 1: Block diagram of the proposed algorithm.
BUILDING APPROACH
The stereo camera used in this work is a Bumblebee [6] sis concern the possibility that the pixel (i, j) corresponds to
stereo camera, which is a Firewire CCD camera with a def- an object or not. In other words, we have eight hundred hy-
inition of 640x480 pixels at 48 fps. We consider images ac- pothesis for each texel, namely θ = {θ1 (0, 0), . . . , θ1 (19, 19),
quired by the stereo camera: they are combined and recti- θ2 (0, 0), . . . , θ2 (19, 19)}, where θ1 (i, j) is the belief that the
fied, and for each pixel its depth is computed. All of these pixel (i, j) of that texel belongs to an object in the environ-
low-level processing are performed by the camera internal ment and θ2 (i, j) is the belief that the pixel (i, j) does not
firmware. It is important to note that we divide the image belong to an object.
in texel of 20 × 20 pixels, as all the subsequent processing is
texel based. 3.1.1 Basic Belief Assignment
The algorithm described in this paper is summarized in The Basic Belief Assignment can be viewed as a gener-
the block diagram reported in Fig. 1. alization of a probability density function. More precisely,
a Basic Belief Assignment m() is a function that assigns a
value in [0, 1] to every subset A of θ that satisfies the fol-
As described in Fig. 1, our 3D objects map approach is per- lowing:
formed using the following steps: basic belief assignment by X
processing the rectified and depth images, data fusion using m(A) = 1, m(∅) = 0
the Dempster-Shafer algorithm, ROI estimation, feature ex- A⊆Θ
traction from the rectified image and object classification. It is worth noting that m(A) is the belief that supports the
Using the object labels and the depth map, which is the dis- subset A of θ, not the elements of A. This reflects some
tance from the camera to all the pixels of the image, a 3D ignorance because this means that we can assign belief only
virtual world is finally built using a 3D vectorial graphical to subsets of θ, not to the individual hypothesis as in classical
application. probability theory.
In the following we summarize some results, limited to
those used in this work, from the Dempster-Shafer theory 3.1.2 Belief function
of evidence. Many good tutorials are available, such as for
The belief function, bel(.), associated with the Basic Belief
example [5] and [19].
Assignment m(.), assigns a value in [0, 1] to every nonempty
3.1 The Dempster-Shafer Fusion subset B of θ. It is defined by
The goal of the Dempster-Shafer theory of evidence [20], is
X
bel(B) = m(A)
to represent uncertainty and lack of knowledge. The theory A⊆B
can combine different measures of evidence. At the base of
the theory is a finite set of possible hypotheses, say θ = The belief function can be viewed as a generalization of a
{θ1 , . . . , θK }. probability function.
In our case, a hypothesis set is defined for each texel in
which is divided the image. Within each texel, the hypothe- 3.1.3 Combination of evidence
Consider two Basic Belief Assignments m1 (.) and m2 (.) the evidence m2 (O) is basically set to the number of pixels
and the corresponding belief functions bel1 (.) and bel2 (.). with the same value in the texel. More precisely, setting the
Let Aj and Bk be subsets of θ. Then m1 (.) and m2 (.) can maximum distance sensed by the camera, which can corre-
be combined to obtain the belief mass assigned to C ⊂ θ spond to a wall or, more generally, to an absence of objects,
according to the following formula [20], we set m2 (O) to the number of pixels in the texel that have
P a distance value different by the maximum distance value.
M j,k,A ∩B =C m1 (Aj )m2 (Bk ) If all the pixels in the texel have a maximum distance value
m(C) = m1 m2 = P j k
1 − j,k,Aj ∩Bk =0 m1 (Aj )m2 (Bk ) the value of m2 (O) is set to zero.
(1)
The denominator is a normalizing factor, which measures 3.3 ROI estimation, object pose and height es-
how much m1 (.) and m2 (.) are conflicting. timation
The final evidence of the presence of an object in a texel
3.1.4 Belief functions combination is computed using (1). In this way, each pixel is assigned a
The combination rule can be easily extended to several value that states the belief that there is an object in it.
belief functions by repeating the rule for new belief func- A 3D view of the final belief image is reported in Fig. 2.
tions. Thus the sum of n belief functions bel1 , bel2 , . . . , beln ,
can be formed as
M M n
M
((bel1 bel2 ) bel3 ) . . . beln = beli
i=1
objects.