Академический Документы
Профессиональный Документы
Культура Документы
Zhi-Qiang Liu
Intelligent Systems Lab.(ISL)
Department of Computer Science and Software Engineering,
The University of Melbourne,
VIC 3010, AUSTRALIA
(z.liu@computer.org)
November 3, 2000
Abstract
Causation plays a critical role in many predictive and inference tasks. Bayesian
networks (BNs) have been used to construct inference systems for diagnostics and
decision making. More recently, fuzzy cognitive maps (FCMs) have gained con-
siderable attention and oer an alternative framework for representing structured
human knowledge and causal inference. In this paper I briey introduce Bayesian
networks and cognitive networks and their causal inference processes in intelligent
systems.
Key Words: Bayesian Networks, Fuzzy Cognitive Map, Causation, Inference,
Decision Making, Dynamic Systems, Intelligent Systems, Graphs
_
w
ij
_
_
,
12
where w
ij
is the value of the arc from node j to node i, i.e., value of arc a
ji
.
Based on causal inputs, the concepts in the FCM can determine their states. This
gives FCMs ability to infer causally. As suggested by Kosko [14, 15], this can be deter-
mined simply by a threshold. In general, we dene a vertex function as follows.
Denition 1 A vertex function f
T,v
i
is dened as:
x
i
= f
T,v
i
() =
_
_
_
1 > T
0 T
,
where is the total input of v
i
, i.e., =
k
w
ik
x
k
.
Usually, if T = 0, f
T,v
i
is denoted as f
v
i
, or simply, f
i
. For the sake of simplicity
and without loss of generality, throughout this paper we assume T = 0 unless specied
otherwise.
Given this denition, an FCM can be dened as a weighted digraph with vertex
functions. denotes an FCM, v(), or simply, v stands for vertex (or concept) of ;
V () represents the set containing all the vertices of ; x() or x is the state of v();
= (x
1
, , x
n
)
T
denotes the state of , where x
i
is the state of vertex v
i
and n is
the number of vertices of ; a() or a stands for an arc in ; A() represents the set
containing all the arcs of ; v
O
(a()) is the start vertex of a() and v
I
(a()) is the end
vertex of a().
As every vertex may take a value in {0, 1}, the state space of the FCM is {0, 1}
n
,
denoted by X
0
() or X
0
. If the state of is after k inference steps from an initial
state
0
, we say that can be reached from
0
, or can reach after k inference steps.
Although the initial state
0
of the FCM may be any state, i.e.
0
X
0
= {0, 1}
n
, some
states can never be reached no matter which initial state it is set to. For example, state
( 1 1)
T
cannot be reached in the FCM shown in Figure 6, where * means that it can
be any value in {0, 1}.
0.5
0
.7
0
.
3
v1
v2
v3
v4
Figure 6: State Space of Fuzzy Cognitive Map
13
We dene X() or X as the reachable state set of which contains all states that
can reach. X
() or X
() X() X
0
().
It is easy to see which state can be reached by the FCM in Figure 6. However, it will
be dicult if the FCM contains a large number of concepts and complex connections.
Whether a state can be reached in a given FCM or not is a fundamental and important
problem in the analysis and design of FCMs.
Clearly, if a state
can be reached in
one step inference if
0
is set as the initial state. The correctness of this assertion can
be proved as follows. Since
(0),
(1), ,
(k) =
.
Obviously,
1 k.
Select a new initial state as (0) =
.
Renumbering the vertices of we can write
slightly dierently:
= (
T
+
,
T
)
T
,
where
+
= (1, 1, , 1)
T
,
= (0, 0, , 0)
T
. Denote W as the adjacency matrix of
the renumbered FCM, n
+
= dim(
+
), n
= dim(
), W
+
= (W
T
1
, , W
T
n
+
)
T
, W
=
(W
T
n
+
+1
, , W
T
n
+
+n
)
T
.
0
has a solution in {0, 1}
n
.
From the above discussion, we may develop an algorithm to determine whether a
state can be reached from a particular initial state.
4 Causal Module of FCM
A general FCM may contain a large number of vertices with very complicated connec-
tions. It is dicult to be handled directly, if at all possible. However, an FCM can be
14
divided into several basic modules, which will be explicitly dened below. Every causal
module is a smaller FCM. Vertices (or concepts) of a causal module infer each other and
are closely connected. Basic FCM modules are the minimum FCM entities that cannot
be divided further.
An FCM is divided as FCM
1
and FCM
2
if
1)V () = V (
1
) V (
2
),
2)A() = A(
1
) A(
2
) B(
1
,
2
),
where
B(
1
,
2
) =
_
a() | V
I
(a()) V (
1
), V
O
(a()) V (
2
),
or V
I
(a()) V (
2
), V
O
(a()) V (
1
) },
V (
1
) V (
2
) = ,
A(
1
) A(
2
) = ,
A(
1
) B(
1
,
2
) = ,
A(
2
) B(
1
,
2
) = .
This operation is denoted as
=
1
B(
1
,
2
)
2
.
Particularly, we consider the causal relationships in subgraphs:
B(
1
,
2
) =
_
a()|V
I
(a()) V (
1
), V
O
(a()) V (
2
)
_
,
this case is described as
2
is caused by
1
, or
1
causes
2
. Such a division is
called a regular division, denoted as
=
1
B(
1
,
2
)
2
.
Denition 2 An FCM containing no circle is called a simple FCM .
A simple FCM has no feedback mechanism and its inference pattern is usually trivial.
Denition 3 A basic FCM is an FCM containing at least one circle, but cannot be
regularly divided into two FCMs, both of which contain at least one circle.
15
From the Denitions 2 and 3, we can see that an FCM containing circles can always
be regularly divided into basic FCMs. In general, the basic FCMs of can be ordered
concatenately as follows.
=
1
B(
1
,
2
)
2
B(
2
,
3
)
3
B(
m1
,
m
)
m
.
We establish a formal result in the following theorem.
Theorem 1 Suppose FCM can be regularly divided into m basic FCMs, then
= (
m
i=1
i
) (
m
i=2
i
j=1
B(
j
,
i
)), (1)
where
V (
i
) V (
j
) = i = j,
A(
i
) A(
j
) = i = j.
B(
i
,
j
) =
_
a()|V
I
(a()) V (
i
), V
O
(a()) V (
j
)}.
. . . . . . . . . . . .
1
2
3
m1
m
Figure 7: Causal Module of FCM
The inference pattern of a basic FCM
i
is determined by its input (external) and
initial state. The inputs of
i
can be determined once the inference pattern of
k
, k < i
are known. Subsequently, the inference pattern of
i
can be analyzed. If we know the
inference patterns of the basic FCMs individually, we will be able to obtain the inference
pattern of the entire FCM, because the basic FCMs collectively contain all the concepts
of the FCM:
V () =
m
i=1
V (
i
).
The following theorem determines if an FCM is not a basic FCM.
16
Theorem 2 Suppose that FCM is input-path standardized and trimmed of aected
branches. If a vertex v
0
of has at least two input arcs and v
0
does not belong to any
circle, then is not a basic FCM.
Theorem 2 provides not only a rule for determining whether an FCM is a basic FCM
or not, but also presents an approach to regularly dividing an FCM if it is not a basic
FCM: In(v
0
) and Out(v
0
). This is illustrated in Figure 8. The FCM in Figure 8(a) can
be regularly divided into FCM in Figure 8(b) and in Figure 8(c), respectively, where
Figure 8(b) is In(v
9
) and Figure 8(c) is Out(v
9
).
v7
v1
v2 v3
v8
v9
v4
v5
v6
v1
v8
v7
v3
v9
v4
v5
v6
v2
(a)
(b)
(c)
Figure 8: Regularly divided and Basic FCMs:According to vertex v
9
, FCM in (a) is
divided as Out(v
9
) in (b) and In(v
9
) in (c).
More specically such a division can be done by the following algorithm.
Algorithm 1
Step 0 If is a simple FCM stop.
Step 1 Select a vertex v of the , mark v.
Step 2 Form In(v).
Step 3 If there is no circle in In(v), go to step 7.
Step 4 Form Out(v).
Step 5 If there is no circle in Out(v), go to step 7.
Step 6 is regularly divided into In(v
9
) and Out(v
9
), stop.
Step 7 If there is no unmarked vertices, is a basic FCM, stop.
Step 8 Select an unmarked vertex v, mark v, go to step 2.
In general, an FCM can be regularly divided into basic FCMs by repeatedly imple-
menting Algorithm 1.
17
5 Inference Patterns of Basic FCMs
The inference pattern of an FCM can be obtained by recursively calculating
(k + 1) = f[W(k)] = (f
1
[W
1
(k)], , f
n
[W
n
(k)])
T
.
However, since in most real applications the FCM contains a large number of vertices and
complicated connections, the state sequence can be very long and dicult to analyze.
It will be most useful to draw out properties or recursive formula for the state sequence.
Proposition 1 If an FCM is a simple FCM, it will become static after L inference
iterations unless it has an external input sequence, where L is the length of the longest
path of the FCM.
The Proof of Proposition 1 is obvious. Consequently, the following is true.
Corollary 1 Vertices except the end vertex of an input path will become static after L
inference iterations unless it has an external input sequence, where L is the length of the
path.
In this section, all FCMs are assumed as basic FCMs, with input paths being stan-
dardized and aected branches trimmed. For the sake of simplicity, we assume that all
FCMs do not have external input sequences unless they are specically indicated. FCMs
with external input sequences can be analyzed in the similar way.
In our study of the inference pattern of the FCM, we found that some vertices may
play more important roles than others. We dene these vertices as key vertices. The
state of every vertex in the FCM can be determined by the state of key vertices. In the
following part of this section, the denition of key vertex is followed by some discussions
of the properties of key vertex.
Denition 4 A vertex is called as a key vertex if
1. it is a common vertex of an input path and a circle, or
2. it is a common vertex of two circles with at least two arcs pointing to it which
belongs to the two circles, or
3. it is any vertex on a circle if the circle contains no other key vertices.
18
Proposition 2 Every circle contains at least one key vertex.
Given Property 3 in Denition 4, the correctness of Proposition 2 is obvious.
Lemma 1 If any vertex, v
0
V (), is not on an input path, then it is on a circle.
Lemma 2 Suppose that v
0
is not a key vertex and not on an input path, then there is
one and only one key vertex (denoted as v
) can aect v
0
via a normal path, and there
is only one normal path from v
to v
0
.
Again suppose the key-vertex set is {v
1
, , v
r
}. If v
j
(1 j r) satises (1) of
Denition 4, it is the end vertex of an input branch, denote v
l=1
r(l,j)
s=0
f
P
s
l,j
{x
l
(l L(P
s
l,j
)}). (2)
If v
j
is the end of an input branch,
x
j
(l) = f
T
1
j
,T
2
j
,v
j
(
r
l=1
r(l,j)
s=0
f
P
s
l,j
{x
l
(k L(P
s
l,j
)} +f
P
v(j),j
(x
v(j)
(l L(P
v(j),j
)))). (3)
Therefore the inference pattern of the FCM is can be determined by the recursive
formula in terms of key vertices. After the states of key vertices are determined, the
states of the remaining vertices can be determined by states of key vertices as follows.
In turn, the state of the entire FCM is also determined.
Theorem 4 Suppose that K
v
() is the key vertex set of , I
v
() is the input vertex set
of , v
0
V (), v
0
/ K
v
() and v
0
/ I
v
(). There exists one and only one vertex x
,
x
K
v
() or x
I
v
() such that there exists a path, P
(v
, v
0
) from v
to v
0
via no
vertices in K
v
().
Theorem 5 P
(v
, v
0
) is a normal path.
19
As P
(v
, v
0
) is a normal path,
x
0
(l) = f
P
(v
,v
0
)
(x
(l L(P
(v
, v
0
))).
Thus the state of the entire FCM is determined. With the help of Theorems 2, 4 and 5,
we discuss some important causal inference properties of typical FCMs in the following
sections.
6 Inference Pattern of General FCMs
When considering the inference pattern of a general FCM(), we should rst regularly
divide it into basic FCMs (
i
, 1 i m) and then determine the inference pattern one
by one according to the causal relationships between of them. For the basic FCM(
i
),
the external input should be formed according to the inference pattern of
j
1 j < i.
Then the input paths should be standardized. After this process, we delete all the
aected branches to simplify the FCM(
i
) for further analysis.
From the simplied FCM, we need to construct the key-vertex set. If the basic FCM
contains only one circle, and every vertex on the circle has only one input arc, then the
key vertex set contains only one vertex. It can be any vertex on the circle according
to Denition 11. If the basic FCM contains more than one circle, the key vertices are
the circle vertices that have at least two input arcs. This can be judged from W. If the
ith column of W has at least two non-zero elements, v
i
has at least two input arcs. If
T
ii
= 0, as the following proposition indicates, v
i
is on a circle.
Proposition 3 Vertex v
i
is on a circle if and only if T
ii
= 0, where T
ii
is the ith row
and ith column element of matrix T =
n
i=k
W
k
.
Following Equations (2) and (3) in Theorem 2, we can obtain the inference pattern
of key vertices. Subsequently the states of other vertices including those on the aected
branches can be determined accordingly.
The steps for analyzing the inference pattern of an FCM are given below.
Algorithm 2
step 1 Divide the regularly into basic FCMs:
1
, ,
m
:
= (
m
i=1
i
) (
m
i=2
i
j=1
B(
j
,
i
)).
20
step 2 i = 1.
step 3 Delete the attached branches of
i
.
step 4 Construct new inputs from basic FCMs
j
by B(
j
,
i
(j < i)
step 5 Standardize the input paths as normal paths.
step 6 Construct the key-vertex set.
step 7 Determine the state-sequence formula of the key vertices according to Theo-
rem 2.
step 8 Determine the state-sequence formula of the remaining vertices.
step 9 Determine the state-sequence formula of the aected branch.
step 10 i=i+1, if i < m, go to step 2, else stop.
The algorithm is illustrated by the following example. In the example, all the arcs
are assumed to be positive, i.e. w
i,j
> 0.
Example
The FCM
a
shown in Figure 9 can be simplied as
b
by trimming o the aected
branches.
v10
v1
v2
v3
v4
v5
v6
v7
v8
v9
v11
v12
v13
v14
v15
v16
v17
v18
v19
v20
v1
v2
v8
v9
v11
v12
v13
v18
v19
v5
v6
v16
v20
a
b
Figure 9: Example of Analyzing Inference Pattern of FCMs:
a
is simplied as
b
by
trimming the aected branches of
a
FCM
c
and
d
in Figure 10 are the two basic FCMs in
b
shown in Figure 9.
Figure 11 shows
e
which is
c
minus the only aected branch, AB(v
20
). v
18
is the only
key vertex of
e
:
x
18
(l) = f
18
(x
18
(l 4) +x
18
(l 6)).
21
As 2 is the common factor of 4 and 6, the nal state sequence of v
18
is: 0, 1, 0, 1, , 0, 1, 0, 1, .
The remaining vertex states of
e
can be completely determined by x
18
. For example,
x
1
(l) = x
18
(l 2).
The nal state sequence of v
1
is also 0, 1, 0, 1, , 0, 1, 0, 1, .
v11
v5
v6
v16 v11
v1
v2
v8
v9
v12
v13
v18
v19
v20
v1
v2
v8
v9
v12
v13
v18
v19
v20
v1
v2
v8
v9
v12
v13
v18
v19
v20
v1
v2
v8
v9
v12
v13
v18
v19
v20
c
d
Figure 10: Example of Analyzing Inference Pattern of FCMs:
b
can be regularly divided
into two basic FCMs:
c
and
d
The aected branch AB(v
20
) of
c
contains only one vertex: v
20
. Its state is deter-
mined by
e
, or more specically, by v
12
.
x
20
(l) = x
12
(l 1).
The nal state sequence of v
20
is also: 0, 1, 0, 1, , 0, 1, 0, 1, .
After all the state patterns of
c
are determined, we can reconstruct inputs for basic
FCM
d
. It is shown in Figure 11 as
f
. The key vertex of
f
is v
16
. By Theorem 2
x
16
(l) = f
16
(x
16
(l 3) +x
20
(l 1)).
As the common factor of 2 and 3 is 1, the nal state of
f
is
x
6
x
16
x
5
1.
With all the state patterns of
b
being determined, it is easy to obtain the state
pattern for the vertices (v
3
, v
4
, v
15
, v
14
, v
7
, v
17
, v
10
) in the aected branches of
a
by the
state of
b
and it is omitted here.
22
v1
v2
v8
v9
v12
v13
v18
v19
v1
v2
v8
v9
v12
v13
v18
v19
v5
v6
v16
v20
e
f
Figure 11: Example of Analyzing Inference Pattern of FCMs:
e
is obtained by deleting
the aected branch of
c
.
f
is derived from
d
by reconstructing the input according
to the inference pattern of
c
7 Conclusions
In this paper I have given a brief discussion of the theory of causation, and its impor-
tance in decision-support and causal discovery systems. One of the well-known causal
frameworks is the Bayesian network which uses conditional probability to model causal
relationships and uncertainty of the system. Whereas the Bayesian network has received
consideration attention in the research community, it still remains an illusive goal for
developing robust, functional systems for real-world applications. One of the most signif-
icant problems is its inability to model vagueness, ambiguity, and natural descriptions.
Fuzzy cognitive maps oer an alternative solution to the problem. I have presented a
simple discussion of the causal inference process using FCMs. A general FCM may be
very complex, but it can be regularly divided into several basic FCMs according to the
results in Section 4. The inference pattern of a basic FCM is the dominant state sequence
of some key vertices whose behaviors are described by a general recursive formula. Based
these results, I have analyzed the inference patterns of several typical FCMs with an
algorithm.
References
[1] S. Andreassen, F.V. Jensen, S.K. Andersen, B. falck, U. Kjaerul, M. Woldbye,
A. Sorensen, A. Rosenfalck, F. Jensen, MUNU - an expert EmG assistant, in
23
Computer-aided Electromyography and Expert Systems, J.E. Desmedt (ed), pp.255-
277, 1989.
[2] R. Axelrod, Structure of Decision: the Cognitive Maps of Political Elites, Princeton,
NJ:Princeton University Press, 1976.
[3] T. Binford, T. Levitt, and W. Mann, Bayesian inference in model-based machine
vision, in Uncertainty in Articial Intelligence, J.F. Lemmer & L.M. Kanal (eds),
Elsevier Science, Amsterdam, 1988.
[4] G.F. Cooper, The Computational complexity of probabilistic inference using
Bayesian networks, Articial Intelligence, Vol.42, pp.393-405, 1990.
[5] G.F. Cooper and E. Herskovits, A Bayesian method for the induction of proba-
bilistic networks from data, Machine Learning, Vol.9, pp.309-347, 1992.
[6] P. Dagum and M. Luby, Approximating probabilistic inference in Bayesian belief
networks is NP-hard, Articial Intelligence, Vol.60, pp.141-153, 1990.
[7] D. Geiger, T. Verma, and J. Pearl, Identifying independence in Bayesian Net-
works, Networks, Vol.20, No.5, pp.507-534, 1990.
[8] C. Glymour and G. Cooper, (Eds) Computation, Causation, and Discovery, AAAI
Press/MIT Press, Cambrige, MA 1999.
[9] F.V. Jensen, An Introduction to Bayesian Networks, Springer-Verlag, Berlin, 1996.
[10] D. Heckerman, Probabilistic Similarity Networks, MIT Press, Cambridge, Mas-
sachusetts, 1991.
[11] D. Heckerman, D. Geiger, and D.M. Chickering, Learning Bayesian networks: the
combination of knowledge and statistical data, Machine Learning, Vol.20, pp.197-
243, 1995.
[12] B. Kosko, Fuzzy Cognitive Maps, International Journal Man-machine Studies,
Vol.24, pp.65-75, 1986.
[13] B. Kosko, Adaptive Inference in Fuzzy Knowledge Networks, Proceeding of the
First Int. Conf. on Neural Networks, Vol.2, pp261-268,1987.
24
[14] B. Kosko, Fuzzy System as Universal Approximators, Proceedings of the 1st IEEE
International Conference on Fuzzy Systems, pp.1153-1162, 1992.
[15] B. Kosko, Neural Networks and Fuzzy Systems: A Dynamical Systems Approach to
Machine Intelligence, Prentice-Hall, Englewood Clis, 1992.
[16] T. Levitt, M. Hedgcock, J. Dye, S. Johnston, V. Shadle, D. Vosky, Bayesian infer-
ence for model-based segmentation of computed radiographs of the hand, Articial
Intelligence, vol.5, No.4, pp.365-387, 1993.
[17] Z. Q. Liu and Y. Miao, Fuzzy Cognitive Map and its causal inference, Proc. IEEE
Int. Conf. on Fuzzy Systems, Seoul, Korea, Vol.III, pp.1540-1545, Aug.22-25, 1999.
[18] Z.Q. Liu, Fuzzy cognitive maps in GIS data analysis, special issue J. Fuzzy Sets
and Systems on The Challenge of Soft Computing for Resolution of Large Scale
Environment Problems, (to appear) 2001.
[19] Z.Q. Liu and R. Satur , Contextual Fuzzy Cognitive Map for Decision Support
in Geographic Information Systems, IEEE Trans. Fuzzy Systems, Vol.7, No.5,
pp.495-507, 1999.
[20] Y. Miao and Z.Q. Liu, On Causal Inference in Fuzzy Cognitive Map, IEEE Trans.
Fuzzy Systems, Vol.8, No.1, pp.107-119, 2000.
[21] Y. Miao, Z.Q. Liu, C.K. Siew, and C.Y. Miao, Dynamical cognitive network-an
extension of fuzzy cognitive map IEEE Trans. Fuzzy Systems, (revised) July 2000.
[22] T. D. Ndousse and T. Okuda, Computational Intelligence for Distributed Fault
Management in Networks Using Fuzzy Cognitive Maps, Proceeding of IEEE Int.
Conf. on Communications Converging Technologies for Tomorrows Application,
IEEE New York, vol.3, pp.1558-1562, 1996.
[23] J. Pearl, A contraint-propagation approach to probablistic reasoning, in Uncer-
tainty in Articial Intelligence, L.M. Kanal & J. Lemmer (eds), pp.357-370, Ams-
terdam, North-Holland, 1986.
[24] J. Pearl, Fusion, proagation, and structuring in belief networks, Articial Intel-
ligence, vol.29, No.3, pp.24-288, 1986.
25
[25] J. Pearl, Probablistic Reasoning in Intelligent Systems, Morgan Kaufmann, San
Mateo, CA, 1988.
[26] J. Pearl, Causality, Cambrige University Press, Cambrige UK, 2000.
[27] L.K. Rasmussen, BOBLO: an expert system based on Bayesian networks to blood
group determination of cattle, Research Report No.16, Research Center Foulum,
Tjele, Denmark, 1995.
[28] S. Russell and P. Norvig, Articial Intelligence, Prentice-Hall, NJ. 1995.
[29] R. Satur and Z.Q. Liu, A context-driven Intelligent Database Processing Sys-
tem Using Object Oriented Fuzzy Cognitive Maps, Int. J. of Intelligent Systems,
Vol.11, No.9, pp.671-689, 1996.
[30] R. Satur and Z.Q. Liu, A Contextual Fuzzy Cognitive Map Framefwork for Ge-
ographic Information Systems, IEEE Trans. Fuzzy Systems, Vol.7, No.5, pp.481-
494,1999.
26