Вы находитесь на странице: 1из 89

1

Empirical Study on Mining Association Rules


Using Population Based Stochastic Search
Algorithms
K.INDIRA

Under the guidance of
Dr. S.KANMANI
Professor,
Department of Information Technology,
Pondicherry Engineering College
2
ORGANIZATION
OBJECTIVES
INTRODUCTION
MOTIVATION
EMPIRICAL STUDY
CONCLUSION
PUBLICATIONS
REFERNECES
3
INDIAN INSTITUTE OF
TECHNOLOGY, ROORKEE
OBJECTIVES

To develop novel hybrid methodology for mining
Association rules both effectively and efficiently using
population based search methods namely Genetic
Algorithm (GA) and Particle Swarm Optimization(PSO).

4
INTRODUCTION
DATA MINING
Extraction of interesting information or
patterns from data in large databases is known
as data mining.
5
ASSOCIATION RULE MINING
Association rule mining finds interesting
associations and/or correlation relationships
among large set of data items.
6
ASSOCIATION RULES
Association Rules are of form X Y with two
control parameters support and confidence

Support, s, probability that a transaction
contains X Y
Confidence, c, conditional probability that a
transaction having X also contains Y
7
ASSOCIATION RULES
Tid Items bought
10 Milk, Nuts, Sugar
20 Milk, Coffee, Sugar
30 Milk, Sugar, Eggs
40 Nuts, Eggs, Bread
50 Nuts, Coffee, Sugar , Eggs,
Bread
Customer
buys sugar
Customer
buys both
Customer
buys milk
Let minsup = 50%, minconf = 50%
Freq. Pat.: Milk:3, Nuts:3, Sugar:4, Eggs:3, {Milk, Sugar}:3

Association rules:
Milk Sugar (60%, 100%)
Sugar Milk (60%, 75%)
8
MOTIVATION
EXISTING SYSTEM
Apriori, FP Growth Tree, clat are some of the
popular algorithms for mining ARs.
Traverse the database many times.
I/O overhead, and computational complexity is
more.
Cannot meet the requirements of large-scale
database mining.
Does not fit in memory and is expensive to build
9
EVOLUTIONARY ALGORITHM
Applicable in problems where no (good) method is
available:
Discontinuities, non-linear constraints, multi-
modalities.
Discrete variable space.
Implicitly defined models (if-then-else constructs).
Noisy problems.
Most suitable in problems where multiple solutions are
required:
Multi-modal optimization problems.
Multi-objective optimization problems.
10
EVOLUTIONARY ALGORITHM
Parallel implementation is easier.
Evolutionary algorithms provide robust and efficient
approach in exploring large search space.
11
GA AND PSO : AN INTRODUCTION
Genetic algorithm (GA) and Particle swarm
optimization (PSO) are both population based search
methods and move from set of points (population) to
another set of points in a single iteration with likely
improvement using set of control operators.


12
GENETIC ALGORITHM
A Genetic Algorithm (GA) is a procedure used to
find approximate solutions to search problems
through the application of the principles of
evolutionary biology.
13
PARTCILE SWARM OPTIMIZATION
PSOs mechanism is inspired by the social and
cooperative behavior displayed by various
species like birds, fish etc including human
beings.
Association Rule
(AR) Mining
Population Based
Evolutionary
Methods
Genetic Algorithm
(GA)
Particle Swarm
Optimization
(PSO)
Mining Association
Rules using GA
Analyzing the role of
Control parameters
in GA for mining ARs
Mining ARs using
Self Adaptive GA
Elitist GA for
Association Rule
Mining
Mining Association
rules with PSO
Mining Association
Rules with
chaotic PSO
Mining ARs with
Neighborhood
Selection in PSO
Mining Association
rules with Self
Adaptive PSO
Hybrid GA/PSO
(GPSO) for AR
Mining
BLOCK DIAGRAM OF RESEARCH MODULES
15
DATASETS DESCRIPTION
Lenses
Habermans Survival
Car Evaluation
Post operative care
Zoo
16
DATASET DESCRIPTION
Dataset Name No. of
Instances
No. of
Attributes
Attribute
characteristics
Lenses 24 3 Categorical
Habermans
Survival
306 3 Integer
Car Evaluation 1728 6 Categorical

Post Operative
Patient
90 8 Categorical,
Integer
Zoo 101 17 Categorical,
Integer

17
Mining AR using
GA
Mining AR using
SAGA
Mining AR using
Elitism based GA
MODULE 1
18
MINING AR USING GA
Methodology

Selection : Tournament

Crossover Probability : Fixed ( Tested with 3 values)

Mutation Probability : No Mutation

Fitness Function :

Dataset : Lenses, Iris, Haberman from
UCI repository.

Population : Fixed ( Tested with 3 values)


19
FLOWCHART OF GA
20
RESULT ANALYSIS
Population Size Vs Accuracy

P
R
E
D
I
C
T
I
V
E

A
C
C
U
R
A
C
Y

21
RESULT ANALYSIS

P
R
E
D
I
C
T
I
V
E

A
C
C
U
R
A
C
Y

Minimum Support and Confidence Vs Accuracy
22
RESULT ANALYSIS
Pc = .25 Pc = .5 Pc = .75
Accuracy
%
No. of
Generations
Accuracy
%
No. of
Generations
Accuracy
%
No. of
Generations
Lenses 95 8 95 16 95 13
Haberman 69 77 71 83 70 80
Iris 84 45 86 51 87 55
Dataset No. of
Instances
No. of
attributes
Population
Size
Min.
Support
Minimum
confidence
Crossover
rate
Accuracy
in %
Lenses 24 4 36 0.2 0.9 0.25 95
Haberman
306 3 306 0.9 0.2 0.5 71
Iris 150 5 225 0.2 0.9 0.75 87
Comparison based on variation in Crossover Probability
Comparison of the optimum value of Parameters for
maximum Accuracy achieved
23
INFERENCES
Values of minimum support, minimum confidence
and population size decides upon the accuracy of
the system than other GA parameters
Crossover rate affects the convergence rate rather
than the accuracy of the system

24
Mining AR using
GA
Mining AR using
SAGA
Mining AR using
Elitism based GA
MODULE 1
25
MINING AR USING SAGA
Methodology

Selection : Roulette Wheel

Crossover Probability : Fixed ( Tested with 3 values)

Mutation Probability : Self Adaptive



Fitness Function :

Dataset : Lenses, Iris, Car from
UCI Irvine repository.

Population : Fixed ( Tested with 3 values)


26
FLOWCHART OF SAGA
SELF ADAPTIVE
27
RESULT ANALYSIS
75
52
85
87.5
68
96
0
10
20
30
40
50
60
70
80
90
100
Lenses Haberman Car Evaluation
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

Dataset
Traditional GA Accuracy
Self Adaptive GA Accuracy
ACCURACY COMPARISON BETWEEN GA AND SAGA WHEN
PARAMETERS ARE IDEAL FOR TRADITIONAL GA
28
0
20
40
60
80
100
Lenses Haberman's Survival Car Evaluation
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

GA
SAGA
GA WITH SAGA
TERMINALCONDITIONS
Accuracy comparison between GA, SAGA and GA with
parameters set to termination values of SAGA
RESULT ANALYSIS
29
INFERENCES

Self Adaptive GA gives better accuracy than
Traditional GA.

30
Mining AR using
GA
Mining AR using
SAGA
Mining AR using
Elitism based GA
MODULE 1
31
Methodology

Selection : Elitism with roulette wheel

Crossover Probability : Fixed to P
c

Mutation Probability : Self Adaptive

Fitness Function : Fitness(x) = con(x)*(log(sup(x) *
length(x) + 1)

Dataset : Lenses, Iris, Car from UCI
repository.

Population : Fixed


GA WITH ELITISM FOR MINING
AR
32
No. Of
Iterations
Lenses Car
Evaluation
Haberman
4 90 94.4 70
8 91.6 92.8 91.6
10 90 87.5 75
15 87.5 90 83.3
20 91.6 87.5 91.6
25 87.5 87.5 92.5
30 83.3 93.75 83.3
50 90 75 75
RESULTS ANALYSIS
Predictive Accuracy for Mining AR based on GA with Elitism
33
Predictive Accuracy for Mining AR based on GA with Elitism
RESULTS ANALYSIS
0
20
40
60
80
100
120
Lenses Car Evaluation Haberman's survival
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

GA
SAGA
GA with Elitism
34
0
1000
2000
3000
4000
5000
6000
4 6 8 10 15 20 25 30 50
E
x
e
c
u
t
i
o
n

t
i
m
e

(
m
s
)

No. of Iterations
Haberman (ms)
Car Evaluation (ms)
Lenses (ms)
RESULTS ANALYSIS
Execution time for Mining AR based on GA
with Elitism
35
INFERENCES
Marginally better accuracy arrived

The execution time increases rapidly once
global optima reached



36
MODULE 2
Mining AR
using PSO
Mining AR
using CPSO
Mining AR
using NPSO
Mining AR
using PSO
37
RELATED WORKS
S.No Variation Application Year
1 Inertia Weight and
Acceleration coefficients
Optimization 2006
2 Global-Local Best Inertia
Weight
computing optimal
control
2006
3 Local Optima Chaos Parameter
Optimization
2010
4 Adaptive Population Size Optimization

2009
5 Cellular PSO Function
Optimization
2011
6 Adaptive inertia weight
strategy
Optimization 2011
38
MODULE 2
Mining AR
using PSO
Mining AR
using CPSO
Mining AR
using NPSO
Mining AR
using PSO
39
MINING ARS USING PSO
Methodology

Each data itemset are represented as particles

The particles moves based on velocity


The particles position are updated based on



Flow chart depicting the General PSO Algorithm:
For each particles position (p)
evaluate fitness
If fitness(p) better than
fitness(pbest) then pbest= p
L
o
o
p

u
n
t
i
l

a
l
l

p
a
r
t
i
c
l
e
s

e
x
h
a
u
s
t

Set best of pBests as gBest
Update particles velocity and
position
L
o
o
p

u
n
t
i
l

m
a
x

i
t
e
r

Start
Initialize particles with random position
and velocity vectors.
Stop: giving gBest, optimal solution.
FLOWCHART OF PSO
41
RESULTS ANALYSIS
Dataset Name
Traditional
GA
Self
Adaptive
GA
PSO
Lenses 87.5 91.6 92.8
Haberman 75.5 92.5 91.6
Car evaluation 85 94.4 95
0
200
400
600
800
1000
1200
4 6 8 10 15 20 25 30 50
E
x
e
c
u
t
i
o
n

T
i
m
e

m
s
e
c

No. of iterations
Haberman
PSO
SAGA
0
20
40
60
80
100
4 6 8 101520253050
E
x
e
c
u
t
i
o
n

T
i
m
e

m
s
e
c

No. of iterations
Lenses
PSO
SAGA
0
200
400
600
800
1000
1200
4 6 8 10 15 20 25 30 50
E
x
e
c
u
t
i
o
n

T
i
m
e

m
s
e
c

No. of Iterations
Car Evaluation
PSO
SAGA
Execution Time
Predictive Accuracy
42

PSO produce results as effective as self adaptive GA

Computational effectiveness of PSO marginally fast
when compared to SAGA.

In PSO only the best particle passes information to
others and hence the computational capability of
PSO is marginally better than SAGA.


INFERENCES
43
MODULE 2
Mining AR
using PSO
Mining AR
using CPSO
Mining AR
using NPSO
Mining AR
using PSO
44
MINING ARS USING CHAOTIC PSO
The new chaotic map model is formulated as
Methodology
Initial point u
0
and V
0
to 0.1
The velocity of each particle is updated by
45
RESULT ANALYSIS
80
82
84
86
88
90
92
94
96
98
100
Haberman lens car evaluation
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

SAGA
PSO
CPSO
Predictive Accuracy Comparison
46
90
75
96.8
70
75
80
85
90
95
100
SAGA pso cpso
4
6
8
10
15
20
25
30
50
Convergence Rate Comparison for Lenses
RESULT ANALYSIS
47
40
50
60
70
80
90
100
SAGA pso cpso
4
6
8
10
15
20
25
30
50
Convergence Rate Comparison for Car Evaluation
RESULT ANALYSIS
48
0
10
20
30
40
50
60
70
80
90
100
SAGA pso cpso
4
6
8
10
15
20
25
30
50
Convergence Rate Comparison for Habermans Survival

RESULT ANALYSIS
49

Better accuracy than PSO
The Chaotic Operators could be changed by
altering the initial values in chaotic operator
function
The balance between exploration and exploitation
is maintained

INFERENCES
50
MODULE 2
Mining AR
using PSO
Mining AR
using CPSO
Mining AR
using NPSO
Mining AR
using PSO
51
The concept of local best particle (lbest) replacing
the particle best (pbest) is introduced

The neighborhood best (lbest) selection is as
follows;
Calculate the distance of the current particle
from other particles
Find the nearest m particles as the neighbor of
the current particle based on distance
calculated
Choose the local optimum lbest among the
neighborhood in terms of fitness values
Mining ARs using NPSO
Methodology
52
INTERESTINGNESS MEASURE

The interestingness measure for a rule is taken from
relative confidence and is as follows:



Where k is the rule, x the antecedent part of the rule
and y the consequent part of the rule k.
)) sup( 1 )( sup(
) )(sup( sup( ) sup(
) (
y x
y x y x
k ness Intersting

53
FLOWFCHART FOR NPSO
Compute x
i
(k+1)
Compute (f(x
i
(k+1))
Reorder the particles
Generate neighborhoods I =1
k K
i = i +1
K = k+1
Start
K =1 ,Initialize x
i
(k), v
i
(k)
Compute f(x
i
(k))
Determine best particles in
the neighborhood of i
Update previous best if
necessary
I N
Stop
54
RESULT ANALYSIS
88
89
90
91
92
93
94
95
96
97
98
Haberman lens car evaluation
P
r
e
d
i
c
t
i
v
e

A
c
u
r
a
c
y

(
%
)

saga
pso
Npso
Predictive Accuracy Comparison for Dynamic
Neighborhood selection in PSO
55
RESULT ANALYSIS
Dataset Interestingness
Value
Lens 0.82
Car Evaluation 0.73
Habermans Survival 0.8
Measure of Interestingness
56
RESULT ANALYSIS
0
200
400
600
800
1000
1200
1400
1600
4 6 8 10 15 20 25 30 50
Lenses PSO
Lenses NPSO
Haberman's Survival PSO
Haberman's Survival NPSO
Car Evaluation PSO
Car Evaluation NPSO
Execution Time Comparison for Dynamic Neighborhood
selection in PSO

57
0
10
20
30
40
50
60
70
80
90
100
PSO NPSO
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

4
6
8
10
15
20
25
30
50
0
10
20
30
40
50
60
70
80
90
100
PSO NPSO

P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)


4
6
8
10
15
20
25
30
50
0
10
20
30
40
50
60
70
80
90
100
PSO NPSO
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

4
6
8
10
15
20
25
30
50
CONVERGENCE RATE

Car Evaluation
Lenses
Habermans Survival
datasets
58
The avoidance of premature convergence at local
optimal points tend to enhance the results
The selection of local best particles based on
neighbors (lbest) rather than particles own best
(pbest) enhances the accuracy of the rules mined
INFERENCES
59
MODULE 2
Mining AR
using PSO
Mining AR
using CPSO
Mining AR
using NPSO
Mining AR
using PSO
60
MINING AR USING SAPSO
inertia-weight parameter Is added to the velocity equation as
where, w is the inertia weight.

Three self adaptive method of inertia weight is proposed as

=
max
(
max

min
) g/G

( t+1) = (t) (
max

min
) /G

where, g is the generation index representing the current
number of evolutionary generations, and G is a redefined
maximum number of generations. Here, the maximal and
minimal weights
max
and
min
are set to 0.9 and 0.4, based on
experimental study.
61
Dataset
Highest PA achieved within 50 runs of iterations
No weight
(Normal PSO)
w = 0.5 w = 0.7
Lenses 87.5 88.09 84.75
Haberman 87.5 96.07 99.80
Car 96.4 99.88 99.84
POP Care 91.6 98.64 97.91
Zoo 83.3 96.88 98.97
EFFECT OF CHANGING W
Predictive Accuracy Comparison
62
RESULT ANALYSIS
Predictive Accuracy Comparison for Lenses Dataset
50
55
60
65
70
75
80
85
90
95
100
10 20 40 60 80 100
P
r
e
d
i
c
t
i
v
e

A
c
c
u
a
r
c
y

(
%
)

No. of Iterations
PSO
WPSO
SAPSO1
SAPSO2
SACPSO
63
Predictive Accuracy Comparison for Habermans Survival Dataset
RESULT ANALYSIS
50
55
60
65
70
75
80
85
90
95
100
10 20 40 60 80 100
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

No. o Iterations
PSO
WPSO
SAPSO1
SAPSO2
SACPSO
64
Predictive Accuracy Comparison for Post operative patient care
Lenses Dataset
RESULT ANALYSIS
50
55
60
65
70
75
80
85
90
95
100
10 20 40 60 80 100
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

No. of Iterations
PSO
WPSO
SAPSO1
SAPSO2
SACPSO
65
Predictive Accuracy Comparison for Zoo Dataset
RESULT ANALYSIS
87.5
95.45
50 50 50 50
40
50
60
70
80
90
100
10 20 40 60 80 100
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

No. of Iterations
PSO
WPSO
SAPSO1
SAPSO2
SACPSO
66
Predictive Accuracy Comparison for Car Evaluation Dataset
RESULT ANALYSIS
50
55
60
65
70
75
80
85
90
95
100
10 20 40 60 80 100
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

No. of Iterations
PSO
WPSO
SAPSO1
SAPSO2
SACPSO
67
RESULT ANALYSIS
Predictive Accuracy Comparison
70
75
80
85
90
95
100
Lenses Car Evaluation Habermans Survival Po-opert Care Zoo
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

PSO
CPSO
NPSO
WPSO
SAPSO2
SAPSO1
SACPSO
68
INFERENCES
Self adaptive methods perform better than other
methods
In term of computational efficiency SAPSO1
performs better
Setting of appropriate values for the control
parameters involved in these heuristics methods is
the key point to success in these methods


69
MODULE 3


Mining Association rules using Hybrid GA/PSO
(GPSO)
70
RELATED WORKS
S.No Method Application Year

1 Hybrid GA and PSO Global
Optimization
2009
2 Chaotic Hybrid GA
& PSO
Circle
Detection
2010
3 Hybrid GA and PSO Job shop
Scheduling
2010
4 PSO with dynamic
Inertia Weight and
PSO
Classification
Rule Mining
2012
71
HYBRID GA/ PSO (GPSO) MODEL
Genetic
Algorithm
Particle
Swarm
Optimization

Evaluate
Fitness
Upper
Lower
Initial
Population
Ranked
Population
Updated
Population
72
RESULT ANALYSIS
Predictive Accuracy Comparison
Dataset GA PSO GPSO
Lenses 75 87.5 87.5
Habermans
survival
50 90 95
Car
Evaluation
75 94.4 98
Post
operative
Patient
85 83.5 90
Zoo 92.5 95.45 95
73
0
10
20
30
40
50
60
70
80
90
100
GA PSO GPSO
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

10
20
30
40
50
0
10
20
30
40
50
60
70
80
90
100
GA PSO GPSO
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

10
20
30
40
50
Car evaluation
Dataset
Habermans
Survival Dataset
RESULT ANALYSIS
Predictive Accuracy Comparison
74
0
10
20
30
40
50
60
70
80
90
100
GA PSO GPSO
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

10
20
30
40
50
0
10
20
30
40
50
60
70
80
90
100
GA PSO GPSO
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

10
20
30
40
50
Post operative
patient care
Dataset
Lenses Dataset
RESULT ANALYSIS
Predictive Accuracy Comparison
75
0
10
20
30
40
50
60
70
80
90
100
GA PSO GPSO
P
r
e
d
i
c
t
i
v
e

A
c
c
u
r
a
c
y

(
%
)

10
20
30
40
50
Zoo Dataset
RESULT ANALYSIS
Predictive Accuracy Comparison
76
INFERENCES
Generates rule with better predictive accuracy when
compared to GA and PSO.
Global search optimization of GA and powerful stochastic
optimization offered by PSO are both combined in GPSO,
resulting in association rules with consistent accuracy.
77
Genetic Algorithm when used for mining association rules
performs better than existing methods accuracy achieved

Particle swarm optimization when adopted for mining
association rules produces results closer to GA but with
minimum execution time

Variations when introduced in both GA and PSO indicates self
adaptive mechanism performs better than others

The premature convergence being the major drawback of PSO
was handled by introducing inertia weights, chaotic maps,
neighborhood selection adaptive inertia weight

Hybrid GA/PSO model produces consistent results in
comparison with GA and PSO
CONCLUSION
78
WORK TO BE COMPLETED


Balancing with exploration an exploitation in PSO
by self adaptive accelerative coefficients

Breeding Swarms: Adding Mutation into gBest

implementing other measures as Execution time
Comparison, number of rules generated, recall.
79
CONTRIBUTION TO RESEARCH


Chaotic operators based on two maps for mining association
rule

Chaotic self adaption in PSO for mining AR

Hybrid GA/PSO (GPSO) for AR mining


80
Papers Published
CONFERENCES
K.Indira and S.Kanmani, Framework for Comparison of Association Rule Mining
Using Genetic Algorithm, In : International Conference On Computers,
Communication & Intelligence , 2010.
K.Indira and S.Kanmani, Mining Association Rules Using Genetic Algorithm: The
role of Estimation Parameters , In : International conference on advances in
computing and communications, Communication in Computer and Information
Science, Springer LNCS, Volume 190, Part 8, 639-648, 2011
K.Indira, S. Kanmani

, Gaurav Sethia.D, Kumaran.S and Prabhakar.J,

Rule
Acquisition in Data Mining Using a Self Adaptive Genetic Algorithm, In : First
International conference on Computer Science and Information Technology,
Springer LNCS Volume 204, Part 1, 171-178, 2011.
K.Indira, S.Kanmani, Prasanth, Harish and Jeeva, Population Based Search
Methods in Mining Association Rules , In : Third International Conference on
Advances in Communication, Network, and Computing CNC 2012, LNICST pp.
255261, 2012.
81
JOURNAL
K.Indira and S.Kanmani, Performance Analysis of Genetic Algorithm
for Mining Association Rules, IJCSI International Journal of Computer
Science Issues, Vol. 9, Issue 2, No 1, 368-376, March 2012

K.Indira and S.Kanmani, Rule Acquisition using Genetic Algorithm,
accepted for publication in Journal of Computing

K.Indira and S.Kanmani, Enhancing Particle Swarm optimization
using chaotic operators for Association Rule Mining, Elixir Computer
Science & Engineering Journal, 46 ,8563-8566, 2912

K.Indira and S.Kanmani, Association Rule Mining by Dynamic
Neighborhood Selection in Particle Swarm Optimization,
communicated to International journal of Swam Intelligence,
Inderscience Publications

82
JOURNAL


K.Indira and S.Kanmani, Association Rule Mining using Self Adaptive
Particle Swarm Optimization, accepted for publication in
International Journal of Computer Science Issues July Issue.

K.Indira and S.Kanmani, Measures for Improving Premature
Convergence in Particle Swarm Optimization for Association Rule
Mining, communicated to International Journal on Data
Warehousing, IGI press.

K.Indira an S.Kanmani, Mining Association Rules using Hybrid
Genetic Algorithm and Particle Swarm Optimization Algorithm
(GPSO), to communicate


83
TIME SCHEDULE
Feb-12 Apr-12 May-12 Jul-12 Aug-12 Oct-12 Dec-12 Jan-13
GA & PSO IN Rule Mining - Journal
Exporation and Exploitation Avoidance
Role of mutation in hybridisation
Synopsis - Preparation
Thesis - Preparation
84
Jing Li, Han Rui-feng, A Self-Adaptive Genetic Algorithm Based On
Real- Coded, International Conference on Biomedical Engineering
and computer Science , Page(s): 1 - 4 , 2010

Chuan-Kang Ting, Wei-Ming Zeng, Tzu- Chieh Lin, Linkage Discovery
through Data Mining, IEEE Magazine on Computational Intelligence,
Volume 5, February 2010.

Shangping Dai, Li Gao, Qiang Zhu, Changwu Zhu, A Novel Genetic
Algorithm Based on Image Databases for Mining Association Rules, 6th
IEEE/ACIS International Conference on Computer and Information
Science, Page(s): 977 980, 2007

Yamina Mohamed Ben Ali, Soft Adaptive Particle Swarm Algorithm for
Large Scale Optimization, IEEE 2010.

Yan Chen, Shingo Mabu, Kotaro Hirasawa, Genetic relation algorithm
with guided mutation for the large-scale portfolio optimization, Expert
Systems with Applications 38 (2011), 33533363.






REFERENCES
85

Xiaoyuan Zhu, Yongquan Yu, Xueyan Guo, Genetic Algorithm Based
on Evolution Strategy and the Application in Data Mining, First
International Workshop on Education Technology and Computer
Science, ETCS '09, Volume: 1 , Page(s): 848 852, 2009

Hong Guo, Ya Zhou, An Algorithm for Mining Association Rules Based
on Improved Genetic Algorithm and its Application, 3rd International
Conference on Genetic and Evolutionary Computing, WGEC '09,
Page(s): 117 120, 2009

Genxiang Zhang, Haishan Chen, Immune Optimization Based Genetic
Algorithm for Incremental Association Rules Mining, International
Conference on Artificial Intelligence and Computational Intelligence,
AICI '09, Volume: 4, Page(s): 341 345, 2009

Maria J. Del Jesus, Jose A. Gamez, Pedro Gonzalez, Jose M. Puerta,
On the Discovery of Association Rules by means of Evolutionary
Algorithms, from Advanced Review of John Wiley & Sons , Inc. 2011

REFERENCES
86
Junli Lu, Fan Yang, Momo Li, Lizhen Wang, Multi-objective Rule
Discovery Using the Improved Niched Pareto Genetic Algorithm, Third
International Conference on Measuring Technology and Mechatronics
Automation, 2011.
Hamid Reza Qodmanan, Mahdi Nasiri, Behrouz Minaei-Bidgoli, Multi
Objective Association Rule Mining with Genetic Algorithm without
specifying Minimum Support and Minimum Confidence, Expert
Systems with Applications 38 (2011) 288298.
Miguel Rodriguez, Diego M. Escalante, Antonio Peregrin, Efficient
Distributed Genetic Algorithm for Rule Extraction, Applied Soft
Computing 11 (2011) 733743.
R.J. Kuo, C.M. Chao, Y.T. Chiu, Application of particle swarm
optimization to association rule mining, Applied Soft Computing 11
(2011) 326336.
Bilal Alatas , Erhan Akin, Multi-objective rule mining using a chaotic
particle swarm optimization algorithm, Knowledge-Based Systems 22
(2009) 455460.


REFERENCES
87
Mourad Ykhlef, A Quantum Swarm Evolutionary Algorithm for mining
association rules in large databases, Journal of King Saud University
Computer and Information Sciences (2011) 23, 16.
Jing Li, Han Rui-feng, A Self-Adaptive Genetic Algorithm Based On
Real- Coded, International Conference on Biomedical Engineering
and computer Science , Page(s): 1 - 4 , 2010

Miguel Rodriguez, Diego M. Escalante, Antonio Peregrin, Efficient
Distributed Genetic Algorithm for Rule extraction, Applied Soft
Computing 11 (2011) 733743.

Hamid Reza Qodmanan , Mahdi Nasiri, Behrouz Minaei-Bidgoli, Multi
objective association rule mining with genetic algorithm without
specifying minimum support and minimum confidence, Expert Systems
with Applications 38 (2011) 288298.








REFERENCES
88
Xiaoyuan Zhu, Yongquan Yu, Xueyan Guo, Genetic Algorithm Based
on Evolution Strategy and the Application in Data Mining, First
International Workshop on Education Technology and Computer
Science, ETCS '09, Volume: 1 , Page(s): 848 852, 2009.

Junli Lu, Fan Yang, Momo Li1, Lizhen Wang, Multi-objective Rule
Discovery Using the Improved Niched Pareto Genetic Algorithm, 2011
Third International Conference on Measuring Technology and
Mechatronics Automation.

Feng Lu, Yanfeng Ge, LiQun Gao, Self-adaptive Particle Swarm
Optimization Algorithm for Global Optimization, 2010 Sixth
International Conference on Natural Computation (ICNC 2010)

Fevrier Valdez, Patricia Melin, Oscar Castillo, An improved
evolutionary method with fuzzy logic for combining Particle Swarm
Optimization and Genetic Algorithms, Applied Soft Computing 11
(2011) ,26252632.
89

Вам также может понравиться