0% found this document useful (0 votes)
10 views10 pages

Fuzzy decision Tree classification...

Uploaded by

Sapo Project
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
10 views10 pages

Fuzzy decision Tree classification...

Uploaded by

Sapo Project
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 10

Kotha Ratna Kumari,et al, International Journal of Research in Computer and Communication

technology, IJRCCT, ISSN 2278-5841, Vol 1, Issue 4, September 2012.

Fuzzy Decision Tree Classification Based On The Peano Count Tree


Representation
Kota Rathna Kumari, P.Venkata Kishore, Dr.M.Seetha,
Computer Science Department, JNT University,Hyderabad, Andhra Pradesh, India
{rathnadaniel,parisakishore}@gmail.com

ABSTRACT

Image classification is one of the important tasks in These methods are time consuming and complex
remote sensing image interpretation in which the there by requiring a large amount of memory. There
image pixels are classified. Many organizations have are many Image representation methods like Quad
large quantities of spatial data collected in various Tree, Object Tree, Containment Tree, which take up
application areas; these data collections are growing a lot of space and need systems capable of robust and
rapidly and can therefore be considered as spatial fast segmentation. But an optimal segmentation of an
data streams. The classification tree is made by image cannot be achieved [10]. These representation
recursive partitioning of feature space and is methods cannot be used for complete picture system
implemented by a set of rules that determine the path retrieval and systems using this kind of representation
to be followed. The Peano Count Tree is a spatial are expensive. A need for a novel image
data organization that provides a lossless compressed representation and compression method has
representation of a spatial data set and facilitates discovered a lossless spatial data compression called
efficient classification and other data mining P-Tree.
techniques. Using P-tree structure, fast calculation of
In many areas, large quantities of data are generated
measurements, information gain, can be achieved.
and collected every day, such as supermarket
Another modification aimed at combining symbolic
transactions, phone call records. These data arrive too
decision trees with approximate reasoning is offered
fast to be analyzed or mined in time. Such kinds of
by fuzzy representation. The intent is to exploit
data are called “data streams” [9, 10]. Classifying
complementary advantages of both: popularity in
open ended data streams brings challenges and
applications to learning from examples, high
opportunities since traditional techniques often
knowledge comprehensibility of decision trees and
cannot complete the work as quickly as the data is
ability to deal with inexact and uncertain information
arriving in the stream [9, 10]. Spatial data collected
of fuzzy representation. However, these spatial data
from sensor platforms in space, from airplanes or
sets are too large to be classified effectively in a
other platforms are typically updated periodically.
reasonable amount of time using existing methods.
For example, AVHRR (Advanced Very High
The main objective of this project is to generate
Resolution Radiometer) data is updated every hour or
Fuzzy Decision tree classification based on the Peano
so (8 times each day during daylight hours). Such
Count Tree representation.
data sets can be very large (multiple gigabytes) and
are often archived in deep storage before valuable
Keywords
information can be obtained from them. An objective
Data mining, Classification, Decision Tree Induction,
of spatial data stream mining is to mine such data in
Spatial Data, Data Streams.
near real time prior to deep storage archiving.
1.INTRODUCTION
Classification is one of the important areas of data
Spatial Data refers to information related to a
mining [6,7,8]. In classification task, a training set (or
location anywhere on the earth's surface, and allows
called learning set) is identified for the construction
users to look at an area or geographic feature in
of a classifier. Each record in the learning set has
relation to other areas. Image compression methods
several attributes, one of which, the goal or class
like Chain code, Deflate are suitable for images like
label attribute, indicates the class to which each
monochrome images and those consisting of a
record belongs. The classifier, once built and tested,
reasonable number of large connected components.
186
Kotha Ratna Kumari,et al, International Journal of Research in Computer and Communication
technology, IJRCCT, ISSN 2278-5841, Vol 1, Issue 4, September 2012.

is used to predict the class label of new records that yield quality, and soil attributes such as moisture and
do not yet have a class label attribute value. nitrate levels, etc.). All the values have been scaled to
values between 0 and 255 for simplicity. The pixel
A test set is used to test the accuracy of the classifier. coordinates in raster order constitute the key
The classifier, once certified, is used to predict the attribute. One
class label of future unclassified data. Different can view such data as table in relational form where
models have been proposed for classification, such as each pixel is a tuple and each band is an attribute.
decision trees, neural networks, Bayesian belief
networks, fuzzy sets, and generic models. Among There are several formats used for spatial data, such
these models, decision trees are widely used for as Band Sequential (BSQ), Band Interleaved by Line
classification. We focus on decision tree induction in (BIL) and Band Interleaved by Pixel (BIP). In our
this paper. ID3 (and its variants such as C4.5) [1, 2] previous works [11], we proposed a new format
and CART [4] are among the best known classifiers called bit Sequential Organization (bSQ). Since each
that use decision trees. Other decision tree classifiers intensity value ranges from 0 to 255, which can be
include Interval Classifier [3] and SPRINT [3, 5] represented as a byte, we try to split each bit in one
which concentrate on making it possible to mine band into a separate file, called a bSQ file. Each bSQ
databases that do not fit in main memory by only file can be reorganized into a quadrant-based tree (P-
requiring sequential scans of the data. tree). The example in Figure 1 shows a bSQ file and
Classification has been applied in many fields, such its P-tree.
as retail target marketing, customer retention, fraud
detection and medical diagnosis [8]. Spatial data is a
promising area for classification. In this paper, we
propose a decision tree based model to perform
classification on spatial data streams. We use the
Peano Count Tree (P-tree) structure [11] to build the
classifier.

P-trees [11] represent spatial data bit-by-bit in a


recursive quadrant-by-quadrant arrangement. With
the information in Ptrees, we can rapidly build the In this example, 55 is the count of 1’s in the entire
decision tree. Each new component in a spatial data image (called root count), the numbers at the next
stream is converted to P-trees and then added to the level, 16, 8, 15 and 16, are the 1-bit counts for the
training set as soon as possible. Typically, a window four major quadrants. Since the first and last quadrant
of data components from the stream is used to build is made up of entirely 1-bits (called pure-1
(or rebuild) the classifier. There are many ways to quadrants), we do not need sub-trees for them.
define the window, depending on the data and Similarly, quadrants made up of entirely 0-bits are
application. In this paper, we focus on a fast called pure-0 quadrant. This pattern is continued
classifier-building algorithm. The rest of the paper is recursively. Recursive raster ordering is called Peano
organized as follows. In section 2, we briefly review or Z-ordering in the literature. The process terminates
the spatial data formats and the P-tree structure. In at the leaf level (level-0) where each quadrant is a 1-
Section 3, we detail our decision tree induction row-1-column quadrant. If we were to expand all
classifier using Ptrees. We also walk through an sub-trees, including those pure quadrants, then the
example to illustrate our approach. Performance leaf sequence is just the Peano space-filling curve for
analysis is given in Section 4. Finally, the original raster image.
there is a conclusion in Section 5.
For each band (assuming 8-bit data values), we get 8
2. PEANO COUNT TREE STRUCTURE basic Ptrees, one for each bit positions. For band, Bi,
A spatial image can be viewed as a 2-dimensional we will label the basic P-trees, Pi,1, Pi,2, …, Pi,8,
array of pixels. Associated with each pixel are thus, Pi,j is a lossless representation of the jth bits of
various descriptive attributes, called “bands”. For the values from the ith band. However, Pij provides
example, visible reflectance bands (Blue, Green and more information and re structured to facilitate data
Red), infrared reflectance bands (e.g., NIR, MIR1, mining processes. Some of the useful features of P-
MIR2 and TIR) and possibly some bands of data trees can be found later in this paper or our earlier
gathered from ground sensors (e.g., yield quantity, work [11].
187
Kotha Ratna Kumari,et al, International Journal of Research in Computer and Communication
technology, IJRCCT, ISSN 2278-5841, Vol 1, Issue 4, September 2012.

The basic P-trees defined above can be combined In this paper, we consider the classification of spatial
using simple logical operations (AND, OR and data in which the resulting classifier is a decision tree
COMPLEMENT) to produce Ptrees for the original (decision tree induction). Our contributions include
values (at any level of precision, 1-bit precision, 2-bit  A set of classification-ready data structures
precision, etc.). We let Pb,v denote the Peano Count called Peano Count trees, which are
Tree for band, b, and value, v, where v can be compact, rich in information and facilitate
expressed in 1-bit, 2-bit,.., or 8-bit precision. For classification;
example, Pb,110 can be constructed from the basic P-
trees as:  A data structure for organizing the inputs to
decision tree induction, the Peano count
Pb,110 = Pb,1 AND Pb,2 AND Pb,3’ cube;

where ’ indicates the bit-complement (which is  A fast decision tree induction algorithm,
simply the count complement in each quadrant). This which employs these structures.
is called the value P-tree. The AND operation is
simply the pixel wise AND of the bits. We point out the classifier is precisely the classifier
built by the ID3 decision tree induction algorithm [4].
The data in the relational format can also be The point of the work is to reduce the time it takes to
represented as Ptrees. For any combination of values, build and rebuild the classifier as new data continue
(v1,v2,…,vn), where vi is from band-i, the quadrant- to arrive. This is very important for performing
wise count of occurrences of this tuple of values is classification on data streams.
given by:
3.1 Data Smoothing and Attribute Relevance
P(v1,v2,…,vn) = P1,V1 AND P2,V2 AND … In the overall classification effort, as in most data
AND Pn,Vn mining approaches, there is a data preparation stage
in which the data are prepared for classification. Data
This is called a tuple P-tree. preparation can involve data cleaning (noise
reduction by applying smoothing techniques and
Finally, we note that the basic P-trees can be missing value management techniques). The P-tree
generated quickly and it is only a one-time cost. The data structure facilitates a proximity-based data
logical operations are also very fast [12]. So this smoothing method, which can reduce the data
structure can be viewed as a “data mining ready” and classification time considerably. The smoothing
lossless format for storing spatial data. method is called bottom-up purity shifting. By
replacing 3 counts with 4 and 1 counts with 0 at
3. THE CLASSIFIER level-1 (and making resultant changes on up the tree),
Classification is a data mining technique that the data is smoothed and the P-tree is compressed. A
typically involves three phases, a learning phase, a more drastic smoothing can be effected. The user can
testing phase and an application phase. A learning determine which set of counts to replace with pure-1
model or classifier is built during the learning phase. and which set of counts to replace with pure-0. The
It may be in the form of classification rules, a most important thing to note is that this smoothing
decision tree, or a mathematical formula. Since the can be done almost instantaneously once P-trees are
class label of each training sample is provided, this constructed. With this method it is feasible to actually
approach is known as supervised learning. In smooth data from the data stream before mining.
unsupervised learning (clustering), the class labels
are not known in advance. Another important pre-classification step is relevance
analysis (selecting only a subset of the feature
In the testing phase test data are used to assess the attributes, so as to improve algorithm efficiency).
accuracy of classifier. If the classifier passes the test This step can involve removal of irrelevant attributes
phase, it is used for the classification of new, or redundant attributes. We can build a cube, called
unclassified data tuples. This is the application phase. Peano Cube (P-cube) in which each dimension is a
The classifier predicts the class label for these new band and each band has several values depending on
data samples. the bit precision. For example, for an image with
three bands using 1-bit precision, the cell (0,0,1)
188
Kotha Ratna Kumari,et al, International Journal of Research in Computer and Communication
technology, IJRCCT, ISSN 2278-5841, Vol 1, Issue 4, September 2012.

gives the count of P1’ AND P2’ AND P3. We can  The algorithm stops when all samples for a
determine relevance by rolling-up the P-cube to the given node belong to the same class or when
class label attribute and each other potential decision there are no remaining attributes (or some
attribute in turn. If any of these roll-ups produce other stopping condition).
counts that are uniformly distributed, then that
attribute is not going to be effective in classifying the The attribute selected at each decision tree level is the
class label attribute. one with the highest information gain. The
information gain of an attribute is computed by using
The roll-up can be computed from the basic P-trees the following algorithm.
without necessitating the actual creation of the P-
cube. This can be done by ANDing the P-trees of Assume B[0] is the class attribute; the others are non-
class label attribute with the P-trees of the potential class attributes. We store the decision path for each
decision attribute. Only an estimate of uniformity in node.
the root counts is all that is needed. Better estimates
can be discovered by ANDing down to a fixed depth For example, in the decision tree below (Figure 2),
of the P-trees. For instance, ANDing to depth=1 the decision path for node N09 is “Band2, value
counts provides the rough set of distribution 0011, Band3, value 1000”. We use RC to denote the
information, ANDing at depth=2 provides better root count of a P-tree, given node N’s decision path
distribution information and so forth. Again, the point B[1], V[1], B[2], V[2], … , B[t], V[t], let P-tree
is that P-trees facilitate simple real-time relevance P=PB[1],v[1]^PB[2],v[2]^…^PB[t],v[t]
analysis, which makes it feasible for data streams.

3.2 Classification by Decision Tree Induction


Using P-trees
A Decision Tree is a flowchart-like structure in
which each node denotes a test on an attribute. Each
branch represents an outcome of the test and the leaf
nodes represent classes or class distributions.
Unknown samples can be classified by testing
attributes against the tree. The path traced from root
to leaf holds the class prediction for that sample. The
basic algorithm for inducing a decision tree from the
learning or training sample set is as follows [2, 7]:
Initially the decision tree is a single node
representing the entire training set.
 If all samples are in the same class, this node
becomes a leaf and is labeled with that class
label. In pseudo code, the general algorithm for building
 Otherwise, an entropy-based measure, decision trees is:
"information gain", is used as a heuristic for
selecting the attribute which best separates 1. Check for base cases
the samples into individual classes (the 2. For each attribute a
“decision" attribute).
 A branch is created for each value of the test Find the normalized information gain from
attribute and samples are partitioned splitting on a .
accordingly.
 The algorithm advances recursively to form 3. Let a_best be the attribute with the highest
the decision tree for the sub-sample set at normalized information gain
each partition. Once an attribute has been 4. Create a decision node that splits on a_best
used, it is not considered in descendent
nodes.
189
Kotha Ratna Kumari,et al, International Journal of Research in Computer and Communication
technology, IJRCCT, ISSN 2278-5841, Vol 1, Issue 4, September 2012.

5. Recur on the sub lists obtained by splitting on year’s income and current employment (a very
a_best, and add those nodes as children of simplistic scenario). Assume that each applicant
node indicated the exact income amount (more informative
than just income brackets) and the number of hours
Improvements from ID3 algorithm worked (which is again more informative than the
information whether they worked at all). Each
application was either rejected or accepted (a binary
C4.5 made a number of improvements to ID3. Some
decision). Or, alternatively, each application could
of these are:
have been given a score.
 Handling both continuous and discrete For example, an applicant whose reported income
attributes - In order to handle continuous was $52,000, who was working 30 hours a week, and
attributes, C4.5 creates a threshold and then who was given credit with some hesitation, could
splits the list into those whose attribute become the following training example
value is above the threshold and those that [Inc=52,000][Emp=30] [Credit=0.7]:- weight=1.
are less than or equal to it. Our fuzzy decision tree can also handle fuzzy
 Handling training data with missing attribute examples such as :
values - C4.5 allows attribute values to be [Inc=52,000][Employment=High][Credit=High],
marked as for missing [5]. Missing attribute but for clarity we will not discuss such natural
values are simply not used in gain and extensions here.
entropy calculations.
 Handling attributes with differing costs. 4.2 Example
 Pruning trees after creation - C4.5 goes back In this example the data is a remotely sensed image
through the tree once it's been created and (e.g., satellite image or aerial photo) of an
attempts to remove branches that do not help agricultural field and the soil moisture levels for the
by replacing them with leaf nodes. field, measured at the same time. We use the whole
data set for mining so as to get as better accuracy as
we can. This data are divided into learning and test
data sets. The goal is to classify the data using soil
moisture as the class label attribute and then to use
4. FUZZY DECISION TREES the resulting classifier to predict the soil moisture
levels for future time (e.g., to determine capacity to
4.1 Fuzzy Sample Representation buffer flooding or to schedule crop planting).
Our fuzzy decision tree differs from traditional Branches are created for each value of the selected
decision trees in two respects: it uses splitting criteria attribute and subsets are partitioned accordingly. The
Based on fuzzy restrictions and its inference following training set contains 4 bands of 4-bit data
procedures are different. Fuzzy sets defining the values (expressed in decimal and binary). B1 stands
fuzzy terms used for building the tree are imposed on for soil-moisture. B2, B3, and B4 stand for the
the algorithm. However, we are currently channel 3, 4, and 5 of AVHRR, respectively
investigating extensions of the algorithm for
generating such fuzzy terms, along with the defining FIELD CLASS REMOTELY SENSED
fuzzy sets, either off-line (such as [9]) or on-line COORDS LABEL REFLECTANCES
(such as [1][6]). X Y B1 B2 B3 B4
0,0 0011 0111 1000 1011
For the sake of presentation clarity, we assume crisp 0,1 0011 0011 1000 1111
data form2. For instance, for attributes such as 0,2 0111 0011 0100 1011
Income, features describing an example might be 0,3 0111 0010 0101 1011
such as Income=$23,500.Examples are also 1,0 0011 0111 1000 1011
augmented with “confidence” weights. 1,1 0011 0011 1000 1011
1,2 0111 0011 0100 1011
As an illustration, consider the same previously 1,3 0111 0010 0101 1011
illustrated fuzzy variables Income and Employment, 2,0 0010 1011 1000 1111
and the fuzzy decision Credit. That is, assume that 2,1 0010 1011 1000 1111
applicants applying for credit indicated only last 2,2 1010 1010 0100 1011

190
Kotha Ratna Kumari,et al, International Journal of Research in Computer and Communication
technology, IJRCCT, ISSN 2278-5841, Vol 1, Issue 4, September 2012.

2,3 1111 1010 0100 1011 Advancing the algorithm recursively to each sub-
3,0 0010 1011 1000 1111 sample set, it is unnecessary to rescan the learning set
3,1 1010 1011 1000 1111 to form these sub-sample sets, since the P-trees for
3,2 1111 1010 0100 1011 those samples have been computed.
3,3 1111 1010 0100 1011
The algorithm will terminate with the decision tree:
B2=0010 B1=0111
B2=0011 B3=0100 B1=0111
This learning dataset (Figure 3) is converted to bSQ  B3=1000B1=0011
format. We display the bSQ bit-bands values in their B2=0111 B1=0011
spatial positions, rather than displaying them in 1-
B2=1010 B1=1111
column files. The Band-1 bit-bands are:
B2=1011 B1=0010
B11 B12 B13 B14
0000 0011 1111 1111
0000 0011 1111 1111 6. PERFORMANCE ANALYSIS
0011 0001 1111 0001 Sl. Decision Tree P-Tree based decision Fuzzy based Decision

0111 0011 1111 0011 NO: Tree Tree

Thus, the Band-1 basic P-trees are as follows (tree Kappa Overall Kappa Overall Kappa Overall
pointers are omitted). Coefficient Accuracy Coefficient accuracy Coefficient accuracy
P1,1 P1,2 P1,3 P1,4
5 7 16 11 1 0.3728 54.3100 0.4432 60.9900 0.4771 61.8700
0014 0403 4403
2 0.4134 72.6120 0.6123 79.3100 0.6537 81.9500
0001 0111 0111
3 0.4982 63.4100 0.7235 81.2100 0.7366 83.1420

4 0.2561 56.5300 0.2734 58.8500 02789 59.1300


Then we generate basic P-trees and value P-trees
similarly to B2, B3 and B4. 5 0.3420 47.2850 62.1100 0.4832 64.9540 64.9540

Start with A = B2. Because the node currently


The maximum accuracy is obtained for fuzzy
dealing is the root node, P is the full P-tree. So pi can
be 3/16, 1/4, 1/4, 1/8, 3/16, thus we can calculate approach i.e. 83.14. On observing the above table we
can also conclude that for every input image the
I(P) = 3/16*log2(3/16) + 4/16*log2(4/16) + performance of fuzzy is greater than the Peano
/16*log2(4/16) + 2/16*log2(2/16) + 3/16*log2(3/16) image. Thus the objective of this project is achieved
and this is evident from the accuracy table.
) = 2.281

To calculate E(B2), first P^PA,VA[i] should be all Accuracy Comparison between Peano and fuzzy
the value P-trees of B2. Then I(P^PA,VA[i]) can be Decision Tree techniques
calculated by ANDing all the B2 value P-trees and The user’s accuracy method is used for computing
the accuracy. The fuzzy approach based classified
B1 value P-trees. Finally we get E(B2)=0.656 and
images is taken as the reference, based on which the
Gain(B2)=1.625.
accuracies of the other method is compared.
Likewise, the Gains of B3 and B4 are computed:
Gain(B3) = 1.084 , Gain(B4) = 0.568. Thus, B2 is
selected as the first level decision attribute. Accuracy Analysis
In many instances, the stratified random sampling
strategy is the most useful tool to use. In this case, the
Branches are created for each value of B2 and
samples are partitioned accordingly. map area is stratified based on either a systematic
breakdown followed by a random sample design in
B2=0010 Sample_Set_1
each of the systematic sub areas, or alternatively
B2=0011 Sample_Set_2
through the application of a random sample within
B2=0111 Sample_Set_3 each of the map classes [21]. The use of this
B2=1010 Sample_Set_4 approach will ensure that one has an adequate cover
B2=1011 Sample_Set_5 for the entire map as well as generating a sufficient

191
Kotha Ratna Kumari,et al, International Journal of Research in Computer and Communication
technology, IJRCCT, ISSN 2278-5841, Vol 1, Issue 4, September 2012.

number of samples for each of the classes on the But just because 83% classifications were accurate
map. overall, it does not mean that each category was
successfully classified at that rate.
The diagonal elements tally the number of pixels
classified correctly in each class. Comparison of Classified Images
An overall measure of classification accuracy is as
given below, The following table lists out the Input Images along
with their respective classified images.
Overall Accuracy =Total number of correct
Each Input Image is classified based on the Decision
classifications Tree, Peano Decision tree and Fuzzy Decision Tree
The maximum accuracy is obtained for fuzzy
Total number of
approach i.e. 83.14. On observing the above figure,
classifications

This in this example amounts to 35+37+41 or 83% for every input image the performance of fuzzy is
136 greater than the other images. Thus the objective of
this project is achieved and this is evident from the
accuracy table.

192
Kotha Ratna Kumari,et al, International Journal of Research in Computer and Communication
technology, IJRCCT, ISSN 2278-5841, Vol 1, Issue 4, September 2012.

Input Image Decision Tree Peano Decision Fuzzy Decision

Tree Tree

Fig 3 : Classified Images

193
Kotha Ratna Kumari,et al, International Journal of Research in Computer and Communication
technology, IJRCCT, ISSN 2278-5841, Vol 1, Issue 4, September 2012.

GA for Patter Recognition, S. Pal & P.


Wang (eds.), CRC Press, pp. 253-282,1995.
[8] C.Z. Janikow, “A Genetic Algorithm
7.CONCLUSION Method for Optimizing Fuzzy Decision
Trees”, Information Sciences, 89(3-4), pp.
275-296, March 1996.
This project classifies the input images according to
[9] D. Drinkov, H. Hellendoorn & M. Rein
the fuzzy Decision tree based on peano count tree for
frank, An Introduction to Fuzzy Control,
performing pixel-based image classification. For this
1996.
purpose, LISS-III satellite images were used. An
[10] Ding Q, Ding Q, and Perrizo W, Decision
image is classified according to the fuzzy based
tree classification of spatial data streams
approach where each feature is divided into five
using Peano count trees, In Proceedings of
ranges being very small, small, medium, large and
the ACM Symposium on Applied
very large. The Fuzzy Decision Tree classification
Computing, Madrid, Spain: 413–7, 2002.
gives better performance compared to Decision Tree
[11] D. Michie, D. Spiegelhalter, and C. Taylor,
classification and Peano Decision Tree classification
Machine Learning, Neural and Statistical
techniques. For the fuzzy classification a total of 150
Classification, 1994.
pixels, from all the 5 input images are used. Then the
[12] Domingos, P. and Hulten, G., “Mining high-
two methods Peano Decision Tree and Fuzzy
speed data streams”, Proceedings of ACM
Decision Tree are compared such that Fuzzy
SIGKDD, 2000.
Decision Tree gives better accuracy than both the
[13] D. Wolpert, Stacked generalization, Neural
methods taken individually. The accuracy table
Networks, 5: 241 -259, 1992.
proves this objective of the project.
[14] H. Schee, The Analysis of Variance, In
International Joint Conference on Neural
Networks, 1959.
REFERENCES [15] I. Enbutsu, K. Baba &N. Hara, “Fuzzy Rule
Extraction from a Multilayered Neural
[1] Amlendu Roy, William Perrizo, "Peano Network”, In International Joint Conference
Count Tree Technology”, Deriving High on Neural Networks, pp. 461-465, 1991.
Confidence Rules from Spatial Data using [16] I. Kononenko and S.J. Hong, Attribute
Peano Count Trees, LNCS 2118, July 2001. Selection for Modeling, 1997.
[2] Buntine. W, Learning Classification Trees, [17] J. Hosking, E. Pednault, and M. Sudan, A
Statistics and Computing, 2:63-73, 1992. Statistical Perspective on Data Mining,
[3] C. Apte, F. Damerau, and S. Weiss, 1997.
Automated Learning of Decision Rules for [18] Jiawei Han, Micheline Kamber, “Data
Text Categorization, ACM Transactions on Mining: Concepts and Techniques”, 2001.
Information Systems, 12(3):233-251, July [19] J. Quinlan, Combining instance-based and
1994. model-based learning, In International
[4] C. Apte and S.J. Hong, Predicting Equity Conference on Machine Learning, pages
Returns from Securities Data with Minimal 236-243, 1993.
Rule Generation, In Advances in Knowledge [20] J. Rissanan, Stochastic Complexity in
Discovery, AAAI Press / The MIT Press, Statistical Inquiry, World Scientific Series in
pages 541-560, 1995. Computer Science, 15, 1989.
[5] C.Z. Janikow, “Fuzzy Processing in [21] J.R. Quinlan, C4.5: Programs for Machine
Decision Trees”, In Proceedings of the Sixth Learning, International Conference on
International Symposium on Artificial Machine Learning, 1993.
Intelligence, pp. 360-367, 1993. [22] L. Breiman, J.H. Friedman, R.A. Olsen &
[6] C.Z. Janikow, “Learning Fuzzy Controllers C.J. Stone, Classification and Regression
by Genetic Algorithms”, In Proceedings of Trees, 1984.
the 1994 ACM Symposium on Applied [23] L. Breiman. Bagging predictors, Machine
Computing, pp. 232-236, 1994. Learning, Classification and Regression
[7] C.Z. Janikow, “A Genetic Algorithm Trees24:123-140, 1996.
Method for Optimizing the Fuzzy [24] L. Breiman, J. Friedman, R. Olshen, and C.
Component of a Fuzzy Decision Tree”, In Stone, Classification and Regression Trees,

194
Kotha Ratna Kumari,et al, International Journal of Research in Computer and Communication
technology, IJRCCT, ISSN 2278-5841, Vol 1, Issue 4, September 2012.

1984. [37] S. Weiss and N. Indurkhya, Optimized Rule


[25] S. M. Weiss and C. A. Kulikowski, Induction, IEEE EXPERT, 8(6):61-69,
“Computer Systems that Learn: December 1993.
Classification and Prediction Methods from [38] S. Weiss and N. Indurkhya, Rule-based
Statistics, Neural Nets, Machine Learning, Machine Learning Methods for Functional
and Expert Systems”, 1991. Prediction,Journal of Artificial Intelligence
[26] Mingers J, “An Empirical Comparison of Research, 3: 383-403, 1995.
Prunning Methods for Decision Tree [39] S.Weiss and C.A. Kulikowski, Computer
Induction”, Machine Learning 4, pp. 227- Systems That Learn. Morgan Kaufmann,
243, 1989. 1991.
[27] M. Mizumoto, “Fuzzy Controls Under [40] T.G. Dietterich, H. Hild & G. Bakiri, “A
Various Fuzzy Reasoning Methods”, Comparative Study of ID3 and Back
Information Science, 45, pp. 129-151, 1988. propagation for English Text-to-Speach
[28] Pal N R and Chakraborty, Fuzzy rule Mapping”, Proceedings of the International
extraction from ID3-type decision trees for Conference on Machine Learning, 1990.
real data, IEEE Transactions on Systems
Man and Cybernetics Part B-Cybernetics,
31: 745–54, S 2001.
[29] P. Clark and T. Niblett, The CN2 Induction
Algorithm, Machine Learning, 3:261-283,
1989.
[30] P. Smyth and R. Goodman, Information The
theoretic Approach to Rule Induction from
Data- bases, IEEE Transactions on
Knowledge and Data Engineering, 4(4):301-
316, August 1992.
[31] R. Gallion, D.C. St.Clair, C. Sabharwal &
W.E. Bond, “Dynamic ID3: A Symbolic
Learning Algorithm for Many-Valued
Attribute Domains”, In Proceedings of the
1993 Symposium on Applied Computing,
ACM Press, pp. 14-20, 1993.
[32] R. Jager, Fuzzy Logic in Control, Ph.D.
dissertation, Technische Universities Delft,
pp. 480-485, 1994.
[33] R. Michalski, I. Mozetic, J. Hong, and N.
Lavrac, The Multi-Purpose Incremental
Learning System AQ15 and its Testing
Application to Three Medical Domains, In
Proceedings of the AAAI-86, pages 1041-
1045, 1986.
[34] Shekhar S, Zhang P, Huang Y, and
Vatsavai, Trends in spatial data mining,In
Kargupta H, Joshi A, Sivakumar K and
Yesha Y (eds) Data Mining: Next
Generation Challenges and Future
Directions, London, AAAI Press: 357–80,
2003.
[35] S.J. Hong, R-MINI, An Iterative Approach
for Generating Minimal Rules from
Examples, IEEE Transactions on
Knowledge and Data Engineering, 1997.
[36] Batty M, Spatial entropy, Geographical
Analysis 6: 1–31, 1974.

195

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy