Ijetr022734 PDF
Ijetr022734 PDF
77 www.erpublication.org
Comparative Study of Various Techniques of Image Segmentation Techniques
B. Preprocessing and Tuning Generally, the text components are aligned horizontally in the
document as compared to the non-text components. Hence,
The various parameters used for the tuning deltaRho, we use the surrounding components also to build the feature
deltaTheta, startTheta, endTheta, connectDistance, vector. Each connected component with its surrounding
pixelCount. The image before getting transformed goes connected area is rescaled to the 40X40 window size for
through a number of stages. The image are pre-processed, generating the context based feature vector. The surrounding
binarized (using Otsu algorithm [4]). The edge detection of context area is not fixed for all connected components but it
the objects is done using various masks. The masks used are is a function of components length (l) and height (h). The
as follow: function is such that, for each connected component the area
of dimensions is 5xl by 2xh. The size of the context based
[ ] [ ] feature vector is 1600.
C. Classification
The technique which is useful in segmentation is shaped IV. CLUSTERING BASED TECHNIQUES
based. These techniques take into consideration the
homogeneity of a particular area (forming a region). The Apart from edge and shaped methods, there are techniques
paper proposed [5] uses discriminative learning connected which are derived from data mining to facilitate the process
component based classification. Here they train a self-tunable of segmentation. The paper [6] proposed the algorithms used
multilayer perceptron (MLP) classifier for distinguishing are K-Means, EM which are useful in terms of segmenting
between text and non-text connected components using shape images.
and context information as a feature vector.
A. K-Means Algorithm
A. Shape of the connected component
K-Means algorithm is an example of unsupervised clustering
In most of the documents, the size of the non-text algorithm. It classifies the input data points into different
components is larger than that of the text components. Thus, clusters based on their Minkowski distance.
size information plays a key role in classification. But it alone
cannot suffice the need of classification and hence we also ( | | ) (2)
use shape of the text and non-text components which can be
learned by the MLP classifier. The algorithm assumes that the bits of the image form a
vector space and tries to clusters them naturally into
Hence for generating the feature vector each connected according to their intensities. The points are clustered around
component is rescaled to a 40X40 pixel window. It is only centroids i i ranging from 1 to k in pursuit of minimizing
downscaling. If the length or height is greater than 40 then it the distance of the data points from the centroids of their
is downscaled to 40 else if it is less than 40 it is fit to the respective clusters. The algorithm uses an iterative approach
center of the window. The advantage of doing so is to to cluster the data points. Here the data points are nothing but
distinguish the shape of the smaller and larger components. the pixel density.
Together with raw rescaled connected component, the shape The algorithm is given below
based feature vector is also composed of four other size based
features: 1. Calculate the histogram graph of the intensities of the pixel
of a particular image.
1. Normalized length - It is the ratio of the length of the
component to the length of the input image. 2. Randomly select k data points that will act as a centroid for
a particular cluster.
2. Normalized height - It is the ratio of the height of the
component to the height of the input image. 3. Follow the given steps again until the cluster a label of the
image does not change anymore.
3. Aspect ratio of a component - It is ratio of length to height
4. Cluster the points based on the metric used for the relative
4. The ratio of the number of foreground pixels to the total change in the intensities from the centroid intensities.
rescaled area. ()
() ()
B. Surrounding context of connected component
5. Compute the new centroid for each of the clusters.
78 www.erpublication.org
International Journal of Engineering and Technical Research (IJETR)
ISSN: 2321-0869, Volume-2, Issue-11, November 2014
()
()
* () + Al competition
Azawi, Faisal test images
* () + Shafait, and circuit
Thomas M. diagrams
The parameter on which the above algorithm is tuned is k. k Breuel
denotes the number of clusters to be formed for given set of Suman
K-means,
data points. The characters in a text are clusters into similar Gray-scaled Expected
Tatiraju, Avi - -
cluster due to the fact that most of the characters are of same images. Maximizati
Mehta
intensities and thus belong to the same cluster. on
B. Expected Maximization
VI. CONCLUSION
When it comes to unsupervised learning, the most
omnipresent algorithm used is Expected Maximization. The
data model is dependent on the hidden variables and the The above survey concludes that remarkable work has been
method depends on computing the maximum a posterior done for image segmentation. But there is more scope for
(MAP) estimate of the parameters. In Expected improvements. Some of the key improvements could be in
Maximization, the steps are performed iteratively till all terms of segmentation of cursive handwriting in images. In
consecutive iterations give the same value. The Expectation conclusion, we hope that this lucid discussion will clarify the
Step (E step) computes the probability of hidden variables approaches and methodologies involved in it and would aid
being observable. The next step i.e. the Maximization Step to the future researchers.
(M step) maximizes the probability of the expected
probability found in the previous step. Now, again the E step ACKNOWLEDGEMENT
and M step are repeated so that the values of the result reach a
constant point. The tuning factor or the parameter is
calculated in the M step are used in the previous step. The Special thanks to Mrs. Khushali Deulkar for guiding us
above explanation can be mathematically expressed as: through the implementation of the algorithm and enriching
the quality of the research.
Given training dataset { ( ) ( ) ( ) } and model ( )
where z is the latent variable, we have: REFERENCES
79 www.erpublication.org