An Insight Into Machine Learning Techniq
An Insight Into Machine Learning Techniq
Published By:
Retrieval Number: I10550789S19/19©BEIESP 342 Blue Eyes Intelligence Engineering
DOI: 10.35940/ijitee.I1055.0789S19 & Sciences Publication
An Insight Into Machine Learning Techniques For Predictive Analysis And Feature Selection
connections between machine learning and “signal learning techniques and statistical techniques and concluded
processing techniques for big data processing”[11]. Landset that the machine learning techniques can predict software
et al. discussed the merits and demerits of three different fault proneness[14].
processing paradigms and also compared the engines that
implement them[12]. The work related to predictive modelling, along with their
objective and techniques involved are summarized in Table
Lykourentzou et al. proposed a dropout prediction method I.
for e-learning courses using “feed-forward neural
networks”, “support vector machines” and “probabilistic
ensemble simplified fuzzy ARTMAP”[19]. Ruchika
Malhotra compared the performance of different machine
1 Clustering “A new soft 2019 “Hoang Nguyen, Xuan- To develop a hybrid model based on
computing Nam Bui, Quang-Hieu “Hierarchical K-means clustering (HKM)”
model for Tran, Ngoc-Luan Mai” and “Cubist algorithm (CA)”, code name
estimating and “HKM-CA model” to predict “blast-induced
controlling PPV on-site to control undesirable effects on
blast-produced the surrounding environment”. This model
ground vibration improved the accuracy of the CA model with
based on a RMSE of 0.475 and MAE of 0.373.
Hierarchical K-
means clustering
and Cubist
algorithms”
2 Linear “A machine 2018 “Ahmed Abdelaziz, To propose a model for Health Care Services
Regression learning model Mohamed Elhoseny, based on cloud environment using “Parallel
and Neural for improving Ahmed S. Salama, A.M. Particle Swarm Optimization (PPSO)” in
Network healthcare Riad” order to optimize the Virtual Machine’s
services on selection and for chronic kidney disease
cloud computing diagnosis
environment”
3 Logistic “Data mining 2018 “Concepción Burgos, To predict if a student will drop out in his
Regression for modelling María L. Campanario, course or not
students’ David de la Peña, Juan
performance: A A. Lara, David Lizcano,
tutoring action María A. Martínez”
plan to prevent
academic
dropout”
4 Logistic “A new hybrid 2018 “Arno De Caigny, To design a new algorithm called “logit leaf
Regression classification Kristof Coussement, model” which is a hybrid of “Decision trees”
algorithm for Koen W. De Bock” and “logistic regression” algorithms in
customer churn customer churn prediction
prediction based
on logistic
regression and
decision trees”
Published By:
Blue Eyes Intelligence Engineering
Retrieval Number: I10550789S19/19©BEIESP 343 & Sciences Publication
DOI: 10.35940/ijitee.I1055.0789S19
International Journal of Innovative Technology and Exploring Engineering (IJITEE)
ISSN: 2278-3075, Volume-8, Issue-9S, July 2019
5 Clustering “Cloud based 2018 “Shakeel, P.M., Baskar, To compare the efficiency of different
framework for S., Dhulipala” clustering techniques in order to diagnose
diagnosis of which gender and age group are mostly
diabetes mellitus affected by diabetes.
using K-means
clustering”
6 Decision “A new hybrid 2018 “Arno De Caigny, To propose a hybrid algorithm, “the logit
Tree classification Kristof Coussement, leaf model (LLM)”, based on “Decision
algorithm for Koen W. De Bock” trees” and “Logistic regression” for customer
customer churn churn prediction.
prediction based
on logistic
regression and
decision trees”
7 Logistic “Early detection 2017 “Anne-Sophie Hoffait, To design the prediction algorithms with
Regression of university Michaël Schyns” high accuracy using “random forest”,
students with “logistic regression” and “artificial neural
potential network” algorithms.
difficulties”
9 Linear “Comparison of 2016 “Jean-Loup Loyer, Elsa To compare the performance of five
Regression Machine Henriques, statistical models for the prediction of
Learning MihailFontul, Steve manufacturing cost of jet engine
methods applied Wiseall” components, during the early design phase.
to the estimation The results proved “Gradient Boosted Trees”
of and “Support Vector Regression” to be very
manufacturing efficient techniques
cost of jet
engine
components”
10 Decision “Towards the 2016 “Mrinal Pandey, S. To propose integrated multiple classifiers for
Tree integration of Taruna” predicting students’ academic performance
multiple using “Decision Tree”, “K-Nearest
classifier Neighbour” and “Aggregating One-
pertaining to the Dependence Estimators (AODE)”
Student's
performance
prediction”
11 Logistic “Using Machine 2016 “Hiba Asri, Hajar To compare the performance of machine
Regression Learning Mousannif, Hassan Al learning algorithms: “Support Vector
Algorithms for Moatassime, Thomas Machine”, “Decision Tree”, “Naive Bayes”
Breast Cancer Noel” and “k Nearest Neighbors” on the Wisconsin
Risk Prediction Breast Cancer (original) datasets. The results
Published By:
Retrieval Number: I10550789S19/19©BEIESP 344 Blue Eyes Intelligence Engineering
DOI: 10.35940/ijitee.I1055.0789S19 & Sciences Publication
An Insight Into Machine Learning Techniques For Predictive Analysis And Feature Selection
and Diagnosis” show that Support Vector Machine gives the
highest accuracy of 97.13%
12 Neural “An empirical 2016 “Ruchika Malhotra” To predict defective classes using 18
Network framework for machine-learning techniques. “The proposed
defect prediction framework has been applied to seven
using machine application packages of well known, widely
learning used Android operating system viz. Contact,
techniques with MMS, Bluetooth, Email, Calendar, Gallery2
Android and Telephony”.
software”
13 Decision “Wrapper 2015 “RattanawadeePanthong, To perform optimization based on ensemble
Tree Feature Subset AnongnartSrivihok” algorithms namely “Bagging” and
Selection for “AdaBoost” by subset evaluations which are
Dimension performed using two classifiers: “Decision
Reduction Tree” and “Naïve Bayes”. “This study shows
Based on that the search technique using SFS based on
Ensemble the bagging algorithm using Decision Tree
Learning obtained better results in average accuracy
Algorithm” (89.60%) than other methods”
14 Clustering “Randomized 2015 “C. Boutsidis, A. To present feature extraction methods based
Dimensionality Zouzias, M. W. on random projections and fast approximate
Reduction fork- Mahoney and P. SVD factorizations alongwith the feature
Means Drineas” selection method for k-means clustering
Clustering”
A. Linear Regression
III. MACHINE LEARNING TECHNIQUES Linear regression is a supervised machine learning algorithm that
is used to model a target value based on the given inputs. It is
Machine learning technique basically works by training an normally used for forecasting and finding the relation between
algorithm on a large data set, also called training set, which dependent and independent variables. If the input variable is u
has known output[11][12]. Later the algorithm is tested on a and the output variable is v, the relation between these two
test set to get the results[19][14]. Machine learning variables can be depicted using
approach can be classified as supervised learning and a linear equation
unsupervised learning. v= α1 + α2.u
The aim is to find the best value
of α1 & α2 so that we get the
Published By:
Blue Eyes Intelligence Engineering
Retrieval Number: I10550789S19/19©BEIESP 345 & Sciences Publication
DOI: 10.35940/ijitee.I1055.0789S19
International Journal of Innovative Technology and Exploring Engineering (IJITEE)
ISSN: 2278-3075, Volume-8, Issue-9S, July 2019
best fit line. For this purpose, the cost function is the Root Mean assigned to the closest cluster centroid. This iteration goes on
Square Error(RMSE), which minimizes the difference between till no more improvements are possible and we reach the
the actual value of v and the predicted value of v. global optima.
Published By:
Retrieval Number: I10550789S19/19©BEIESP 346 Blue Eyes Intelligence Engineering
DOI: 10.35940/ijitee.I1055.0789S19 & Sciences Publication
An Insight Into Machine Learning Techniques For Predictive Analysis And Feature Selection
There is one input layer and one output layer. The design of
input and output layers is application specific. For example,
in a sentimental analysis, the input can be the sentence or
paragraphs to be analysed. The output could be a classifier,
which classifies it as neutral, negative or positive. It could
be designed for ranking (viz. from 1 to 5) on various aspects
like anger or happiness. The inner layers are called hidden
layers. In fig. 2, there is one hidden layer. The input and
hidden layer each have 4 units.
F. Decision Trees On the other side, Decision trees are not appropriate for
Decision tree is a supervised machine learning tool used for estimation tasks, they are prone to errors in classification
predictive modelling. It can be used for classification as well as problems where there are many classes and a small data set.
regression tasks[7]. Another name for decision trees is They can also be computationally expensive to train.
CART(Classification and Regression Trees). It is a tree like
structure that can be used to visually represent the flow of IV. FEATURE SELECTION
decision making. It is one of the simplest machine learning
algorithms. The rules in decision trees are in the form of if- With the availability of huge data sets, feature selection as a
then-else statements. The rules get complex as we go deep pre-processing step to machine learning is highly effective
in the tree and the deeper the tree, fitter the model. The in removing unimportant and redundant data, improving
nodes represent the place where we pick a feature and ask a result comprehensibility and increasing learning
question; the edges represent the answers to that question; accuracy[9][13]. Feature selection is a technique in which
and the leaf nodes represent the actual output or target the most influential features/attributes from a data set are
label. E.g. Fig. 3 shows an example of decision tree for selected for predictive analysis[1]. The irrelevant and
classifying whether a flower is Iris-Sentosa, Iris-versicolor or Iris- redundant features are eliminated from the dataset and a
Verginica based on petal width and petal length. subset of features is kept. Some supervised machine
learning algorithms have built-in feature selection like
Random Forests and Regularized Regression. Feature
selection can be supervised or unsupervised or a
combination of both[6][17]. The two most commonly used
feature selection methods are:
Published By:
Blue Eyes Intelligence Engineering
Retrieval Number: I10550789S19/19©BEIESP 347 & Sciences Publication
DOI: 10.35940/ijitee.I1055.0789S19
International Journal of Innovative Technology and Exploring Engineering (IJITEE)
ISSN: 2278-3075, Volume-8, Issue-9S, July 2019
95% of observations are for women who are 40-year- 3. Ajmer Singh, Rajesh Bhatia, Anita Singhrova, “Taxonomy of
old, then the features 'Age' and 'Gender' can be easily machine learning algorithms in software fault prediction
eliminated without any major loss of using object oriented metrics”, Procedia Computer Science,
information[21][10]. Normalization of features is a very Volume 132, 2018, Pages 993-1001, ISSN 1877-0509,
https://doi.org/10.1016/j.procs.2018.05.115.
important step because variance is dependent on scale.
4. Arno De Caigny, Kristof Coussement, Koen W. De Bock, “A
Intuition is the base for applying variance thresholds: new hybrid classification algorithm for customer churn
features that remain same do not add much information prediction based on logistic regression and decision trees”,
to the data. It is a relatively safe and easy way for European Journal of Operational Research, Volume 269,
dimensionality reduction. It can be implemented easily Issue 2, 2018, Pages 760-772, ISSN 0377-2217,
in Python/R programming language https://doi.org/10.1016/j.ejor.2018.02.009.
5. Aggarwal D., “Sentiment Analysis: An insight into
Correlation Thresholds: Correlation thresholds remove Techniques, Application and Challenges”, International
the highly correlated features (i.e. the features whose Journal of Computer Science and Engineering, Volume 6,
value change similarly to other's). These features do not Issue 5, 2018, pages 697-703,
add much to the information. E.g , if you had a weather https://doi.org/10.26438/ijcse/v6i5.697703
dataset with 'Temperature (Celsius)' and 'Temperature 6. RaziehSheikhpour, Mehdi Agha Sarram, Sajjad Gharaghani,
(Fahrenheit)' as different features, one of these can be Mohammad Ali ZareChahooki, “A Survey on semi-
supervised feature selection methods”, Pattern Recognition,
safely removed. The question arises which one of these
Volume 64, 2017, Pages 141-158, ISSN 0031-3203,
features should be removed. For this purpose, all pair- https://doi.org/10.1016/j.patcog.2016.11.003.
wise correlations need to be calculated. Then, if the 7. Anne-Sophie Hoffait, Michaël Schyns, “Early detection of
correlation> threshold, then the feature, that has greater university students with potential difficulties”, Decision
mean absolute correlation with rest of the features, Support Systems, Volume 101, 2017, Pages 1-11, ISSN
should be removed.Intuition is the base for applying 0167-9236, https://doi.org/10.1016/j.dss.2017.05.003.
correlation thresholds also: features that do not change 8. Chuan Liu, Wenyong Wang, Qiang Zhao, Xiaoming Shen,
much gives redundant information[22][8]. The tricky Martin Konan, “A new feature selection method based on a
part is that the threshold value needs to be set manually validity index of feature subset, Pattern Recognition Letters,
Volume 92, 2017, Pages 1-8, ISSN 0167-8655,
because if the threshold is set very low, the useful
https://doi.org/10.1016/j.patrec.2017.03.018.
information may be dropped. So, the algorithms 9. S. Sasikala, S. Appavu alias Balamurugan, S. Geetha, “Multi
with built-in feature selection method are a better Filtration Feature Selection (MFFS) to improve
choice. It can be implemented easily in Python / R discriminatory ability in clinical data set”, Applied
programming language. Computing and Informatics, Volume 12, Issue 2, 2016, Pages
117-127, ISSN 2210-8327,
V. CONCLUSION https://doi.org/10.1016/j.aci.2014.03.002.
10. Jianyu Miao, LingfengNiu, “A Survey on Feature Selection”,
Procedia Computer Science, Volume 91, 2016, Pages 919-
Predictive analysis is a process that overlaps with machine
926, ISSN 1877-0509,
learning and both go hand-in-hand. This paper tells about https://doi.org/10.1016/j.procs.2016.07.111.
the various supervised and unsupervised machine learning 11. Junfei Qiu, Qihui Wu, Guoru Ding, Yuhua Xu and
algorithms used for predictive analytics.The two most Shuo Feng, “A survey of machine learning for big data
commonly methods used for feature selection are also processing”, EURASIP Journal on Advances in Signal
discussed. Processing, 2016:67, https://doi.org/10.1186/s13634-016-
0355-x
12. Sara Landset, TaghiM. Khoshgoftaar, Aaron N. RichterEmail
ACKNOWLEDGEMENT
author and Tawfiq Hasanin, “A survey of open source tools
for machine learning with big data in the Hadoop ecosystem”,
The authors are highly grateful to the Principal, Journal of Big Data, 2015, 2:24,
Management and Department of Computer Science and https://doi.org/10.1186/s40537-015-0032-1
Engineering of JSS Academy of Technical Education, 13. RattanawadeePanthong, AnongnartSrivihok, ”Wrapper
Noida, Uttar Pradesh to provide complete support in Feature Subset Selection for Dimension Reduction Based on
carrying out the research work and writing this paper. Ensemble Learning Algorithm”, Procedia Computer Science,
Volume 72, 2015, Pages 162-169, ISSN 1877-0509,
https://doi.org/10.1016/j.procs.2015.12.117.
REFERENCES 14. Ruchika Malhotra, “A systematic review of machine learning
techniques for software fault prediction”, Applied Soft
1. Jie Cai, Jiawei Luo, Shulin Wang, Sheng Yang, “Feature Computing, Volume 27, 2015, Pages 504-518, ISSN 1568-
selection in machine learning: A new perspective”, 4946, https://doi.org/10.1016/j.asoc.2014.11.023.
Neurocomputing, Volume 300, 2018, Pages 70-79, ISSN 15. Konstantina Kourou, Themis P. Exarchos, Konstantinos P.
0925-2312, https://doi.org/10.1016/j.neucom.2017.11.077 Exarchos, Michalis V. Karamouzis, Dimitrios I. Fotiadis,
2. Concepción Burgos, María L. Campanario, David de la Peña, “Machine learning applications in cancer prognosis and
Juan A. Lara, David Lizcano, María A. Martínez, “Data prediction”, Computational and Structural Biotechnology
mining for modeling students’ performance: A tutoring action Journal, Volume 13, 2015, Pages 8-17, ISSN 2001-0370,
plan to prevent academic dropout”, Computers & Electrical
Engineering, Volume 66, 2018, Pages 541-556, ISSN 0045-
7906, https://doi.org/10.1016/j.compeleceng.2017.03.005.
Published By:
Retrieval Number: I10550789S19/19©BEIESP 348 Blue Eyes Intelligence Engineering
DOI: 10.35940/ijitee.I1055.0789S19 & Sciences Publication
An Insight Into Machine Learning Techniques For Predictive Analysis And Feature Selection
AUTHORS PROFILE
Deepti Aggarwal has received her B. Tech.(CSE)
from MDU, Rohtak, M.Tech.(CSE) from
Rajasthan Vidyapeeth, Udaipur and pursuing
Ph.D. from Jaipur National University, Jaipur.
She has overall teaching experience of more than
18 years and currently working as Assistant
Professor at JSS Academy of Technical
Education, Noida since 2005. She has 4 papers in
international journals and conference. She has
written a book on ‘Computer Organisation’. Her
area of interest is machine learning, data mining,
operating system and compiler design.
Published By:
Blue Eyes Intelligence Engineering
Retrieval Number: I10550789S19/19©BEIESP 349 & Sciences Publication
DOI: 10.35940/ijitee.I1055.0789S19