Unit 4 Part 1
Unit 4 Part 1
1. MULTIEXPERT COMBINATION
a. Multiexpert combination methods have base-learners that work in parallel.
b. These methods can in turn be divided into two:
i. In the global approach, also called learner fusion, given an input, all base-learners
generate an output and all these outputs are used. Examples are voting and stacking.
ii. In the local approach, or learner selection, for example, in mixture of experts, there is
a gating model, which looks at the input and chooses one (or very few) of the learners
as responsible for generating the output.
2. MULTISTAGE COMBINATION
a. Multistage combination methods use a serial approach where the next combination
base-learner is trained with or tested on only the instances where the previous
base-learners are not accurate enough.
b. The idea is that the base-learners (or the different representations they use) are sorted in
increasing complexity so that a complex base-learner is not used (or its complex
representation is not extracted) unless the preceding simpler base-learners are not
confident. An example is cascading.
METHOD 1 VOTING
Decision Tree Random Forest Support Vector Machine
Decision Tree Random Forest Support Vector Machine
Max Voting
The max voting method is generally used for classification problems. In this technique,
multiple models are used to make predictions for each data point. The predictions by
each model are considered as a ‘vote’. The predictions which we get from the majority of
the models are used as the final prediction.
Averaging
Similar to the max voting technique, multiple predictions are made for each data point in
averaging. In this method, we take an average of predictions from all the models and use
it to make the final prediction. Averaging can be used for making predictions in
regression problems or while calculating probabilities for classification problems.
Weighted Average
This is an extension of the averaging method. All models are assigned different weights
defining the importance of each model for prediction.
BAGGING
Steps in Bagging
● Bagging is also known as Bootstrap aggregating. It consists of two steps:
bootstrapping and aggregation.
● Bootstrapping
○ Involves resampling subsets of data with replacement from an initial
dataset. In other words, subsets of data are taken from the initial dataset.
○ These subsets of data are called bootstrapped datasets or, simply,
bootstraps.
○ Resampled ‘with replacement’ means an individual data point can be
sampled multiple times. Each bootstrap dataset is used to train a weak
learner.
● Aggregating
○ The individual weak learners are trained independently from each other.
Each learner makes independent predictions.
○ The results of those predictions are aggregated at the end to get the overall
prediction. The predictions are aggregated using either max voting or
averaging.
Detailed Steps
● We test the trained weak learner using the training data. As a result of the testing, some data points
● Each data point with the wrong prediction is sent into the second subset of data, and this subset is
updated.
● Using this updated subset, we train and test the second weak learner.
● We continue with the following subset until the total number of subsets is reached.
● We now have the total prediction. The overall prediction has already been aggregated at each step, so
TECHNICAL
ROUND 1
CEO
TECHNICAL
ROUND 2
APTITUDE
ROUND
Steps
Step 1:
Step 2 : Compute Distance of Every point from
centroids
Step 2 Continued…
C1 (2,10) C2(5,8) C3(1,2) Assigned
Cluster
A1(2,10)
A2(2,5)
A3(8,4)
A4(5,8)
A5(7,5)
A6(6,4)
A7(1,2)
A8(4,9)
Step 3 Update Centroid
Step 4: Repeat Step 2 with new centroids
Steps
DISTANCE BETWEEN TWO POINTS
1. Euclidean
2. Manhattan
3. Cosine
4. Jaccard
Euclidean Distance
Manhattan Distance