Self-Organizing Maps (SOM), developed by Kohonen, are a type of unsupervised neural network used for dimensionality reduction and data visualization by creating a low-dimensional representation of high-dimensional data. They utilize competitive learning techniques, involving competition, cooperation, and adaptation among neurons, to preserve topological properties of the input space. SOMs have applications in various fields, including seismic facies analysis, text clustering, and exploratory data analysis, though they have limitations in handling categorical data and slow training times.
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0 ratings0% found this document useful (0 votes)
15 views31 pages
Self Organizing Maps
Self-Organizing Maps (SOM), developed by Kohonen, are a type of unsupervised neural network used for dimensionality reduction and data visualization by creating a low-dimensional representation of high-dimensional data. They utilize competitive learning techniques, involving competition, cooperation, and adaptation among neurons, to preserve topological properties of the input space. SOMs have applications in various fields, including seismic facies analysis, text clustering, and exploratory data analysis, though they have limitations in handling categorical data and slow training times.
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 31
Self-Organizing Maps
Introduction
• Neural Networks use processing, inspired by the
human brain, as a basis to develop algorithms that can be used to model and understand complex patterns and prediction problems. • There are several types of neural networks and each has its own unique use. • The Self Organizing Map (SOM) is one such variant of the neural network, also known as Kohonen’s Map. Self-Organizing Maps
• A self-organizing map is also known as SOM and
it was proposed by Kohonen. • It is an unsupervised neural network that is trained using unsupervised learning techniques to produce a low dimensional, discretized representation from the input space of the training samples, known as a map and is, therefore, a method to reduce data dimensions. Self-Organizing Maps
• Self-Organizing Maps are very di)erent from
other arti*cial neural networks as they apply competitive learning techniques unlike others using error-correction learning methods such as backpropagation with gradient descent, and use a neighbourhood function to preserve all the topological properties within the input space. Self-Organizing Maps
• Self-Organizing Maps were initially only being
used for data visualization, but these days, it has been applied to di)erent problems, including as a solution to the Traveling Salesman Problem as well. • Map units or neurons usually form a two- dimensional space and hence a mapping from high dimensional space onto a plane is created. Self-Organizing Maps
• The map retains the calculated relative distance between
the points. Points closer to each other within the input space are mapped to the nearby map units in Self- Organizing Maps. • Self-Organizing Maps can thus serve as a cluster analyzing tool for high dimensional data. • Self-Organizing Maps also have the capability to generalize. • During generalization, the network can recognize or characterize inputs that it has never seen as data before. • New input is taken up with the map unit and is therefore mapped. Uses of Self-Organizing Maps
• Self-Organizing Maps provide an advantage in
maintaining the structural information from the training data and are not inherently linear. • Using Principal Component Analysis on high dimensional data may just cause loss of data when the dimension gets reduced into two. • If the data comprises a lot of dimensions and if every dimension preset is useful, in such cases Self-Organizing Maps can be very useful over PCA for dimensionality reduction. Uses of Self-Organizing Maps
• Seismic facies analysis generates groups based
on the identi*cation of di)erent individual features. • This method *nds feature organizations in the dataset and forms organized relational clusters. Uses of Self-Organizing Maps
• However, these clusters sometimes may or may
not have any physical analogs. • Therefore a calibration method to relate these clusters to reality is required and Self- Organizing Maps do the job. • This calibration method de*nes the mapping between the groups and the measured physical properties. Uses of Self-Organizing Maps
• Text clustering is another important
preprocessing step that can be performed through Self-Organizing Maps. • It is a method that helps to verify how the present text can be converted into a mathematical expression for further analysis and processing. • Exploratory data analysis and visualization are also the most important applications of Self- Organizing Maps. Uses of Self-Organizing Maps SOM Training
• SOM doesn’t use backpropagation with SGD to
update weights, this type of unsupervised arti*cial neural network uses competetive learning to update its weights. • Competitive learning is based on three processes : – Competition – Cooperation – Adaptation Competition
• As we said before each neuron in a SOM is assigned a
weight vector with the same dimensionality as the input space. • In the example below, in each neuron of the output layer we will have a vector with dimension n. • We compute distance between each neuron (neuron from the output layer) and the input data, and the neuron with the lowest distance will be the winner of the competetion. • The Euclidean metric is commonly used to measure distance. Competition Cooperation
• We will update the vector of the winner neuron
in the *nal process (adaptation) but it is not the only one, also it’s neighbor will be updated. • How do we choose the neighbors ? • To choose neighbors we use neighborhood kernel function, this function depends on two factor : time ( time incremented each new input data) and distance between the winner neuron and the other neuron (How far is the neuron from the winner neuron). Cooperation
• The image below show us how the winner
neuron’s ( The most green one in the center) neighbors are choosen depending on distance and time factors. Adaptation
• After choosing the winner neuron and it’s
neighbors we compute neurons update. • Those choosen neurons will be updated but not the same update, more the distance between neuron and the input data grow less we adjust it like shown in the image below : Adaptation Adaptation
• The winner neuron and it’s neighbors will be
updated using this formula: Adaptation
• This learning rate indicates how much we want
to adjust our weights. • After time t (positive in*nite), this learning rate will converge to zero so we will have no update even for the neuron winner . Adaptation
• The neighborhood kernel depends on the
distance between winner neuron and the other neuron (they are proportionally reversed : d increase make h(t) decrease) and the neighborhood size wich itself depends on time ( decrease while time incrementing) and this make neighborhood kernel function decrease also. Full SOM Algorithm Architecture
• Self-Organizing Maps consist of two important
layers, the *rst one is the input layer, and the second one is the output layer, which is also known as a feature map. • Each data point in the dataset recognizes itself by competing for a representation. • The Self-Organizing Maps’ mapping steps start from initializing the weight to vectors. Architecture
• After this, a random vector as the sample is selected
and the mapped vectors are searched to *nd which weight best represents the chosen sample. • Each weighted vector has neighboring weights present that are close to it. The chosen weight is then rewarded by being able to become a random sample vector. • This helps the map to grow and form di)erent shapes. Most generally, they form square or hexagonal shapes in a 2D feature space. • This whole process is repeatedly performed a large number of times and more than 1000 times. Architecture
• Self-Organizing Maps do not use
backpropagation with SGD to update weights, this unsupervised ANN uses competitive learning to update its weights i.e Competition, Cooperation and Adaptation. • Each neuron of the output layer is present with a vector with dimension n. • The distance between each neuron present at the output layer and the input data is computed. Architecture
• The neuron with the lowest distance is termed
as the most suitable *t. • Updating the vector of the suitable neuron in the *nal process is known as adaptation, along with its neighbour in cooperation. • After selecting the suitable neuron and its neighbours, we process the neuron to update. • The more the distance between the neuron and the input, the more the data grows. Architecture
• To simply explain, learning occurs in the
following ways: – Every node is examined to calculate which suitable weights are similar to the input vector. The suitable node is commonly known as the Best Matching Unit. – The neighbourhood value of the Best Matching Unit is then calculated. The number of neighbours tends to decrease over time. Architecture
• The suitable weight is further rewarded with
transitioning into more like the sample vector. The neighbours transition like the sample vector chosen. – The closer a node is to the Best Matching Unit, the more its weights get altered and the farther away the neighbour is from the node, the less it learns. • Repeat the second step for N iterations. Pros
• Data can be easily interpreted and understood
with the help of techniques like reduction of dimensionality and grid clustering. • Self-Organizing Maps are capable of handling several types of classi*cation problems while providing a useful, and intelligent summary from the data at the same time. Cons
• It does not create a generative model for the
data and therefore the model does not understand how data is being created. • Self-Organizing Maps do not perform well while working with categorical data and even worse for mixed types of data. • The model preparation time is comparatively very slow and hard to train against the slowly evolving data. Thanks