ANN Based Facial Emotion Detection and Music Selection
ANN Based Facial Emotion Detection and Music Selection
effort. When anticipating a person's emotions and mood, the of doing so is through facial expression. Feelings,
face is crucial. We create a prototype for a dynamic music emotions, and mood may all be communicated through the
recommendation system based on human emotions in this face.
system that is being suggested. Songs for each emotion are
taught based on human listening patterns. The emotion on a Fear, disgust, anger, surprise, sad, happy and neutral,
real person's face is recognised using an integration of and a state of neutrality are the basic human emotions that
feature extraction and machine learning techniques. Once can be broken down into more specific categories. These
the mood is determined from the input image, the feelings can also be used as an umbrella term to describe a
appropriate music will be played to keep the users' attention. wide range of different states of mind, including concept
It consists of two stages: training and real-time emotion and cheerfulness, amongst others. These feelings are
recognition and music selection. The suggested study has perfectly suited for the situation. Because of this, facial
demonstrated a notable level of performance in terms of muscular contortions are quite subtle, yet being able to
recognition and music choice. differentiate between them can result in a wide range of
expressions. Because of how strongly an emotion is
Keywords—, emotion recognition, facial recognition.
influenced by its surrounding context, different people —
or even the same people — may express the same feeling
I. INTRODUCTION in different ways..
The human face is an important organ of the body, and Machine learning and neural networks have shown
it plays an especially vital function in determining an promising outcomes when applied to such categorised
individual's behaviour and emotional condition. It is a tasks. Machine learning algorithms have already been put
highly laborious, time-consuming, and time-consuming to good use in the fields of pattern identification and
operation to manually segregate the list of songs and classification; this suggests that they may also be
generate a suitable playlist depending on an individual's applicable to the detection of emotional states. Because of
emotional qualities. However, this is something that must the rise of digital music, it is crucial to create a system that
be done. can propose songs based on individual tastes.
A number of different algorithms have been proposed
and developed in order to automate the process of creating II. LITERATURE SURVEY
playlists. However, the proposed existing algorithms that Interaction with a human being is necessary for the
are now being used are computationally slow, have a lower traditional method[5] of playing music that corresponds to
level of accuracy, and at times even necessitate the usage a person's emotional state. The transition to technologies
of supplementary gear such as EEG or sensors. This based on computer vision will make the automation of
proposed method, which is based on the extraction of such systems possible. In order to do this task, an
facial expressions, will automatically build a playlist, algorithm is utilised to categorise human expressions, and
hence decreasing the amount of time and effort that are then a music track is played in accordance with the current
required to render the process manually. emotion that has been identified. It saves time and effort
The integration of Facial Emotion Recognition (FER) that would otherwise be required to manually browse
and Music Information Retrieval (MIR) into conventional through a collection of songs in order to find one that
music players made it possible to automatically categorise corresponds to a person's current mood. The facial features
the playlist according to a wide range of feelings and states of a person are extracted with the use of an algorithm
of mind. MER is a technique that is used to identify a called Principle Component Aanalysis and a classifier
facial extraction that has been received by taking into called Euclidean Distance. This allows the expressions of
consideration the numerous facial features of the extracted the person to be recognised. When compared to alternative
face and how they correspond to different categories of ways, the use of an integrated camera to record a person's
feelings and states of mind. Even though both MER and facial expressions results in a reduction in the amount of
MIR featured the capabilities of avoiding the manual money spent on the system's design.
segregation of songs and development of playlists, it is still A wide variety of human experiences and emotions can
unable to fully incorporate a human emotion-controlled be captured and understood through the medium of song.
music player. This is the case despite the fact that both Emotion-based classification systems that can be relied
928
Authorized licensed use limited to: Amrita School of Engineering. Downloaded on September 04,2023 at 04:01:14 UTC from IEEE Xplore. Restrictions apply.
Fig 2: Music recommendation Module
929
Authorized licensed use limited to: Amrita School of Engineering. Downloaded on September 04,2023 at 04:01:14 UTC from IEEE Xplore. Restrictions apply.
faces of the pupils. Adaboost is the learning algorithm that CONCLUSION
forms the foundation of Haar Cascades. An important step forward in understanding human
behaviour has been developed as part of the proposed
Algorithm: effort. 7 common human emotions are explored here. An
Emotion-Based Music Player satisfies the need to classify
Training phase musical selections in accordance with emotional states.
Select the FER2013 database Both a training and a testing phase make up the suggested
system. In the training phase, the Adaboost learning
Preprocess the Dataset algorithm is utilised to train on the FER2013 database.
Read the images of N x N HarrCascades is used to analyse face expressions in real
time, allowing for the identification of seven distinct
Resize the images human emotions and the subsequent selection of an
Extract the facial features appropriate soundtrack.
930
Authorized licensed use limited to: Amrita School of Engineering. Downloaded on September 04,2023 at 04:01:14 UTC from IEEE Xplore. Restrictions apply.
[18] D.Fistum,D.Mali,andM.Ismail,―Bandwidthenhancementofrectangul
armicrostrippatchantennausing defected ground structure,‖
Indonesian Journal ofElectrical Engineering and Computer
Science, vol. 3, no.2, pp.428-434,2016.
[19] A. Zaidi et al., "High gain microstrip patch antenna, with PBG
substrateand PBG cover, for millimetre wave applications," 2018
4th InternationalConference on Optimization and Applications
(ICOA), IEEE, pp. 1-6,2018.
[20] S. S. Kumar, B. K. Aishwarya, K. N. Bhanutheja and M. Chaitra,
"Breath to speech communication with fall detection for
elder/patient with take care analytics," 2016 IEEE International
Conference on Recent Trends in Electronics, Information &
Communication Technology (RTEICT), Bangalore, 2016, pp. 527-
531, doi: 10.1109/RTEICT.2016.7807877.
931
Authorized licensed use limited to: Amrita School of Engineering. Downloaded on September 04,2023 at 04:01:14 UTC from IEEE Xplore. Restrictions apply.