Proceedings of CIC 2016 Track 1 Final PD
Proceedings of CIC 2016 Track 1 Final PD
Published & Edited by
International Journal of Computer Science and
Information Security (IJCSIS)
Vol. 14 Special Issue CIC 2016
ISSN 1947-5500
© IJCSIS PUBLICATION 2016
Pennsylvania, USA
Indexed and technically co‐sponsored by :
Editorial
Message from Editorial Board
It is our great pleasure to present the CIC 2016 Special Issue (Volume 14 Track 1, 2, 3, 4, 5, 6)
of the International Journal of Computer Science and Information Security (IJCSIS). High
quality research, survey & review articles are proposed from experts in the field, promoting insight
and understanding of the state of the art, and trends in computer science and technology. It
especially provides a platform for high-caliber academics, practitioners and PhD/Doctoral
graduates to publish completed work and latest research outcomes. According to Google Scholar,
up to now papers published in IJCSIS have been cited over 6818 times and the number is quickly
increasing. This statistics shows that IJCSIS has established the first step to be an international
and prestigious journal in the field of Computer Science and Information Security. There have
been many improvements to the processing & indexing of papers; we have also witnessed a
significant growth in interest through a higher number of submissions as well as through the
breadth and quality of those submissions. IJCSIS is indexed in major academic/scientific
databases and important repositories, such as: Google Scholar, Thomson Reuters, ArXiv,
CiteSeerX, Cornell’s University Library, Ei Compendex, ISI Scopus, DBLP, DOAJ, ProQuest,
ResearchGate, Academia.edu and EBSCO among others.
On behalf of IJCSIS community and the sponsors, we congratulate the authors, the reviewers
and thank the committees of International Conference On Advances In Computational
Intelligence In Communication (CIC 2016) for their outstanding efforts to review and
recommend high quality papers for publication. In particular, we would like to thank the
international academia and researchers for continued support by citing papers published in
IJCSIS. Without their sustained and unselfish commitments, IJCSIS would not have achieved its
current premier status.
“We support researchers to succeed by providing high visibility & impact value, prestige and
excellence in research publication.” For further questions or other suggestions please do not
hesitate to contact us at ijcsiseditor@gmail.com.
Open Access This Journal is distributed under the terms of the Creative Commons Attribution 4.0 International License
(http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium,
provided you give appropriate credit to the original author(s) and the source.
Bibliographic Information
ISSN: 1947-5500
Monthly publication (Regular Special Issues)
Commenced Publication since May 2009
CIC 2016 Committees
Dr. T.G. Palanivelu Former Principal, PEC & Professor, SMVEC, Puducherry
CIC2016_paper 60: Recognition of Gait in Arbitrary Views using Model Free Methods
M. Shafiya Banu, Department of Information science and Technology, Anna University, Chennai.
M. Sivarathinabala, Department of Information science and Technology, Anna University, Chennai.
S. Abirami, Department of Information science and Technology, Anna University, Chennai.
CIC2016_paper 67: Study on Watermarking Effect on Different Sub Bands in Joint DWT-DCT
based Watermarking Scheme
Mohiul Islam, Department of Electronics & Communication Engineering National Institute of
Technology Silchar, Assam, India
Amarjit Roy, Department of Electronics & Communication Engineering National Institute of Technology
Silchar, Assam, India
Rabul Hussain Laskar, Department of Electronics & Communication Engineering National Institute of
Technology Silchar, Assam, India
CIC2016_paper 70: Image Denoising using Hybrid of Bilateral Filter and Histogram Based Multi-
Thresholding With Optimization Technique for WSN
H. Rekha, Research Scholar, Department of Electronics engineering, Pondicherry University,
Pondicherry, India
P. Samundiswary, Assistant Professor, Department of Electronics Engineering, Pondicherry University
Pondicherry, India
CIC2016_paper 86: Chaos Based Study on Association of Color with Music in the Perspective of
Cross-Modal Bias of the Brain
Chandrima Roy, Department of Electronics &Communication Engineering, Heritage Institute of
Technology, Kolkata, India
Souparno Roy (1), Dipak Ghosh (2)
(1) Researcher, (2) Professor Emeritus, Sir C.V. Raman Centre for Physics & Music, Kolkata, India
CIC2016_paper 99: Estimation of Visual Focus of Attention from Head Orientations in a Single
Top-View Image
Viswanath K. Reddy, Assistant Professor, Department of Electronic and Communication Engineering,
M. S. Ramaiah University of Applied Sciences, Bangalore, India
CIC2016_paper 107: Face Recognition Under Varying Blur, Illumination and Expression in an
Unconstrained Environment
Anubha Pearline.S, M.Tech, Information Technology, Madras Institute of Technology, Chennai, India
Hemalatha.M, Assistant Professor, Information Technology, Madras Institute of Technology, Chennai,
India
CIC2016_paper 117: Efficient Stereoscopic 3D Video Transmission over Multiple Network Paths
Vishwa kiran S, Thriveni J, Venugopal K R, Dept. Computer Science and Engineering
University Visvesvaraya College of Engineering, Bangalore, India
Raghuram S, Pushkala Technologies Pvt. Ltd., Bangalore, India
CIC2016_paper 131: Formant Frequency Based Analysis of English vowels for various Indian
Speakers at different conditions using LPC & default AR modeling
Anil Kumar C., Research Scholar, Electronics Engg, Jain University, Bengaluru., India.
M. B. Manjunatha, Principal, A.I.T., Tumkur.,India
G. N. Kodanda Ramaiah, Professor, HOD and Dean R & D, Dept of ECE. K.E.C., Kuppam., India
TRACK 2 Computational Intelligence in Wireless Communication Networks
CIC2016_paper 17: Design of Cooperative Spectrum Sensing based spectrum access in CR
networks using game theory
Lavanya Shanmugavel, Fenila Janet, M. A. Bhagyaveni
Dept. of ECE, CEG, Anna University, Chennai, INDIA
CIC2016_paper 22: An Intelligent Cognitive Radio Receiver for Future Trend Wireless
Applications
M. Venkata Subbarao, Research Scholar, Department of EE, School of Engineering & Technology,
Pondicherry University, Pondicherry, India.
P. Samundiswary, Assistant Professor, Department of EE, School of Engineering & Technology,
Pondicherry University, Pondicherry, India.
CIC2016_paper 31: Vehicular Ad Hoc Networks: New Challenges in Carpooling and Parking
Services
Amit Kumar Tyagi, Research Scholar, Department of CS&E, Pondicherry Engineering College,
Puducherry-605014, India.
Sreenath Niladhuri, Professor, Department of CS&E, Pondicherry Engineering College, Puducherry-
605014, India.
CIC2016_paper 36: Efficient Energy Utilisation in Zigbee WDSN using Clustering protocol and
RSSI Algorithm
Maria Brigitta.R, Department of Electronics Engineering, School of Engineering and Technology
Pondicherry University, Puducherry-605 014, India
Samundiswary.P, Department of Electronics Engineering, School of Engineering and Technology
Pondicherry University, Puducherry-605 014, India
CIC2016_paper 37: Prediction of Spacecraft Position by Particle Filter based GPS/INS integrated
system
Vijayanandh R [1], Raj Kumar G [2]
[1], [2] – Assistant Professor, Department of Aeronautical Engineering, Kumaraguru College of
Technology, Coimbatore, Tamil Nadu, India
Senthil Kumar M [3], Samyuktha S [4]
[3] – Assistant Professor (SRG), [4] – BE Student, Department of Aeronautical Engineering,
Kumaraguru College of Technology, Coimbatore, Tamil Nadu, India
CIC2016_paper 68: Dynamic Application Centric Resource Provisioning Algorithm for Wireless
Broadband Interworking Network
S. Kokila, Department of Electronics and Communication Engineering, Pondicherry Engineering
College Puducherry, India
G. Sivaradje, Department of Electronics and Communication Engineering, Pondicherry Engineering
College, Puducherry, India
CIC2016_paper 69: RSSI based Tree Climbing mechanism for dynamic path planning in WSN
Thilagavathi P, Research scholar, Department of Information Technology, Jerusalem college of
Engineering, Chennai 600100, India
Martin Leo Manickam J, Professor, Electronics and Communication Engineering, St. Joseph’s college
of Engineering, Chennai 600119, India
CIC2016_paper 73: Conductor Backed CPW Fed Slot Antenna for LTE application
M. Saranya, S. Robinson,Gulfa Rani
Department of ECE, Mount Zion College of Engineering and Technology, Pudukkottai, India
CIC2016_paper 74: Power Optimized and Low Noise Tunable BPF using CMOS Active Inductors
for RF Applications
A. Narayana Kiran, Assistant Professor, Department of ECE, Shri Vishnu Engineering College for
Women, Bhimavaram, India,
P. Akhendra Kumar, Research Scholar, Department of ECE, National Institute of Technology,
Warangal, Warangal, India
CIC2016_paper 76: Butterfly Shaped Microstrip Patch Antenna with Probe Feed for Space
Applications
Deepanshu Kaushal, PG student, Department of Electronics Engineering, Pondicherry University
Pondicherry, India
T. Shanmuganatham, Assistant Professor, Department of Electronics Engineering, Pondicherry
University, Pondicherry, India
CIC2016_paper 78: Primary User Emulation Attack Analysis in Filter Bank Based Spectrum
Sensing Cognitive Radio Networks
Sabiq P.V. & D. Saraswady
Dept. of ECE, Pondicherry Engineering College, Puducherry, India
CIC2016_paper 85: Ant Colony Multicast Routing for Delay Tolerant Networks
E. Haripriya, Assistant Professor of Computer Science, J.K.K.Nataraja College of Arts & Science,
Namakkal, TamilNadu, India
K. R. Valluvan, Professor and Head of ECE, Velalar College of Engineering & Technology, Erode,
TamilNadu, India
CIC2016_paper 87: An Energy-Efficient Key Management Scheme using Trust Model for
Wireless Sensor Network
P. Raja, Associate Professor, Department of ECE, Sri Manakula Vinayagar Engineering College,
Pondicherry, India
E. Karthikeyan, Department of ECE, Sri Manakula Vinayagar Engineering College, Pondicherry, India
CIC2016_paper 88: Power Efficiency analysis of Four State Markov Model based DRX
mechanism with OTSC ratio for Long Term Evolution User Equipment
R. Vassoudevan, Research Scholar, Department of Electronics Engineering, Pondicherry University,
Puducherry, India
P. Samundiswary, Assistant Professor, Department of Electronics Engineering, Pondicherry University
Puducherry, India
TRACK 3 Computational Intelligence in Wireless Communication Networks
CIC2016_paper 32: Ensuring Trust and Privacy in Large Carpooling Problems
Amit Kumar Tyagi, Research Scholar, Department of CS&E, Pondicherry Engineering College,
Puducherry-605014, India.
Sreenath Niladhuri, Professor, Department of CS&E, Pondicherry Engineering College, Puducherry-
605014, India.
CIC2016_paper 92: New Joint Non Linear Companding and Selective Mapping Method for PAPR
Reduction in OFDM System
Sandeep Dwivedi, M.Tech. student, Department of Electronics Engineering, School of Engineering and
Technology, Pondicherry University, Puducherry-605014
P. Samundiswary, Assistant Professor, Department of Electronics Engineering, School of Engineering
and Technology, Pondicherry University, Puducherry-605014
CIC2016_paper 96: Collaborative Location Based Sleep Scheduling With Load Balancing In
Sensor-Cloud
N. Mahendran, Assistant Professor, Dept of ECE, M. Kumarasamy College of Engineering, Karur,
Tamilnadu
CIC2016_paper 105: A layer based survey on the security issues of cognitive radio networks
Tephillah.S, J.Martin Leo Manickam
ECE, St.Joseph’s College of Engineering, Chennai, India
CIC2016_paper 106: TCOR- Energy Efficient and Power Saving Routing Architecture for Mobile
AD HOC Networks
S. Sargunavathi Associate professor, ECE, Sriram Engineering College Chennai, India
Dr. J.Martin Leo Manickam, Professor, ECE, St Joseph’s Engineering College Chennai-119, India
CIC2016_paper 113: Cluster Head Selection in Cognitive Radio Networks using Fruit Fly
Algorithm
Umadevi K.S., School of Computing Science and Engineering, VIT University, Vellore, India.
CIC2016_paper 116: Reputation Based IDS for Gray hole Attack in MANETs
K. Ganesh Reddy, K. Radharani, K. V. Sravani, K. Mounika, K. Poojitha,
Dept. of Computer Science and Engineering, Shri Vishnu Engineering College for Women,
Bhimavaram, India
P. Santhi Thilagam, Dept. of Computer Science and Engineering, NITK Surathkal, Mangalore, India
CIC2016_paper 120: Customer friendly Fast and Dynamic Handover in Heterogeneous Network
Environment
T. Senthil Kumar, M. A. Bhagyaveni
Department of Electronics and Communication Engineering, College of Engineering, Guindy, Anna
University, Chennai, India
CIC2016_paper 138: Type-2 Fuzzy based GPS Satellite Selection algorithm for better
Geometrical Dilution of Precision
Arul Elango G, ECE Department, Pondicherry Engineering College, Puducherry, India
Murukesh C and Rajeswari K, EIE Department, Velammal Engineering College, Chennai, India
CIC2016_paper 146: Dual Microphone Speech Enhancement Utilizing General Kalman Filter in
Mobile Communication
Vijay Kiran Battula, Department of ECE, University College of Engineering Vizianagaram, JNTUK,
Vizianagaram, INDIA
Appala Naidu Gottapu, Department of ECE, University College of Engineering Vizianagaram, JNTUK,
Vizianagaram, INDIA
CIC2016_paper 147: Artificial Bee Colony Algorithm Based Trustworthy Energy Efficient
Routing Protocol
D. Sathian, Department of Computer Science, Pondicherry University
M. Gunashanthi, Department of Computer Science, Pondicherry University
P. Dhavachelvan, Department of Computer Science, Pondicherry University
TRACK 4 Computational Methods in Biosignal Processing for Telemedicine
CIC2016_paper 04: Modified Local Gradient Pattern Based Computation Analysis for the
Classification of Mammogram
Narain Ponraj (1), Poongodi (2), Merlin Mercy (3)
Dept of ECE (1,2) Dept of CSE (3) Karunya University (1) KCE (2) SKCT (3) India
CIC2016_paper 09: Study on the use of Multi frequency Bioelectrical Impedance for
Classification of Risk of Dengue fever in Indian Children
Neelamegam Devarasu (1) and Gnanou Florence Sudha (2)
(1,2) Department of Electronics and Communication Engineering,
Pondicherry Engineering College, Puducherry, India.
CIC2016_paper 21: An Effective Liver Cancer Diagnosis through Multi – Temporal Fusion and
Decorrelation Stretching Techniques
B. Lakshmi Priya, S. Joshi Adaikalamarie, K. Jayanthi
Department of ECE, Pondicherry Engineering College, Puducherry
CIC2016_paper 56: Non-Invasive Measurement of Cholesterol Levels Using Eye Image Analysis
S.V. Mahesh Kumar (a,*), R. Gunasundari (a) and N. Ezhilvathani (b)
(a) Department of Electronics and Communication Engineering, Pondicherry Engineering College,
Puducherry, India.
(b) Department of Ophthalmology, Indira Gandhi Medical College and Research Institute, Puducherry,
India.
CIC2016_paper 93: An Efficient Noise Cancellation Approach suitable for Respiratory Sound
Signals
Prashanth B.S., Department of Electronics & Communication Engineering, Pondicherry Engineering
College, Pillaichavadi, Puducherry,
Jayanthi K., Department of Electronics & Communication Engineering, Pondicherry Engineering
College, Pillaichavadi, Puducherry,
CIC2016_paper 109: Evaluation of Hypotension using Wavelet and Time Frequency Analysis of
Photoplethysmography (PPG) Signal
Remya Raj, Research scholar, Department of ECE, SRM University, Kattankulathur, Chennai, Tamil
Nadu-603203, India
Dr. J. Selvakumar, Asst. Professor(SG), Department of ECE, SRM University, Kattankulathur, Chennai,
Tamil Nadu-603203, India
Dr. M. Anburajan, Dept of Biomedical Engineering, SRM University, Kancheepuram, Tamil Nadu, India
CIC2016_paper 33: 3D Ultrasound Imaging For Automated Kidney Stone Detector On FPGA
K. Viswanath, Research Scholar, Member IEEE, Department of ECE, Pondicherry Engineering College
Pondicherry, India
Dr. R. Gunsundari, Professor, Department of ECE, Pondicherry Engineering College, Pondicherry,
India
TRACK 5 Computational Intelligence Methodologies
CIC2016_paper 44: Context Aware Web Service Discovery Optimization By Chameleon Inspired
Algorithm
(1) A. Amirthasaravanan, (2) Paul Rodrigues, (3) R. Sudhesh
(1) Department of Information Technology, University College of Engineering, Villupuram, Tamilnadu,
India
(2) DMI Engineering College, Chennai, Tamilnadu, India
(3) Department of Mathematics, BIT Campus, Tiruchirappalli, Tamilnadu, India
CIC2016_paper 95: Various Computing models in Hadoop Eco System along with the
Perspective of Analytics using R and Machine learning
Uma Pavan Kumar Kethavarapu, Research Scholar, Department of Computer Science and
Engineering, Pondicherry Engineering College
Dr. Lakshma Reddy Bhavanam, Principal, BCC College, Bangalore
CIC2016_paper 115: Using Semantic Fields For Generating Research Paper Summaries
A. L. Agasta Adline, Department of Information Technology, Easwari Engineering College, Chennai
600089, TamilNadu, India
Harish M, Department of Computer Science and Engineering, College of Engineering Guindy, Anna
University, Chennai 600025, TamilNadu, India
G.S. Mahalakshmi, Department of Computer Science and Engineering, College of Engineering Guindy,
Anna University, Chennai 600025, TamilNadu, India
S. Sendhilkumar, Department of Computer Science and Engineering, College of Engineering Guindy,
Anna University, Chennai 600025, TamilNadu, India
CIC2016_paper 142: Two Run Morphological Analysis for POS Tagging of Untagged Words
Betina Antony J, Dept of Computer Science and Engineering, College of Engineering Guindy, Anna
University, Chennai 600025, Tamil Nadu, India.
G. S. Mahalakshmi, Dept of Computer Science and Engineering, College of Engineering Guindy, Anna
University, Chennai 600025, Tamil Nadu, India
CIC2016_paper 143: Bio-Inspired Schedulers for Public Cloud Environment
Vaithianathan Geetha, Department of Information Technology, Pondicherry Engineering College,
Puducherry-605014.
CIC2016_paper 148: A Scrutiny and Appraisal of Various Optimization Algorithm to Solve Multi-
Objective Nurse Scheduling Problem
M. Rajeswari, Research Scholar, Department of Computer Science, Pondicherry University
S. Jaiganesh, Research Scholar, Department of CS, R&D Centre, Bharathiar University, Coimbatore.
P. Sujatha, Assistant Professor, Department of Computer Science, Pondicherry University
T. Vengattaraman, Assistant Professor, Department of Computer Science, Pondicherry University
P. Dhavachelvan, Professor, Department of Computer Science, Pondicherry University,
TRACK 6 Computational Intelligence Applications
CIC2016_paper 2: Information Detection System using 4T Dual Port CAM
V. Bharathi, Associate Professor, ECE Department, Sri Manakula Vinayagar Engineering College,
Puducherry, India
A. Ragasaratha Preethee, ECE Department, Sri Manakula Vinayagar Engineering College,
Puducherry, India
CIC2016_paper 16: VLSI Implementation of Reverse Converter via Parallel Prefix Adder for
Signed Integers
(1) P. Rajagopalan , Puducherry – 605 014, India
(2) A. V. Ananthalakshmi, Assistant Professor, Pondicherry Engineering College, Puducherry – 605
014, India.
CIC2016_paper 63: Optimal Tuning of Coordinated Controller using BBO Algorithm for Stability
Enhancement in Power System
Gowrishankar Kasilingam (1*), Jagadeesh Pasipuleti (2)
(1*) Research Scholar, Department of Electrical Power, Universiti Tenaga Nasional (UNITEN),
Malaysia
(1*) Associate Professor, Department of ECE, Rajiv Gandhi College of Engg. & Tech.,Pondicherry,
INDIA
(2) Associate Professor, Department of Electrical Power, Universiti Tenaga Nasional (UNITEN),
Malaysia
CIC2016_paper 80: Smart Phone Based Speed Breaker Early Warning System
Viswanath K. Reddy (1), and Nagesh B. S (2)
(1) Assistant Professor in the Department of Electronic and Communication Engineering in M.S.
Ramaiah University of Applied Sciences, Bangalore,
(2) Robert Bosch Engineering and Services India Pvt.Ltd, Bangalore
CIC2016_paper 82: B+ Indexing for Biometric Authentication using Fused Multimodal Biometric
Jagadiswary. D, Electronics and Communication Engineering, Pondicherry Engineering College,
Puducherry, India
Dr. D. Saraswady, Electronics and Communication Engineering, Pondicherry Engineering College,
Puducherry, India
CIC2016_paper 118: Smart Logistics for Pharmaceutical Industry based on Internet of Things
(IoT)
M. Pachayappan (1), Nelavala Rajesh (2), G. Saravanan (3)
(1) Assistant Professor, Department of International Business, School of Management, Pondicherry
University , Puducherry – 605 014, India
(2) Assistant Professor, Department of Electronic and Communication, Arunai College of Engineering,
Tiruvannamalai – 606601, India
(3) Assistant Professor, Department of Electronic and Communication, Valliammai Engineering
College, Kattankulathur - 603 203, India
CIC2016_paper 126: Smart Phone Keylogger Detection Technique Using Support Vector
Machine
S. Geetha, Research Scholar, Dept. of Banking Technology, Pondicherry University
G. Shanmugasundaram, Assistant Professor, Department of IT, SMVEC
BharathKumar V., UG Students, Department of IT, SMVEC
V. Prasanna Venkatesan, Associate Professor, Dept. of Banking Technology, Pondicherry University
CIC2016_paper 145: Standby Mode Subthreshold Leakage Power Analysis in Digital Circuits
with Variations in Temperature
Amuthavalli. G, Department of ECE, Pondicherry Engineering College, Puducherry, India
Gunasundari. R, Department of ECE, Pondicherry Engineering College, Puducherry, India
TRACK 1
Computational Intelligence in
Signal & Image Processing
Vol. 14 CIC 2016 Special Issue International Journal of Computer Science and Information Security (IJCSIS)
https://sites.google.com/site/ijcsis/
ISSN 1947-5500
Madhu Chandra G
Research Scholar Sreerama Reddy G.M
Dept.of ECE, Professor & HOD
MS Engineering College, Bangalore, Dept. of ECE,
VTU, Belagavi, India CBIT, Kolar, India
madhu.guru1984@gmail.com
Abstract— The area of video analytics has recently gained a pace II. RELATED WORK
with the advancement of distributed mining methodologies. For a
given scene, there are various forms of event, which are quite The significant study for video analytics was found in the
unpredictable in nature especially for unscripted data analysis. work of Aradhya and Pavithra [1] have introduced a study that
Such unpredictable events with respect to context are termed as performs recognizes text from the multiple language video. The
outliers. After reviewing the recent research contribution in this procedure utilizes k-mean clustering, wavelet transform, and
regards, we find that the existing video analytics are quite in Gabor filter by keeping in mind the end goal to perform video
infancy stage to understand such outliers and apply mining indexing. Despite the fact that, the study is not
operation. The presented manuscript introduces a technique
straightforwardly a use of video analytics but rather its system
which performs precise detection of outliers with better com-
putational performance. The technique extracts features using
is exceptionally assistive in video analysis. The latest study
dictionary-based approach and extraction contextual data of an towards video analysis was completed by Angelov et al. [2]
object using unsupervised learning technique. The outcome of the and planned to do continuous analysis of video. The system
study was compared with existing system to excel better utilizes iterative density estimation and local clustering average
performance with respect to accuracy factor and reduced estimation. The system used the Lucas Kanade mechanism and
computational capability. in addition RANSAC mechanism to perform better objects
detection from air borne vehicle. In the work of Ayed et al. [3],
Keywords- Video Analytics; Video Mining; Data Mining; utilized MapReduce model as a part of request to perform
Outlier Detection; Multimedia Analytics
mining of video data set. Aryanfar et al. [4] have utilized a
I. INTRODUCTION customary classifier (bolster vector machine and innocent
Bayes) with a specific end goal to perceive the human activity.
A typical video consists of multiple information which are
The result of the study was contrasted with existing component
the source of interest for any data mining operation. There are
with find around 90% of identification rate. The study result
three types of information existing within a video e.g. i) low-
level information (e.g. texture, color, shape, etc.), ii) semantic was affirmed with visual results, exemplary model time, and
infor-mation (e.g. spatial factors-characters, location, objects handling time taken by MapReduce. Cai et al. [5] have talked
etc. and temporal factor-video sequences and movement of about the video analysis keeping in mind the end goal to Gauge
objects), and iii) Syntactic information (salient objects, the engagement of clients. The statistical approach was used to
respective position and timing attributes etc.). Another perform video prediction. Xu et al. [6] have introduced a
significant factors that plays a crucial role in designing video strategy that empowers, mining operation on 3D data of video
analytics are modalities. There are three types of video over cloud environment. The method additionally crypto-
modalities viz. i) auditory (music, speech, surrounding sounds), realistic performed over the video data. The result of the study
ii) visual (seen of video), and iii) textual modalities (text was assessed regarding the time required for encrypted and
contents of video). Presences of outliers are the most original video coding. Chen et al. [7] have exhibited a model
challenging to be identified owing to lack of predefined called as PeakVizor which means to extricate the examples of
knowledge about the scene. This paper presents technique of advanced online courses mechanism. The study utilizes Glyph
developing a new video analytical model that can perform cost representation strategy as a part of request to perform
effective mining operation for solving the problem of outlier information reflection and also crest location from MOOC
detection of input video. Section II discusses about the related (Massive Open Online Courses) recordings. The concentrate
work followed by brief discussion of problem identification in additionally helps with defining relationship among various
Section III. Proposed system is discussed in Section IV online users. Study on video observation was likewise
followed by research methodology in Section V. Finally result
completed by Mao et al. [8]. The researchers have exhibited a
discussion is made in Section 6 followed by conclusion in
novel traceability component for framework over video
Section VI.
reconnaissance framework. The method likewise presents a
• A framework will likewise utilize pack of visual words in outlier detection. However, Cong et al. [12] uses a method
model for video pattern extraction. called as dynamic patch grouping where the author uses k-
• An empirical model is needed to develop to minimize the means clustering resulting in non-consideration of local
errors during vector spar city based image reconstruction. patterns. This causes declination of performance metrics to a
• An identification model is needed to be developed to significant level. Moreover Cong et al. [12] have not
enhance the video abnormality or outliers characteristics. incorporated optimization technique like us. Proposed system
• Different scenarios are needed to be considered for uses a simple optimization technique discussed by Li and
performance analysis of the method with both real time Ngom [13]. The processing time for identification of outlier is
approximately 0.0265 seconds in 4 GB RAM with Core i3
and other video data sets.
processor and the proposed system doesn’t store any values
• The proposed method effectiveness is needed to identify
over run time leading to extremely less time complexity.
as true or false. Hence, the outcome of the proposed study is good in outlier
detection while the approach used can be said to be a cost
VI. RESEULT DISCUSSION effective one with respect to computational complexity factor.
The proposed technique was evaluated with UCSD
pedestrian dataset. Our analysis was done in two parts i.e.
VII. CONCLUSION
developing a trained video and given a query of new test video
to identify the outlier objects in mobile condition. We make the This model gives genuine abnormality identification for the
discussion of result analysis with respect to both visual and desired video. The proposed study with the probability
numerical outcomes. mechanism, the objectives achieved are correct and accurate.
Also the proposed framework has achieved the accuracy of
A. Visual Outcomes Analysis 95% in video abnormality in different scenarios.
Fig.3 shows the 3 visual outcomes. The first visual outcome
is for all objects with similar context i.e. pedestrian. The REFERENCES
second and third visual outcomes are for highlighting the [1] V.N.M. Aradhya and M. S. Pavithra, “A comprehensive of transforms,
outlier object which is basically a non-pedestrian. We have Gabor filter and k-means clustering for text detection in images and
testified with other test images to observe similar behavior of video", Applied Computing and Informatics, 2014
the proposed mechanism.
[2] P. Angelov, P. S-Tehran, C. Clarke, “AURORA: autonomous real-time
on-board video analytics”, Neural Comput & Applica, 2016
[3] A.B. Ayed, M. B. Halima, and A. M. Alimi, “MapReduce Based Text
Detection in Big Data Natural Scene Videos”, Procedia Computer
Science, Vol. 53, pp.216-223, 2015.
[4] A. Aryanfar, R. Yaakob, A. A. Halin, N. Sulaiman, K. A. Kasmiran, and
L. Mohammadpour, “Multi-View Human Action Recognition Using
Wavelet Data Reduction and Multi-Class Classification”, Procedia
Computer Science, Vol. 62, pp.585-592, 2015.
[5] P. Vollucci, B. Le, J. Lai, W. Cai, Y. Ye, G. Necula, and D. Wroblewski,
Figure. 3 Visual Outcomes of Study
“Online Video Data Analytics”, 2015
B. Numerical Outcome Analysis [6] T. Xu, W. Xiang, Q. Guo, and L. Mo, “Mining cloud 3D video data for
For the purpose of better benchmarking, we compare the interactive video services”, Mobile Networks and Applications, Vol. 20,
numerical outcomes of proposed system with Cong et al. [12] No. 3, pp. 320-327, 2015.
work with respect to standard performance metrics e.g. recall
[7] C.H.E. N. Qing, Y. Chen, D. Liu, C. Shi, Y. Wu, and H. Qu,
rate, precision rate, specificity, and F1-Score (Table 1).
“PeakVizor: Visual Analytics of Peaks in Video Clickstreams from
Table 1 Comparative Numerical Outcome Massive Open Online Courses”, 2015.
Techniq Reca Precis Specifi F1- [8] B. Mao, J.He, J. Cao, S.W. Bigger, and T. Vasiljevic, “A framework for
ues ll ion city Score food traceability information extraction based on a video surveillance
system”, Procedia Computer Science, Vol. 55, pp.1285-1292, 2015.
Propose 0.76 0.998 0.9992 0.76
[9] Y.B. Kim, S. J. Kang, Sang Hyeok Lee, Jang Young Jung, Hyeong
d 7 62 5 282
Ryeol Kam, Jung Lee, Young Sun Kim, Joonsoo Lee, and Chang Hun
Cong 0.34 0.917 0.8672 0.51 Kim. "Efficiently detecting outlying behavior in video-game
[12] 816 65 2 62 players." PeerJ 3 (2015): e1502.
The similarity between proposed and Cong et al. [12] is [10] M. Shao, and Y.Fu, “Deeply Self-Taught Multi-View Video Analytics
usage of similar dataset, similar goal, and adoption of context Machine for Situation Awareness”, AFA Cyber Workshop, White Paper,
2015
Abstract--Steganography is the technique used for information the pixels of the images are
concealing messages inside the appropriate cover files misrepresented so as unviewable to the others and the
such as images, video, audio files. The steganography changes applied in the image are indefinable. The
main objective is to share securely in such a way that image for embedding the information is called
the required information is not visible to the viewer.
The information which is to be transmitted over the
envelop image and it becomes stego image after
envelop image is converted to 0’s and 1’s form by trouncing secret information[2].
Huffman coding. This converted code is embedded The compression is done for covert
inside the envelop image by varying the Least information for enhancing the hiding capability and
Significant Bit (LSB) of each of the pixel values of protection. The general standard of information
information can be decrypted by Huffman Table which compression algorithms on content files is to
is embedded in envelop image. The information can be transform a sequence of characters into a new
decrypted by Huffman Table which is embedded in sequence which contains the same information but
envelop image itself so that the stego image becomes with new length as small as possible. The proficient
impartial information to the viewer. The algorithm has
elevated capacity and superior invisibility by Huffman
information compression algorithm is elected
Encoding. Furthermore, Peak Signal to Noise Ratio according to some scales like: compression size,
(PSNR) of stego image with envelop image shows better compression ratio, processing time or speed, and
result, and agreeable protection is maintained since the entropy[4].
covert information cannot be extracted without Steganography techniques can be used in
knowing the decoding rules and Huffman Table. spatial domain or in frequency domain. Spatial
domain techniques are easier than frequency domain
Keywords--Steganography;Huffman Encoding; LSB techniques and many steganographic methods are
proposed in spatial domain. The basic steganography
I. INTRODUCTION technique in spatial domain is LSB replacement,
where the least significant bit in each pixel of image
Steganography is mainly used to share is replaced with the covert information bit[6].
information securely such that the true message is not The projected approach uses Huffman
perceptible to the viewer. Hiding the covert coding for information compression and then the
information in an image is called image compressed information is finally hidden within an
steganography. image using basic steganography technique i.e., LSB
Replacement.
Lenna
Huffman Coding Arithmetic Coding
Size Encoded MSE PSNR Size Encoded MSE PSNR
Bit Size Bit Size
4.02KB 16910 0.0093 68.4892 4.02KB 19760 0.0112 67.6366
10KB 25835 0.0139 66.7415 10KB 49758 0.0245 63.3225
50KB 210087 0.0812 59.0713 50KB 245654 0.0844 58.9003
Cameraman
Huffman Coding Arithmetic Coding
Size Encoded MSE PSNR Size Encoded MSE PSNR
Boat:
Bit Size Bit Size
4.02KB 16910 0.0115 67.5424 4.02KB 19760 0.0140 66.6794
10KB 25835 0.0172 68.4442 10KB 49758 0.0377 62.4064
50KB 210087 0.0791 59.1837 50KB 245654 0.0816 59.0474
Boat
Huffman Coding Arithmetic Coding
Size Encoded MSE PSNR Size Encoded MSE PSNR
Bit Size Bit Size
4.02KB 16910 0.0093 68.4442 4.02KB 19760 0.0114 67.5858
10KB 25835 0.0141 66.6768 10KB 49758 0.0309 63.2591
50KB 210087 0.0813 59.0642 50KB 245654 0.0843 58.9081
IV. CONCLUSION
A. Plot Representation The proposed image steganography based on
Huffman Encoding and Arithmetic Coding. The
The graph is plotted for Huffman Coding and
Huffman Algorithm improves the security and
Arithmetic Coding on PSNR and Text Size for
quality of stego image and is better in comparison
different images. We consider text size on X-axis and
with other existing algorithms. According to results,
PSNR on Y-axis. The resultant is as follows:
the stego images are identical to that of cover images
-----: Huffman Coding
and is difficult to differentiate between them. We
-----: Arithmetic Coding
have achieved 100% recovery of the secret image
Lenna:
that means original and extracted secret images are
identical. The stego image also contains the size of
the text which is embedded using compression
algorithms,so that the stego image itself is the
standalone information to the decoder.The decoder
only needs to know the extraction algorithm to
extract the secret message.
REFERENCES
[3] A.Nag,S.Biswas,D.Sarkar, P. P. Sarkar, "A Novel Processing as ECE faculty. Interesting Areas are
Technique for Image Steganography Based on Digital Signal Precessing, Array Signal Processing,
Huffman Encoding". International Journal of Image Processing, Video surveillance, Embedded
Computer Science and Information Technology,
Systems, Digital Communications.
Volume 2, Number 3, June 2010.
[4] Abbas Cheddad, Joan Condell, Kevin Curran,
Paul McKevitt. "Digital Image Steganography:
SurSurvey and Analysis of Current Methods".
ELSEVIER Journal on Signal Processing 90 (2010)
7727-752.
[5] Gonzalez, R.C. and Woods, R.E., Digital Image
Processing using MA TLAB, Pearson
C.Maheswari, Assistant Professor, Dept of ECE, Sree
Education,Tndia,2006.
Vidyanikethan Engineering College, A. Rangampet,
[6] Alkhrais Habes, "4 least Significant Bits
Tirupati received M.Tech in DECS From JNTUA
Information Hiding Implementation and Analysis",
COLLEGE OF ENGINEERING PULIVENDULA,
ICGST Int. Conf. on Graphics, Vision and Image
B.Tech in Electronics and Communication
Processing (GVIP-05), Cairo, Egypt, 2005.
Engineering from YSR Engineering college of yogi
[7] Neil F. Jhonson, Sushil Jajodia. "Exploring
vemana university, proddatoor kadapa district.
Steganography: Seeing the Unseen". IEEE paper of
Interesting Areas Image Processing, Digital Signal
February 1998.
Processing, Image Processing, ERTS.
AUTHORS PROFILE:
Abstract- A new image encryption algorithm based on pixel categories such as value transformation, pixel position
integration is proposed in this paper. This paper presents permutation, and chaotic systems [6]. There are two groups of
implementing security for multiple image that enables to select cryptography [7] image encryption algorithms (a)Chaos-based
one of the several images displayed simultaneously with a unique selective [8] and (b) Non-chaos selective methods or non-
security key for each image. This process involves encrypting the selective methods. It can be admitting, no particular encryption
images with Elliptic Curve Cryptography followed by applying algorithm which satisfies all image type requirements. An
block based interleaving and integrating the image matrix using Encryption Algorithm should be strong against all types of
the pixel based integration technique. Decrypting the image is attacks, including statistical and brute force attacks. Different
done with the key specifically generated for the image from
security algorithms [5] have been used to provide the required
multiple images. The proposed method provides a high level
security of interactive information requirements in the fields of
protection and many encryption algorithms have been proposed
military, confidential, aerospace, financial and economic, to enhance the image security.
national security and so on. The algorithm is evaluated by In this paper we presented, Elliptic Curve Cryptography
calculating the entropy and correlation values. [9], [10] algorithm is applied on the number of input images to
generate encrypted image and interleaving with pixel based
Keywords—Image Encryption; Elliptic Curve Cryptography; integration to enhance security, which is considered as
Pixel Grouping; Entropy; Correlation; Pixel Integration;Security; complex process, in particular, the information is two-
Digital Communication dimensional, due its size and redundant in nature. In our
I. INTRODUCTION proposed method, set of nine image inputs are resized to 64*64
and divided as Sub-block size of 32*32 is applied to encryption
Image security [1] is mainly considered in this paper. algorithm. Decryption process is done by selection of the key
Encryption is a common technique to increase the security of to retrieve the original image. Improved security level of the
image. Image and video encryption have vast applications. encrypted images can achieve by increasing the image entropy
With the emerging growth in digital technologies and value and to decrease the high correlation among pixels. The
improvement of computer networks, a large amount of digital results for entropy and correlation of the encrypted images are
data is being transferred. A huge part digital data transferred calculated and evaluated various image inputs.
over network either be private or confidential information. In
today’s digital communication [2]- [3], exchange of date over
networks presents certain risks factor, which requires the II. RELATED WORK
existence of appropriate security measures. For example, the Guiliang Zhu et al. [11] proposed image encryption algorithm
images are transmitted and can be copied or saved during their based on pixels. First, scrambling the image pixels, then
transmission without loss of image quality. During an through the method of watermark increasing the difficulty of
exchange images can be hacked in time of digital information its decoded. At last, choose a camouflaged image to vision or
storage and reproduce illegally. It is therefore necessary to the pixels of the true image, getting the final encryption image.
develop software for effective protection of transferred data The key parameters are encrypted by Elliptic Curve
Ągainst arbitrary interference. Data encryption with image Cryptography (ECC).
merging is very often the only effective way to meet these
Laiphrakpam Dolendro Singh et al. [12] proposed an image
requirements.
encryption using Elliptic Curve Cryptography based on pixel
Most of the security algorithms specifically designed to grouping to reduce the number of computation. The group of
encrypt digital data are proposed in the mid-1990s. Encrypt pixels are transformed into big integer single digits keeping in
and decrypt of images can be done by different encryption mind that it should not exceed ‘p’ value which is one of the
algorithms. Most of the traditional [4] public key standards parameter in elliptic curve equation of finite field. These big
algorithm such as Rivest Shamir Adleman (RSA), private key integer values are paired and given as input denoted by ’Pm’ in
encryption standard (DES) and the family of elliptic-curve- ECC operation. This operation helps us to ignore the mapping
based encryption (ECC), as well as the international data operation and the need to share mapping table between sender
encryption algorithm (IDEA), can be classified into several and receiver.
Ahmed Bashir Abugharsa et al. [13] proposed an traditional method of generation as the product of very large
encryption algorithm based on the rotation of the faces of a prime numbers.
Magic Cube. This process involves dividing the original image
into six sub-images and further these sub images are divided B. Image Encryption
into small blocks and attached to the faces of magic cubes. The encryption procedure is based on encrypting the image
intensity and thus converting it into a new intensity. This new
Rogelio Hasimoto Beltran et al. [14] proposed interleaving
scheme where the de-correlation process is applied not only at intensity is decrypted to obtain the original intensity.
a block, but also at a co-efficient or pixel level in the
compressed domain. 1. Read the image and find the intensity I from the
image intensity matrix.
Frank Dellaert et al. [15] proposed image-based tracking 2. Convert the intensity of the image I into an elliptic
algorithm, which relies on the selective integration of a small curve point E using Mapping-1.
subset of pixels that contain a lot of information about the state 3. Elliptic curve point from Mapping-1 Encrypted to a
variables to be estimated. new point(E').
4. The new point (E') is converted to a corresponding
III. PROPOSED METHOD integer M, using Mapping-2.
5. This integer M is used to calculate the new encrypted
intensity I.
C. Image Decryption
1. The decryption is done by reverse process of
encryption.
2. The encrypted image intensity I’ is read from the
received files.
3. The intensity I parameter is to calculate the integer M.
4. Integer M is converted to encrypted elliptic curve
point E’, using reverse mapping-2.
5. The encrypted elliptic curve point E’ is decrypted to
get the original point E.
6. By reverse mapping-1, the original intensity I is
obtained.
D. Image Interleaving
The encrypted image from ECC is divided as 4*4 sub-
blocks, for better security the block [16] size should be small.
These sub-block images are interleaved column wise. The
number of sub-block pixel values in each block is fixed for a
given interleaver [17]. The interleaver operation on a set of
image pixel values is independent of its operation on all other
sets of symbols Applying block based interleaving by selecting
the initial location randomly.
In 2D type of interleaves [16], [17], the idea of extending
from 1D prime interleaved into two dimensional is utilized.
The concept of proposed 2D prime block based interleaved is
follows: Consider the two dimensional interleaving of nr by nc
matrix. Firstly, we divide the interleaving scheme into column-
wise interleaving and row-wise interleaving. Secondly, we
assign the value of seed as column-wise seed and row-wise
seed to column-wise interleaver and row-wise respectively.
Figure. 1. ECC Algorithm with Pixel Integration
Therefore, the location of bits after interleaving will be as
follows.
A. Elliptic Curve Cryptography Algorithm
Elliptical curve cryptography (ECC) [3] function is based
on public key encryption technique based on elliptic curve
theory that can be used to create efficient cryptographic keys
which is smaller and faster. Key generated through the
properties of the elliptic curve equation instead of the
Where prow and pcol are row and column-wise seeds. With
the new location of bits after interleaving on both row and
column-wise, the new locations are mapped back into 2D
interleaver to get the resulted interleaved bits in 2D.
Figure. 4 (a) Image Pixel value for 4*4 Image Size (b)Sub-blocks
image pixel value for (a)
E. Color Representation
16-bit Colour: Each pixel is represented [3] using 2 bytes or
16 bits. The bits are divided as red, blue and green each having
values i.e. 5-bits for red, 6-bits for green, and 5-bits for blue.
A. Correlation Co-Efficient
The correlation [18] is calculated between input image and
encrypted image, which is called the correlation coefficient.
The values -1 to +1 ranges the correlation coefficient value.
The encrypted image and original image are totally different, if
the correlation value of encrypted image is equal to zero or
very near to zero.The encrypted image is a reverse of the
original image, if the correlation is equal to -1. Correlation
coefficients were calculated by using the equation (1),(2) and
(3),
Figure. 7. Texture Input Image
(1)
Where x, y is input image and encrypted image values of
two adjacent pixels in the image. In numerical computation, the
following formulas were used
\ (2)
The obtained correlation coefficient for encrypted image is
shown in Table II and III.
Figure. 8. Encrypted Image
(3)
B. Information Entropy
Entropy [19] is defined to express the degree of
uncertainties in the system. A secure encryption of encrypted
Figure.9. Decrypted Image with Key No.2
image should not provide any information about the original
image. Higher entropy images such as an image of heavily
cratered on the moon have a great deal of contrast from one
pixel to the next and consequently pixel cannot be compressed
as much as low entropy images. Entropy H indicated that each
symbol has an equal probability. The information entropy for
encrypted image is calculated using equation (4),
V. CONCLUSION
Digital image security has become highly important since
the communication by transmitting of digital products over the
network occur very frequently. The image encryption
algorithm is proposed, based on pixels interleaving with image
Figure.13. Decrypted Image with Key No.5 integration in this paper. First, interleaving the image pixels,
then through the method of pixel integration increasing the
difficulty of decoded. At last, a camouflaged image for all the
input images, getting the final encryption image. Experimental
result shows good performance with low correlation and high
entropy which shows that the pixel based algorithm is highly
secure. With this approach it is also able to encrypt large
volume of data more securely and simultaneously. Our new
approach is expected to be useful for transmission applications
Figure.14. Decrypted Image with Key No.6 and real time system. Future work includes the incorporation of
other encryption algorithm and extending the images to videos.
C. Result Analaysis
REFERENCES
High entropy value and low correlation values provides
[1] Christof Paar and Jan Pelzl,” Understanding Cryptography: A Textbook
good encryption. The time taken for encrypting and decrypting for Students and Practitioners," Springer, 2010,pp.1-24
mosaic image with key is 262.4035 seconds and for texture [2] Darrel Hankerson , Alfred J. Menezes and Scott Vanstone," Guide to
image is 614.3073 seconds. Elliptic Curve Cryptography,” Spinger,2004
Results for the correlation and the entropy values are shown [3] Chris Solomon and Toby Breckon,"Fundamentals of Digital Image
Processing,” Wiley,2010, pp1-18
in Tables I and II.
[4] I. Ozturk and I. Sogukpinar,"Analysis and comparison of image
encryption algorithm,”Journal of transactions on engineering, computing
and technology, pp.38, Dec 2004.
[5] Rinki Pakshwar, Vijay Kumar Trivedi and Vineet Richhariya,"A Survey Annamalai University, chidamabram, India and has put in 15
On Different Image Encryption and Decryption Techniques,"
International Journal of Computer Science and Information
years of service. She has produced one Ph.D and guiding 6
Technologies, pp.113-116, April 2013. Ph.D scholars. Her research interests are in power converters,
[6] G.Zhi-Hong, H.Fangjun, and G.Wenjie, "Chaos-base, Image Encryption control techniques for multiple connected power converters,
Algorithm,”Elsevier, pp. 153-157, Oct 2005. embedded controllers for power converters, renewable energy
[7] Norman D. Jorstad, "Cryptographic Algorithm Metrics,"Institute for based power converters. She is a life member of Indian society
Defense Analyses Science and Technology Division-Jan 1997. for technical education.
[8] Li. Shujun and X. Zheng,"Cryptanalysis of a chaotic image encryption
method,"IEEE International Symposium on Circuits and Systems,
ISCAS, May 2002.
[9] Moncef Amara and Amar Siad,"Elliptic Curve Cryptography and its
applications ",Systems, Signal Processing and their Applications
(WOSSPA),pp 247-250,May 2011.
[10] Kamlesh Gupta and Sanjay Silakari,"Performance Analysis for Image
Encryption Using ECC",Computational Intelligence and
Communication Networks (CICN),pp 79-82,Nov 2010
[11] Guiliang Zhu ,Weiping Wang ,Xiaoqiang Zhang and Mengmeng Wang
,”Digital Image Encryption Algorithm Based on Pixels”,IEEE 2010.
[12] Laiphrakpam Dolendro Singh and Khumanthem Manglem Singh,"Image
Encryption using Elliptic Curve Cryptography"Eleventh International
Multi-Conference on Information Processing,pp 472-481,2015.
[13] Ahmed Bashir Abugharsa, Abd Samad Bin Hasan Basari and Hamida
Almangush,"A Novel Image Encryption Using an Integration Technique
of Blocks Rotation Based on the Magic Cube and the AES Algorithm,"
International Journal of Computer Applications, pp.38-45, March 2012.
[14] Rogelio Hasimoto-Beltran and Ashfaq Khichari,"Pixel Level
Interleaving scheme for Robust Image Communication," Scalable and
Parallel Algorithm Labs, University of Delaware, Newark, Oct 1998.
[15] Frank Dellaert and Robert Collins, "Fast Image-Based Tracking by
Selective Pixel Integration,"Computer Science Department and Robotics
Institute Carnegie Mellon University,Pittsburgh,Sep 1999
[16] Hanpinitsak and C. Charoenlarpnopparut,"2D Interleaver Design for
Image Transmission over Severe Burst-Error Environment,"
International Journal of Future Computer and Communication, pp.308-
312, Aug 2013.
[17] Shengyong Guan, Fuqiang Yao and Chang Wen Chen,"A novel
interleaver for image communications with theoretical analysis of
characteristics," Communications, Circuits and Systems and West Sino
Expositions,IEEE 2002 International Conference (Volume:1), July 2002.
[18] Satoru Yoneyama and Go Murasawa, "Digital Image Correlation,"
Encyclopedia of Life Support Systems, Digit Imaging. 2008 Sep.
[19] Du-Yih Tsai, Yongbum Lee and Eri Matsuyama,“Information Entropy
Measure for Evaluation of Image Quality", 2008 Sep.
AUTHORS PROFILE
Abstract— Gait feature is a relatively useful biometric for and important tasks in gait recognition. Two different families
application in security surveillance, because it can be obtained by of approaches for gait recognition have been proposed:
a camera just kept at a long distance away without disturbing appearance-based and model-based methods. Appearance-
that person. Gait is robust feature even if people change their based approaches use captured image sequences directly to
appearance, on the contrary appearance features are helpless in
those situations. The problem of multi-view human gait
extract gait features, while model-based methods extract
recognition along any straight walking paths is investigated in model parameters from the 2 images. Several methods have
this work. It is observed that the gait appearance changes as the been proposed for gait recognition from various different
view changes, while a certain amount of correlated information perspectives. The state of the art methods mainly fall into
exists among different views. This work is considered as a three categories namely, view- invariant models estimation,
classification problem, where the classification features are the viewing angle rectification and view transformation model
elements of the transformation matrix that is estimated by the methods. One example that estimates the view- invariant
Transformation Invariant Low-Rank Texture (TILT) algorithm. model is the Joint Subspace Learning (JSL) method, where the
Later on the gallery gait appearances are converted to the view of view-invariant model is represented by a weighted sum of
the probe subject, where the spatially neighboring pixels of the
gait feature are considered as the correlated information of the
sufficiently small number of prototypes of the same view.
two views. Finally, a similarity measurement is applied on the Similarly, a distance metric is learned with a good
converted gait appearance and the testing gait appearance. The discrimination ability based on the clustered and averaged
proposed method is tested on the CASIA-B multi-view gait GEI. However, this kind of methods can achieve good
databases to examine how the proposed gait recognition method recognition results only for similar views. When the probe gait
performs under most views. sequences are significantly different from the gallery
sequences, they usually have poor performances.
Keywords-component; formatting; style; styling; insert (key
words) The main disadvantages of appearance features can
be categorized as follows: 1) Due to the change in view angles
I. INTRODUCTION and illumination, appearance of people changes largely. 2)
Color can get distorted by several camera parameters that may
Gait is one of the well recognized biometric features result in the variation of the same color as a different color in
to re-identify a human at a distance from the camera. A large different cameras. 3) The Peoples clothing condition may also
number of successful gait recognition techniques have been vary according to the season and occasion, which may cause
continuously contributed. However, appearance changes of variation in the way people, appears. The main objective of
individuals due to viewing angle changes cause many this work is to design a system for gait recognition using
difficulties for 2D appearance-based gait recognition. This model free methods. For recognition purpose, the effective
situation cannot be easily avoided in a practical surveillance gait feature GEI is used. The walking path image of the gait
application. There are three major approach categories to solve sequence is constructed. The low rank textures for the features
the problem, namely; 1) seeking a gait feature invariant to are obtained using the TILT algorithm. The features obtained
view change; 2) reconstructing gait under any viewing angle are classified to make the proposed method more
using 3D structure information, which is obtained by discriminative and effective.
calibration; 3) projecting gait feature from one viewing angle
to the other by using view transformation. Gait features of
uncooperative subjects may contain covariates that influence II. RELATED WORKS
the gait itself and/or the appearance of a walking person, so Chen et al. [1] proposed a novel cross-view gait
the robustness to such covariates is quite important for recognition method based on projection of Gravity Center
accurate gait recognition. Among the covariates, a change in Trajectory (GCT). The coefficients of 3-D GCT in reality to
view occurs frequently for real situations and has a large different view planes for the complete view variation are
impact on the appearance of the walking person. Matching gait projected. The view of a silhouette sequence is estimated using
across different views is therefore one of the most challenging this matrix to complete view variance of gait features.
Calculation of the body part trajectory on silhouette sequence (FDFP), which is robust to high frequency noise. Zheng Liu et
is done to improve recognition accuracy by using correlation al.[10] proposed a framework consists of the hierarchical
strength as similarity measure.Finally nested match method is feature extraction and descriptor matching with learned metric
used to calculate the final matching score of two kinds of matrices.
features. Huang et al. [2] proposed the concept of extreme
Burhan et al.[11] used the concept of Multi-view gait
learning classification for regression and multi class
recognition using enhanced gait energy image and radon
classification. Both LS-SVM and PSVM can be simplified
transform techniques. A gait representation model for multi-
further and a unified learning framework of LS-SVM, PSVM
view gait recognition systems based on Gait Energy Image
and other regularization algorithms referred to Extreme
(GEI) and Radon Transform (RT) on human silhouettes. The
Learning Machine (ELM) can be built. Zhang et al.[3]
recognition of gait is based on similarity in measurements
proposed the low rank textures capture geometrically
using Euclidean distance. Zeng et al.[13] used the concept of
meaningful structures in an image, which encompass
silhouette based gait recognition through deterministic
conventional local features such as edges and corners as well
learning. The new silhouette based gait recognition via
as all kinds of regular, symmetric patterns ubiquitous in urban
deterministic learning is proposed in order to combine spatio-
environments and man-made objects. Worapan et al.[4]
temporal motion characteristics and physical parameters of a
proposed the Gait as one of the well recognized biometrics
human subject for recognition. Lu et al. [14] proposed a
that has been widely used for human identification. This paper
method based on joint distribution of motion angles is
proposes a new multi-view gait recognition approach which
proposed for gait recognition.
tackles the problems mentioned above. This method differs
from other by creating a so called View Transformation Model
(VTM) based on spatial-domain Gait Energy Image (GEI) by III. SYSTEM ARCHITECTURE
adopting Singular Value Decomposition (SVD) technique. To This section explains the system architecture for the
further improve the performance of the proposed VTM, Linear recognition of gait in arbitrary views by extracting features
Discriminant Analysis (LDA) is used to optimize the obtained from the surveillance videos.
GEI feature vectors.
A. Gait Recognition Systems for arbitrary views
Yibo Li et al. [5] proposed a gait recognition method The architecture of the system proposed in this work
based on ankle joint motion trajectory and bending angle. First consists of major components like feature extraction,
it obtains lower limb joint points according to each part of the transformation matrix generation and classification using ELM
body and height proportion. It obtains the position coordinates as shown in Figure 1.
of the toe by using skeleton algorithm. The feature vector is
made up of the relative velocity of ankle joint motion
trajectory and the bending angle. Support Vector Machine
(SVM) classifier and the Nearest Neighbor (NN) classifier are
used for the gait classification. Worapan et al.[6] proposed the
approach using regression-based View Transformation Model
(VTM) is proposed to address this challenge. Gait features
from across views can be normalized into a common view
using learned VTM(s). In principle, a VTM is used to
transform gait feature from one viewing angle (source) into
another viewing angle (target).
Sruti Das et al.[7] proposed the concept of a two-
phase View-Invariant Multi scale Gait Recognition method
(VI-MGR) which is robust to variation in clothing and Figure 1. Gait Recognition System for Arbitrary Views
presence of a carried item. Wei Zeng et al. [8] proposed the
method eliminates the effect of view angle for efficient gait The framework consists of feature extraction, transformation
recognition through deterministic learning theory. The width matrix generation using TILT algorithm and ELM
of the binarized silhouette models, the periodic deformation of classification. Feature extraction phase involves the extraction
human gait shape are selected as the gait feature. Makoto et of Gait Energy Image (GEI) feature. Gait Energy Image [12]
al.[9] proposed a novel gait recognition approach, which is generated by the summation of silhouette images to obtain
differs a lot from existing approaches in the subjects the spatio-temporal information. This process is followed by
sequential 3D models and his/her motion are directly the convex hull creation and construction of walking path
reconstructed from captured images. Arbitrary viewpoint image (WPI). TILT algorithm is used for transformation
images are synthesized from the reconstructed 3D models for matrix generation and then classification is done using ELM.
the purpose of robustness in gait recognition to changes in the
walking direction is proposed. Moreover, a gait feature is
proposed in this work named Frame Difference Frieze Pattern
B. System Description walking path information of the subject. This is because the
The description of each module based on the system TILT algorithm is much more computation efficient with a
architecture for the recognition of gait in arbitrary views by low rank image input.
model free method is explained in this section. This section C. Transformation Matrix Generation
includes the feature extraction process, usage of TILT
The walking path image is taken as the input for the
algorithm of transformation and finally classification with the
transformation matrix generation using TILT algorithm. The
help of ELM classifier.
main aim of this algorithm is to efficiently and effectively
The videos of large number of people are used as the
extract a class of low rank textures in a 3D scene from 2D
input for preprocessing phase. The obtained input video is
images. The low-rank textures capture geometrically
converted into frames and then background subtraction
meaningful structures in an image, which includes local
process is done to the frames in order to obtain the foreground
features such as edges and corners as well as all kinds of
images. The foreground images are the silhouette images
regular, symmetric patterns. Usage of Transformation
which are used as the input for further processing.
Invariant Low Rank Texture (TILT) algorithm [3] is to get the
Feature extraction involves reducing the amount of
low rank image of the walking path image. The coordinates
resources required to describe a large set of data. When
for this process is given graphically to the image during
performing analysis of complex data, one of the major
execution. The output produced by the TILT algorithm has
problems stems from the number of variables involved in the
been considered as features for the classification process. This
extraction. Analysis with a large number of variables generally
method can accurately recover both the low-rank texture and
requires a large amount of memory and computation power or
the domain transformation.
a classification algorithm which over fits the training sample
and generalizes poorly to new samples. Feature extraction is a D. Extreme Learning Machine (ELM)
general term for methods of constructing combinations of the The transformation matrix generated by the
variables to get around these problems while still describing Transformation Invariant Low Rank Texture (TILT) algorithm
the data with sufficient accuracy. The features which are is considered as the input for this module. The main objective
extracted for the entire process includes extraction of GEI, of Extreme Learning Machine (ELM) is that the weights of the
construction of convex hull followed by the creation of hidden layer can be initialized randomly, which makes the
Walking Path Image(WPI). optimization of the weights of the input layer and the biases of
1) Gait Energy Image the output layer. The time and accuracy of both training and
The input silhouettes obtained after the preprocessing testing is obtained as output after the classification. Extreme
stage is used as the input for the construction of Gait Energy Learning Machine (ELM) was originally developed for the
Image (GEI). It is obtained by the summation of all the single hidden layer feed forward neural networks and then
silhouettes of a person during the gait cycle divided by the extended to the generalized SLFN as per equation 3.1 which
total number of frames in that gait cycle. The resultant GEI is may not be neuron alike
of 240*320 dimensional vector.
2) Convex Hull (CH) Creation
The GEI obtained is used as the input for this Where h(x) is the hidden-layer output corresponding to the
process. The convex hull for the Gait Energy Image (GEI) is input sample x and is the output weight vector between the
obtained as the output. hidden layer and the output layer. One of the salient features
of ELM is that the hidden 18 layer need not be tuned.
Essentially, ELM originally proposes to apply random
computational nodes in the hidden layer, which are
independent of the training data. ELM and its variants mainly
focus on the regression applications. In ELM the weights
connecting inputs to hidden nodes are randomly assigned and
never updated. The ELM achieves good generalization
Figure 2. Convex Hull Figure 3 Sample Image after TILT
performance and solves the regression problem.
The formal definition of convex hull is that the IV. IMPLEMENTATION
smallest convex polygon S contains all the points of S as
This section discusses about the implementation of gait
depicted in figure 2.
recognition method in the arbitrary view. Here, the feature
3) Walking Path Image
extraction process involves extraction of GEI from the
The convex hull created for the Gait Energy Image is
silhouettes followed by the convex hull creation and
considered as the input for the construction of Walking Path
construction of walking path image. The transformation matrix
Image (WPI). The image representing the walking path of the
is obtained for WPI using TILT algorithm followed by the
subject can be constructed by averaging the gait sequence of
ELM classifier, which is then used for classification.
the subject. The main aim is to construct a relatively low rank
image with a sequence of gait images while preserving the
A. Feature Extraction Walking Path Image (WPI) of the probe subject. The WPI
The videos of people’s gait obtained from the provides a relatively low rank version of image while
surveillance networks is converted into frames and reserving the walking path information of the probe subject.
background subtraction is done in order to extract the D. Transformation Matrix Generation
foreground images that is the silhouettes which is used as the
The Transform Invariant Low-Rank Textures (TILT)
input for the feature extraction process.
technique is used to estimate the transformation matrix for
The normalized silhouette is considered as input for this
transforming the WPI into a low rank image. The Walking
module. The silhouette sample obtained from the dataset as
Path Image (WPI) which reveals the walking path information
shown in figure 4.1 is taken as the input for this process. Gait
of the subject is given as input for the generation of low rank
Energy Image (GEI) Gait Energy image is the gait feature
textures as well as the transformation matrix. The angle for the
used in this system. An average of all the images over a gait
TILT of an image is specified by the user graphically by
cycle for a time period is called as gait energy image (GEI) as
marking the coordinate points of the image during the process.
denoted in equation 4.1. It represents the silhouette shape and
The transformation matrix obtained is considered as the
motion information in a single frame by maintaining its
features of different viewing angles and are used to train the
temporal information.
ELM classification model. The transformation matrix of the
WPI of the new coming subject is estimated and fed into the
trained classifiers so that its viewing angle can be estimated.
Where Nij is the number of frames and Sijt is the silhouettes. This transformation matrix describes the degree of difficulty
The following figure 5 depicts the gait energy image of a of the transformation. Moreover, subjects under the same view
person for a number of frames which is based on the persons have similar transformation matrix and subjects of different
gait cycle. GEI reflects major shapes of silhouettes and their views have distinct transformation matrix. According to the
changes over the gait cycle. It is referred as gait energy image TILT algorithm, any image can be transformed to its low rank
because of the following reasons: Each silhouette image is the version and the low rank image can be recovered from its
space-normalized energy image of human walking at this deformed image. Since the transformation matrix reveals the
moment. GEI is the time-normalized accumulative energy degree of transformation, it is able to be employed as the
image of human walking in the complete cycle(s).A pixel with feature for the classification process.
higher intensity value in GEI means that human walking
occurs more frequently at this position. E. ELM Classification
The transformation matrix generated after applying
B. Convex Hull Creation the TILT algorithm to the Walking Path Image (WPI) is stored
The Gait Energy Image (GEI) which is extracted by into file. The values in the file are being read for the
adding all the frames during a gait cycle divided by the total classification process which is done using Extreme Learning
number of frames from the input silhouettes is given as input Machines (ELM) classifier. The ELM classifier is used for
for the creation of convex hull. Then the GEI is converted into classification or regression purpose with a single layer of
binary image by setting a threshold value, the resultant hidden nodes, where the weights connecting inputs to hidden
obtained is used for further processing. The convex hull is nodes are randomly assigned and never updated. During the
depicted in figure 4. classification process the value of the image which is to be
tested is fed into the classifier using another such as test data.
The number of training classes used for the classification is 60
and the testing class is 6. The person is identified in 6 viewing
angles that is 18, 36,54, 72, 90, 108.
After classification using ELM the time and accuracy
of the training and testing data is obtained which is further
used for the recognition process. These models produce good
generalization performance and learn faster than networks
trained using back propagation. The simplest ELM training
Figure. 4. Convex hull and walking path of the image algorithm learns a model of the form
Yˆ = W2σ(W1x)
where W1 is the matrix of input-to-hidden-layer weights, σ is
C. Walking Path Image some activation function, and W2 is the matrix of hidden-to-
The image of the convex hull obtained from the GEI output-layer weights.
is given as input for the construction of Walking Path Image. The summarization of the ELM algorithm is as follows:
The upper and lower boundary of the WPI reveals the walking Given a training set, N = (xi,ti) | i = 1....N, kernel function f(x)
path of the subject and the left, right boundary reveals the and hidden neuron.
body movement which is not necessary for the estimation. Step 1: Select suitable activation function and number of
Therefore, the border protruding parts are cropped in order to hidden neurons for the given problem.
achieve a quadrangle like shape. The figure 4 denotes the
Abstract— With the advent of technology, digital image applications. In this work, a blind watermarking scheme is
watermarking turns to be an effective technique for the presented. Based on the robustness of the algorithm,
protection of digital images against illegal use and duplication. watermarking can be classified into three categories: robust,
Watermarking in wavelet domain has drawn significant attention fragile and semi-fragile. A robust watermarking should be able
due to its multiresolution attributes. In this paper, analysis has
been presented on a digital image watermarking scheme that
to survive different kind of attacks. Watermarking can be
combines the features of Discrete Wavelet Transform (DWT) and performed either in spatial or frequency domain. However, this
Discrete Cosine Transform (DCT). A binary watermark image is study focused on developing digital image watermarking
embedded into certain sub-bands of a 3-level DWT transformed scheme in transform domains. Unlike spatial-domain
DCT coefficients of a host image. To attain maximum watermarking, transform domains are complex however gives
imperceptibility, all the 3-level DWT sub-band covering lower to better robustness. The transform domain method based on
higher frequency range are explored to find the optimal DWT discrete wavelet transform (DWT) [3–4], discrete cosine
sub-band suitable for embedding a binary watermark. transform (DCT) [5–6], discrete Fourier transform (DFT) [7–
Eventually, the same procedure as of the embedding process is 8], and singular value decomposition (SVD) [9–10] utilizes its
applied to extract the DCT middle frequencies in each sub-band
also. The watermark bit is then determined based on the
signal characteristics and human perception properties so as to
correlation between DCT mid-band coefficients and PN- attain better robustness and invisibility. Since the inception of
sequences and is again processed by the 2-D generated key to the very development, many watermarking algorithms have
obtain the actual watermark. been proposed based on a combination of all the aforesaid
transforms [11–13].
Keywords- Discrete Cosine Transform (DCT); Discrete Wavelet
Transform (DWT); Encryption; PN-Sequences; Watermarking. The DWT is a renowned transform for image processing
because of its multi-resolution properties in time and
frequency. The DCT gives tremendous energy compaction for
highly correlated image data. Using the combination of both
I. INTRODUCTION DWT and DCT for watermarking generally exhibits good
With the growing development of the internet and performance in regards of both the invisibility and robustness.
multimedia outfits over the past few decades, the access and However, apart from these two transforms, SVD is another
the unauthorized copying, modification and distribution of potential numeric tool that can be useful for applications like
digital data has become easier and widespread. The easy access data hiding and image compression. But it is not always
of multimedia data brings itself with the challenge of content possible to use SVD since most of the algorithm based on it are
protection. The increasing demand of copyright protection has non blind or semi blind in nature [14-15].
drawn the attention in digital watermarking [1-2], in which it Literature studies reveals that several research works have
embeds information for ownership claim and authentication been proposed to develop robust watermarking algorithms in a
into digital data to avert from illegal copying. Based on the combined domain involving the DWT, DCT and SVD. The
extraction algorithm, watermarking scheme can be broadly developed techniques differ not only in embedding targets but
classified into two distinct categories: non-blind or blind also in processing measures. DWT is used to decompose the
depending. In blind algorithm, the original image is not host image into LL, LH, HL, and HH sub-bands. Bhatnagar et
necessary for watermark extraction. Whereas in non-blind al. [16] has chosen the LL sub-band for watermark embedding.
algorithm, the original image is required during extraction Whereas, Ali et al. [17] has selected the LL and HH sub-band,
stage which may not be suitable for many practical while Laskar et al. [18] has considered the mid frequency LH
and HL sub-band. So, in this paper, thorough analysis and
Figure 1. The different set of 3-level DWT sub-bands used for embedding a 32×32 binary watermark
In these sub-bands representation, LL, HL, LH and HH 8.1 based Matlab(R2013a). The hardware used is Dell
represent application, horizontal, vertical and determinant sub- computer with RAM 16 GB and Intel core i5 processor.
band of first level DWT decomposition respectively. Similarly,
LL1, HL1, LH1 and HH1 represent application, horizontal, There are total two set of experiments carried out for the
vertical and determinant sub-band of 2-level DWT joint DWT-DCT based image watermarking scheme.
decomposition respectively. Likewise, LL2, HL2, LH2 and 1. In this experiment, watermarking is performed on
HH2 represent application, horizontal, vertical and determinant different set of DWT sub-bands to find the sub-bands
sub-band of 3-level DWT decomposition respectively. A sub- suitable which gives better imperceptibility and
band HL2/HL1/HL implies the 3-level horizontal DWT sub- robustness (Experiment I).
band present in second level horizontal DWT sub-band which 2. The robustness of the algorithm has been observed
itself is present in first level horizontal DWT sub-band.
against various attacks.(Experiment II)
D. Performance against different attacks.
The performance of the algorithm against different The performance is measured in terms of visual observation
intentional and non-intentional attacks has been studied. The and peak signal to noise ratio (PSNR), mean square
attacks include various image processing attacks, noising error(MSE) and structural similarity index measure (SSIM) for
attacks, geometrical attacks, filtering, JPEG compression etc. both the watermarked image and the extracted watermark.
Image processing attacks comprise histogram equalization, Normalized cross-correlation is used to check robustness
contrast enhancement, image sharpening, gamma correction against various attacks. The PSNR and MSE are calculated as
and geometrical attacks include cropping, rotation, resizing, mentioned in [18].
translation etc. The experiment has been performed using sizes Instead of using traditional error summation methods, the
of watermark over a variety of host images. The performance is SSIM is designed by modeling any image distortion as a
measured in terms of normalized cross correlation (NC) combination of three factors that are loss of correlation,
between original watermark and extracted watermark. luminance distortion and contrast distortion [19]. The SSIM
III. RESULTS AND DISCUSSION index is calculated as discussed in [20].
The imperceptibility is calculated between original host
The joint DWT-DCT based watermarking scheme is tested image and watermarked image in terms of performance
on 300 gray images of 512×512 size and 200 binary measures like PSNR, MSE and SSIM. Similarly, Performance
watermarks of variable size. These images are collected from measures are compared between the original watermark and the
http://www.imageprocessingplace.com, USC-SIPI image extracted watermark for different values of α. From the
database and CVG-UGR image database. The binary experimental results and visual observation, it has been
watermarks are collected from the MPEG7_CE_shape observed that the performance of the algorithm for
descriptor database along with our collected database. Samples watermarking is satisfactory. Moreover, by choosing a suitable
of gray image of 512×512 size are shown in figure 2. Samples value of α, it has been observed that the algorithm works for
of binary watermarks are shown in figure 3. This image different variety of gray scale images. However, a trade-off has
watermarking technique has been tested for different standard to be made between the quality of the watermarked image and
images like Lena, Baboon, Barbara, biomedical images, Aerials the extracted watermark while choosing the value of α. Higher
image, texture image, Astronomical galaxies image etc. The value of α degrades the quality of the watermarked image but
detailed analysis and experimentation is performed in windows gives a better extracted watermark image and vice versa. In
result tables, HIGH indicates infinite PSNR value.
Identify applicable sponsor/s here. (sponsors)
Figure 2. Samples of gray test images collected from the Computer Vision Group at the University of Granada and USC-SIPI image database.
A. Experiment I: Watermarking on different set of DWT sub- sub-band for the optimum gain factor α= 8. Similarly, the
bands SSIM between original watermark and extracted watermark
The embedding is performed on different combination of with sub-band is shown in figure 6. Here the graphs are shown
sub-bands but the representative result is shown for 9 different for five different cover images: Lena(standard image), Brick
set of sub-bands. The results are shown for the algorithm using wall(texture image), Chest X-ray(Biomedical Images), Galaxy
haar wavelet only. Table 1 and table 2 show the performance image and Aerial image. From PSNR and SSIM graph for
of the algorithm on watermarked image and extracted imperceptibility and robustness, it has been observed that sub-
watermark for different set of sub-bands respectively. In table band 8 which is high frequency diagonal sub-band, gives
1, the performance parameters MSE, PSNR and SSIM are optimum values of both the imperceptibility and robustness.
calculated between watermarked image and original image for Figure 7 shows the shows watermarked Lena image and
different values of gain factor α. Similarly table 2 shows extracted watermark for different sub-band for α= 5.
performance parameters between the original watermark and B. Experiment II: Robustness against different attacks
extracted watermark under no attack condition. From the result
In this experiment, the robustness of the scheme against
it can be said that imperceptibility increases when embedding
different intentional and non-intentional attacks have been
is performed in higher frequency sub-bands. In this experiment
studied. Eight different image processing operations are chosen
, it is found that sub-band 8 which is 3-level sub-bands present
as attacks to evaluate the robustness of the proposed scheme.
in high frequency diagonal bands are the most optimal sub-
These are: 1) histogram equalization (HE), 2) gamma
band in terms of imperceptibility and robustness. The sub-band
correction (GC) with 1.8, 3) JPEG compression with quality
8 gives highest average PSNR and SSIM values of 50.3090 dB
factor 90, 4) salt and pepper noise(SPN) (0.1%), 5) image
and 0.9996 respectively over 300 gray images. This is because
sharpening(IS) with radius 0.5 and amount 2.0, 6) Contrast
the high frequency determinant bands carries edge information
enhancement(CE) (default), 7) cropping(CR) one eighth from
and effect of watermarking is least on the watermarked image.
The algorithm gives poor performance if we embed the left side, 8)image resizing (RS)(512→ 256→ 512). This
watermark in low frequency approximation sub-bands. Sub experiment is performed on 300 gray images. The robustness is
band 1 lies in the low frequency LL band. Since most of tested by embedding different sizes of watermark. For the sake
images have higher low frequency contents, so the effect of of briefness, the results are shown only 10 gray images of size
embedding is more visible if we modify the coefficients of sub 512×512 where a binary watermark of size 32×32 is embedded
band 1 for embedding binary bits of watermark. So, it provides in sub band 8. Table 3 describes the performance of this
high value of MSE for sub band 1. Imperceptibility decreases algorithm in terms of normalized cross-correlation (NC).
with increasing gain factor α where robustness increases with Normalized cross-correlation is calculated in between extracted
increasing gain factor α. This experiment is performed for α watermark and original watermark under different attack
values ranging from 1 to 10 to find the most optimum values of condition. The performance of this algorithm is satisfactory
α in terms of imperceptibility and robustness. The optimum against image processing attacks like histogram equalization,
value of gain factor is found to be equal to 8 for this joint gamma correction, image sharpening and contrast
DWT-DCT based image watermarking scheme. The enhancement. But this algorithm fails to give adequate
imperceptibility of different watermarked images in terms of performance against JPEG compression, noising, cropping,
PSNR and SSIM is shown in figure 4 and figure 5 for different resizing etc.
TABLE 1. PERFORMANCE COMPARISON OF DIFFERENT SET OF BANDS FOR THE WATERMARKED IMAGE FOR DIFFERENT VALUES OF Α
AVERAGED OVER 300 IMAGES
α= 5 α= 5 α= 5 α=8 α= 8 α= 8 α= 10 α= 10 α= 10
Set of bands
MSE PSNR SSIM MSE PSNR SSIM MSE PSNR SSIM
Sub-band 1 25.8516 35.8618 0.9225 25.9609 35.8161 0.9210 26.0533 35.7765 0.9196
Sub-band 2 6.5441 41.9086 0.9748 6.6462 41.9086 0.9689 6.7366 41.7627 0.9673
Sub-band 3 1.9439 47.0981 0.9993 2.0416 46.6856 0.9992 2.1367 46.3357 0.9992
Sub-band 4 1.8454 47.6016 0.9994 1.9411 47.1139 0.9992 2.0367 46.7162 0.9992
Sub-band 5 1.5548 48.2103 0.9995 2.1367 46.3357 0.9992 1.7498 47.2314 0.9993
Sub-band 6 1.4049 48.6122 0.9995 1.5068 48.0095 0.9994 1.5955 47.5640 0.9993
Sub-band 7 1.2277 49.3919 0.9996 1.3276 48.6476 0.9995 1.4202 48.1127 0.9994
Sub-band 8 0.7145 51.4232 0.9997 0.8174 50.3090 0.9996 0.9110 49.5921 0.9995
Sub-band 9 0.7514 51.1948 0.9997 0.8526 50.1388 0.9996 0.9447 49.4415 0.9995
TABLE 2. PERFORMANCE COMPARISON OF DIFFERENT SET OF BANDS FOR THE EXTRACTED WATERMARK FOR DIFFERENT VALUES OF
Α AVERAGED OVER 300 IMAGES.
α= 5 α= 5 α= 5 α=8 α= 8 α= 8 α= 10 α= 10 α= 10
Set of bands
MSE PSNR SSIM MSE PSNR SSIM MSE PSNR SSIM
Sub-band 1 25.8516 35.8618 0.9225 25.9609 35.8161 0.9210 26.0533 35.7765 0.9196
Sub-band 2 6.5441 41.9086 0.9748 6.6462 41.9086 0.9689 6.7366 41.7627 0.9673
Sub-band 3 1.9439 47.0981 0.9993 2.0416 46.6856 0.9992 2.1367 46.3357 0.9992
Sub-band 4 1.8454 47.6016 0.9994 1.9411 47.1139 0.9992 2.0367 46.7162 0.9992
Sub-band 5 1.5548 48.2103 0.9995 2.1367 46.3357 0.9992 1.7498 47.2314 0.9993
Sub-band 6 1.4049 48.6122 0.9995 1.5068 48.0095 0.9994 1.5955 47.5640 0.9993
Sub-band 7 1.2277 49.3919 0.9996 1.3276 48.6476 0.9995 1.4202 48.1127 0.9994
Sub-band 8 0.7145 51.4232 0.9997 0.8174 50.3090 0.9996 0.9110 49.5921 0.9995
Sub-band 9 0.7514 51.1948 0.9997 0.8526 50.1388 0.9996 0.9447 49.4415 0.9995
head CT-
0.7566 0.3963 0.2293 0.7689 0.8642 0.8616 0.6943 0.2052
Vandy
(a) (b)
(c) (d)
(e) (f)
(g) (h)
Figure 7. Lena watermarked image and extracted watermark for different set
of sub-bands for α= 5.(a) Watermarked image for sub band 1,(b) Watermarked
image for sub band 3, (c) Extracted watermark for sub band 1, (d) Extracted
watermark for sub band 3, (e) Watermarked image for sub band 5, (f)
Watermarked image for sub band 8, (g) Extracted watermark for sub band 5,
(h) Extracted watermark for sub band 8.
IV. CONCLUSION
A detail analysis on joint DWT-DCT based digital image
watermarking scheme has been presented in this work. The 3-
level DWT transformed DCT coefficient is modified to embed
a watermark. This blind algorithm provides a maximum
Figure 5. SSIM of different watermarked image with sub-band for α= 8. average (over 300 images) imperceptibility of 50.3090 dB
when the watermark (32×32) is embedded in the most
optimum sub-bands of the host image. The sub-band 8 is
found as the optimal sub-band that is taken as diagonal sub-
bands of high frequency (HH) band. Both the subjective and
objective evaluation analysis show that the performance of the
watermarking algorithm is satisfactory to achieve
imperceptibility and robustness. With the increase of gain
factor, though imperceptibility decreases but robustness
increases and vice-versa. Gain factor α=8 provides the
optimum performance for both imperceptibility and robustness
in terms of quality measurement. It has been observed that
satisfactory performance using this algorithm is achieved
against various attacks likely histogram equalization, image
sharpening, contrast enhancement, gamma correction but fails
to achieve adequate performance against geometrical attacks,
JPEG compression, de-noising.
Figure 6. SSIM of extracted watermark for five different cover image with As an extension of this work, we will try to implement a
sub-band for α= 8.
robust system that can resists against geometrical attack, JPEG
compression, filtering etc keeping the imperceptibility in [13] J. Varghese, S. Subash, O. B. Hussain, K. Nallaperumal,
satisfactory level. Also this algorithm may be extended for M. R. Saady, and M. S. Khan, "An improved digital
embedding and extraction of a gray scale watermark in a color image watermarking scheme using the discrete Fourier
image. transform and singular value decomposition," Turkish
Journal of Electrical Engineering & Computer Sciences.
ACKNOWLEDGMENT
2015
The authors would like to acknowledge the all the members [14] O. Jane, and E. ELBAŞI, "A new approach of nonblind
of Speech and Image Processing Laboratory, Department of watermarking methods based on DWT and SVD via LU
Electronics and Communication Engineering, National Institute decomposition," Turkish Journal of Electrical
of Technology Silchar, India for providing support and Engineering & Computer Sciences. 2014 Sep
necessary facilities for carrying out this work.. 3;22(5):1354-66.
REFERENCES [15] J. M Shieh, D. C. Lou, and M. C. Chang, "A semi-blind
digital watermarking scheme based on singular value
decomposition," Computer Standards & Interfaces. 2006
[1] I. J. Cox, M. L. Miller, J. A. Bloom, and C. Honsinger, Apr 30;28(4):428-40.
Digital watermarking vol. 53, 2002. [16] G. Bhatnagar, and Q. J. Wu, "A new logo watermarking
[2] M. U. Celik, G. Sharma, E. Saber, and A. M. Tekalp, based on redundant fractional wavelet transform,"
"Hierarchical watermarking for secure image Mathematical and Computer Modelling. 2013 Jul
authentication with localization," IEEE Transactions on 31;58(1):204-18.
Image Processing. 2002 Jun;11(6):585-95. [17] M. Ali, C. W. Ahn, and P. Siarry, "Differential evolution
[3] N. Yadav, K. Singh, "An efficient robust watermarking algorithm for the selection of optimal scaling factors in
scheme for varying sized blocks," Turkish Journal of image watermarking," Engineering Applications of
Electrical Engineering & Computer Sciences. 2016 Jul Artificial Intelligence. 2014 May 31;31:15-26.
1;24(4). [18] R. H. Laskar, M. Choudhury, K. Chakraborty, and S.
[4] N. Muhammad, and N. Bibi, "Digital image Chakraborty, "A joint DWT-DCT based robust digital
watermarking using partial pivoting lower and upper watermarking algorithm for ownership verification of
triangular decomposition into the wavelet domain," IET digital images," In Computer networks and intelligent
Image Processing. 2015 Jun 17;9(9):795-803. computing 2011 (pp. 482-491). Springer Berlin
[5] M. Barni, F. Bartolini, V. Cappellini, and A. Piva, "A Heidelberg.
DCT-domain system for robust image watermarking. [19] A. Hore, and D. Ziou, "Image quality metrics: PSNR vs.
Signal processing," 1998 May 28;66(3):357-72. SSIM," InPattern recognition (icpr), 2010 20th
[6] S. D. Lin, and C. F. Chen, "A robust DCT-based international conference on 2010 Aug 23 (pp. 2366-
watermarking for copyright protection," IEEE 2369). IEEE.
Transactions on Consumer Electronics. 2000 Aug [20] Z. Wang, A. C. Bovik, H. R. Sheikh, E. P. Simoncelli,
1;46(3):415-21. "Image quality assessment: from error visibility to
[7] V. Solachidis, and L. Pitas, "Circularly symmetric structural similarity," IEEE transactions on image
watermark embedding in 2-D DFT domain," IEEE processing. 2004 Apr; 13(4):600-12.
transactions on image processing. 2001
Nov;10(11):1741-53. AUTHORS PROFILE
[8] A. Poljicak, L. Mandic, and D. Agic, "Discrete Fourier
Mohiul Islam has received his M.Tech
transform–based watermarking method with an optimal degree in 2014 from National Institute
implementation radius," Journal of Electronic Imaging. Technology Agartala and B.E degree from
2011 Jul 1;20(3):033008. Assam Engineering College in 2011.He is
currently pursuing Ph.D. in the Department
[9] R. Liu, and T. Tan, "An SVD-based watermarking of Electronics and Communication
scheme for protecting rightful ownership," IEEE Engineering at National Institute of
transactions on multimedia. 2002 Mar;4(1):121-8. Technology, Silchar. His research interests
[10] C. C. Lai, "A digital watermarking scheme based on include Image processing, Machine
Learning, Digital Signal Processing.
singular value decomposition and tiny genetic
E-mail: mohiul292@gmail.com
algorithm," Digital Signal Processing. 2011 Jul
31;21(4):522-7.
[11] C. C. Lai, and C. C. Tsai, "Digital image watermarking
using discrete wavelet transform and singular value
decomposition," IEEE Transactions on instrumentation
and measurement. 2010 Nov;59(11):3060-3.
[12] J. Guo, and P. Zheng, and J. Huang, "Secure
watermarking scheme against watermark attacks in the
encrypted domain," Journal of Visual Communication
and Image Representation. 2015 Jul 31;30:125-35.
Abstract—The growth of real time image based applications in density level of AWGN over image. But the computational cost
Wireless Sensor Network (WSN) leads to the necessity of learning to implement the image denoising algorithm is yet to be the
the behavior of noises. The quality of the image is mainly problem. In particular, the transform based techniques [5]-[8]
determined by the amount of noise density present in the image. are succeeded to reduce the noise level with better image
Hence, in order to suppress the impact of noise level over image, quality at the cost of high energy consumption. Recent years,
image denoising is utilized. In this context, the implementation of spatial based techniques [9] - [15] are also introduced in the
the proposed denoising algorithm is done in two phases. The first field of denoising techniques to reduce the computation cost.
phase exploits the image denoising by applying bilateral filtering However the complexity has reduced to a certain extent only.
and the second phase utilizes the Histogram based Multilevel
Thresholding (HMT) with optimization technique. The HMT is In the meantime, the researchers are focused on meta-heuristic
mainly used to find the optimal thresholds and enhance the image optimization algorithms and incorporated them in many image
features and edges with fine tuning. Further, to reduce the processing applications in order to reduce the searching period
computation time, the harmony search algorithm based
as well to find the optimal solution [16][17]. Hence, Jonates
optimization technique is appended with the proposed algorithm
and tested for various noisy images with different entropies by
Lopes De Paiva and et.al [18] has made an attempt to
using the MATLAB simulation. hybridize the one of the optimization algorithms called genetic
algorithm with three different denoising methods such as
Keywords-Denoising; Entropy; Histogram; Multi-thresholding; Block Matching and 3D filtering (BM3D) [19], wiener-chop
Optimization [20] and anisotropic diffusion [21] to improve the denoising
quality with less complexity. Apart from that, Abhijit Chandra
I. INTRODUCTION and Sudipta Chattopadhyay [22] designed a better low pass
Recently, many applications including object tracking, filter for denoising along with the incorporation of the
forest monitoring, tele-medicine and traffic control and so on, Differential Evolution (DE) based optimization method to
choose WSN as a basic networking technology for evaluate the fittest optimal filter coefficients. Later A.
communication. The main advantage of using WSN is that it K.Bhandari and et.al [23] used the wavelet based adaptive
can be deployed anywhere without infrastructure and less thresholding technique with the differential evolution
computation cost. Despite the advantages of WSN, it has algorithm to eradicate the previous research problems like
limited power, less storage space and low communication edge smoothening and poor image quality. Though the above
range. But, when it comes to the image based applications, the methods show very good results in terms of image quality and
major factor affecting the image during sensing, transmitting
performance metrics such as Peak Signal to Noise Ratio
and receiving is noise. From the case study [1]-[4], it is
(PSNR), Image Quality Index (IQI), Mean Squared Error
understood that, there are several types of noises present during
transmission and reception of images using WSN. The noises (MSE) and Normalized Absolute Error (NAE), the time
are categorized in terms of their density level and required to compute the algorithm is large and this made the
characteristics with respect to the atmospheric condition and algorithm to not applicable for low power wireless networks
type of applications. The practical noises frequently raise like wireless sensor networks.
damage to the images are Additive White Gaussion Noise Hence, it is necessary to propose an efficient denoising
(AWGN), salt and pepper noise, shot noise and so on. Out of algorithm with less computation time for efficient AWGN
these kind of practical noises, the impact of the gaussion noise noise reduction. This paper concentrates on bilateral filtering
during image transmission produce severe problems than that based denoising technique along with histogram based multi-
of other practical noises. Until now, there are lots of researches thresholding instead of using the wavelet based method. In
prevailing in image denoising techniques which are trying to addition, a better optimization mechanism called harmony
find the best solution for suppressing the impact of noise search algorithm is appended with the proposed denoising
method to reduce the computational complexity without details. Hence to improve the quality of the image by
compromising the other performance metrics. preserving the edges without compromising the other metrics,
The rest of the paper is organized as follows: section II the HMT [27] with Harmony Search Algorithm (HSA) based
gives the brief description about the existing techniques. optimization algorithm is proposed in this paper. The working
Section III explains the concept of the proposed denoising principle of the proposed model can be explained by using the
method which includes the detailed introduction about the Fig.1.
bilateral filtering and histogram based multi thresholding with Generally the difference between the input image and its noise
optimization technique. Section IV deals with the discussion free output image shows the impact of the denoising method
about the comparison of proposed algorithm for different over noisy images. Bilateral filter is one of the nonlinear
optimization techniques with the existing in terms of their weighted averaging filters that utilize non-iterative technique
performance metrics. Finally section V, concludes the proposed
to reduce the noise density of the image by considering both
work based upon the result analysis and the future work is also
the spatial and the intensity distance of the neighboring pixels.
given in Section V.
Here each pixel value is replaced by its weighted average.
II. EXISITING TECHNIQUES USING OPTIMIZATION Mathematically, the outcome of the bilateral filter [9] at a
pixel location P is denoted as IF which is expressed by,
Denoising is one of the efficient ways to suppress the noise
over image. The noise model taken into consideration for most
of the denoising techniques is AWGN. For the past three = ∑ ∈ ‖ − ‖ | − | (1)
decades, there are different types of filters used in image
denoising like bilateral filtering, FIR filtering and stochastic Where,
filtering and so on to reduce the noise variance (σ). But Jing −Geometric closeness function
Tian [24] proposed an algorithm that, the wavelet based
−Gray level similarity function
algorithms are more effective to tackle the image denoising
problem than that of other filters. The selection of the w - Normalization constant
thresholds from the wavelet coefficients offers the noise free S - Spatial neighborhood of P
image. However it leads into an intra-scale dependency to ‖ − ‖ −Equalidean distance between P and q
estimate the value of the signal variance using neighboring
coefficients. So, to address the above mentioned problem, the Then the difference between the input image and the bilateral
homogeneous local neighboring coefficients are taken into output image is fed into the histogram based multi
account to solve the problem instead of using the whole thresholding to find the optimal thresholds. This can be
wavelet coefficients. This selection process is achieved by the described using the following steps,
Ant Colony Optimization (ACO) method [25]. The
performance of the ACO based denoising illustrated very good M=I−IF (2)
results in terms of PSNR, MSE and image quality. However
the running time of the ACO based denoising is more where, I is the original image and IF is the output of denoising
compared to that of existing wavelet based denoising operator called bilateral filter. The major application of the
techniques. Also the use of genetic algorithm in image bilateral filtering technique on the noisy image is to average
denoising [26] suffers in terms of computation time and
the noise level along with the image fine details to preserve
average fitness value when the population size varied from
minimum to maximum. the edges and ignores the outliers. In case of applying the
wavelet thresholding after the bilateral filter, the
On the other hand, Jonatas and et.al.,[18] examined the computational complexity is more along with the blurred
property of the denoising methods using different evolutionary texture which is not applicable for WSN. As a matter of fact,
algorithms and developed hybrid of genetic algorithm and the difference image M is partitioned by using histogram and
differential evolution to improve the performance of the the selection of the thresholds from each histogram bin will
existing denoising method such as BM3D, antisotropic determine the quality of the output image. If more number of
diffusion method and Wiener chop method. The computational thresholds (G0-GK) is chosen [28] then the sharpness of the
cost of the above method mainly depends upon the number of image is better i.e., the output of the reconstructed image in
iterations utilized to improve the performance of the denoising
enhanced.
method. Hence, from the existing researches, it is understood
that the trade-off between the computation time and the
performance metrics such as PSNR, NAE and IQI is G0= {(x, y) є M | 0≤ f(x, y) ≤ t1-1} (3)
complicated. G1 = {(x, y) є M |t1 ≤ f(x, y) ≤ t2-1} (4)
G2 = {(x, y) є M | t2≤ f(x, y) ≤ t3-1}............ (5)
III. PROPOSED DENOISING TECHNIQUE Gk= {(x, y) є M |tk ≤ f(x, y) ≤ L-1} (6)
In this paper, without using the wavelet thresholding, the
combination of the histogram based multi-thresholding and After separating the image into bins, average the pixels of
bilateral filtering is used to suppress the noise level over the each bin by using the entropy. The expected value of the
input image. Generally the effect of bilateral filter over noisy information H(X) using Shannon entropy[27] is expressed in
images is not perfect at the point of image edges and fine Eq.(7),
= −∑ ∗ log (7) This can be corrected by fixing the parameter values such as
pitch adjusting rate, harmony memory consideration rate and
To improve the performance and to reduce the computational bandwidth. The design and implementation of the HSA is
complexity of the entropy, the optimization algorithm is more efficient compared to that of the existing optimization
incorporated with the Shannon entropy. In this paper, the HSA algorithms like genetic algorithm, firefly algorithm and
is used as optimization to find the best threshold values of particle swarm optimization and also cuckoo search algorithm
each bin. Unlike other optimization methods, the HSA [30][31]. Finally, the output of the HMT (D) and the output of
[28][29] imposes fewer mathematical requirements and needs the bilateral filter (IF) are added to improve the image edge
less memory space with better accuracy. But the low sharpness and to preserve more image details without noise.
convergence rate and problem in local search performance
provide large processing time.
IF
TABLE -4 Parameters of ACO The design of the denoising algorithm to be used for wireless
Parameter value sensor network mainly depends on their computation time.
Step size (L) 15 Hence, the computation time is taken as a primary selection
Pheromone information (α) 1 criterion for selecting better image denoising method. Here,
Heuristic information (β) 2
the proposed HMT based denoising method is analysed with
four optimization methods such as differential evolution,
Pheromone decay coefficient(ψ) 0.3
particle swarm optimization, harmony search algorithm and
ant colony optimization algorithms and compared with the
After the parameter selection of the optimization algorithm, proposed method to check the performance of the proposed
the proposed algorithm and the existing techniques are method using HSA based optimization technique. The
simulated by using MATLAB R2014. The number of computation time of the existing and the proposed denoising
iterations for all algorithms is fixed to 50. For the sake of the method are shown in Table-5. The overall running time of the
performance analysis of the proposed algorithm, two standard proposed algorithm with the combination of Shannon entropy
test images with different sizes are considered. LENA is very less when compared to that of the other optimization
grayscale image is considered as the first image with the size algorithms.
of 256×256 and another is a Boat image of size 512×512. The
TABLE-5 The average computation time of denoising using different
AWGN is added to the input images by distributing the image optimization comparison for LENA image (in seconds)
pixels using the gaussion random variable with zero mean and
variance [2]. The noisy input images and the denoised output Algorithm Run time
images for different gaussian noise density levels (σ) are
Existing algorithms
illustrated in Fig 2. The performance metrics [15-20] used to
ACO 29.98
judge the efficiency of the proposed method are PSNR, IQI,
PSO 143
MSE, NAE and computation time so on. Before the detailed
DE 16.11
analysis, a brief introduction about each metric is given below,
Proposed algorithm:
HSA with kapur 11.12
a) PSNR: PSNR is one of the best metrics mainly used to
HSA with Shannon 10.33
measure the quality of the reconstructed signal and it can be
HSA with fuzzy 11.21
calculated by finding the ratio between the maximum signal
power to noise power. The formula for calculating the PSNR
To improve the performance of the proposed algorithm
is shown in the below equation,
further, three different entropies such as Shannon, fuzzy and
kapur entropy are taken into account for analysis. Out of these,
PSNR=10log (8) the performance of the harmony search algorithm with
Shannon entropy provides less computation time than that of
where MSE- Mean Squared Error between the original and the other methods. From Table-5, it is clearly visible that the
reconstructed image, this can be formulated as, running time of the proposed denoising algorithm using HSA
with Shannon entropy is better than that of the other entropy
combinations. Hence for further analysis, this paper
1
= , − . concentrates only on the combination of bilateral filtering and
× the HMT using HSA with Shannon entropy.
(9)
To understand the overall performance of the proposed
b) NAE: It is a criterion to evaluate the ability of preserving denoising algorithm with optimization technique, each
the information of the original image. The quality of the image algorithm has taken 100 numbers of trials. Also the proposed
can be determined from the NAE value. If it is low, the image optimization technique is tested with various entropies to
quality is good otherwise the image is not suited for further improve the sharpness of the image edges. It is inferred
processing. It is defined as follows: through the Table 6 and Table 7 that the proposed denoising
method with HSA has low NAE and performs well by techniques with less average computation time. Finally, the
eliminating the noise and also restores the image quality overall simulation results prove that the proposed algorithm is
comparatively better than other approaches. efficient and more appropriate for wireless sensor network in
Moreover, the comparison table shown in Table-7 depicts that terms of less running time and better image quality which is
the proposed algorithm achieves approximately more or less shown in Fig.2.
equal values of the PSNR and IQI as that of existing
TABLE 6 Performance comparison of the proposed denoising method with optimizing technique for different entropies (Boat image)
Noise level σ = 10 σ = 20 σ = 30
HSA with HSA with HSA with HSA with HSA with HSA with HSA with HSA with HSA with
Metric/
fuzzy shannon kapur fuzzy shannon kapur fuzzy shannon kapur
Optimization
entropy entropy entropy entropy entropy entropy entropy entropy entropy
PSNR 30.69 30.758 30.625 28.399 28.263 28.694 26.63 26.184 26.189
NAE 0.0197 0.0203 0.0192 0.0260 0.0251 0.0278 0.0355 0.0314 0.0327
IQI 0.9896 0.9896 0.9895 0.9804 0.9806 0.9818 0.9663 0.9661 0.9645
TABLE 7 Performance comparison of the different optimization based denoising techniques with the proposed denoising algorithm
I
m
a Noise level σ = 10 σ = 20 σ = 30 σ = 40
g
e
Metric/
Optimizatio ACO DE HSA ACO DE HSA ACO DE HSA ACO DE HSA
B n
o
a PSNR 31.47 30.70 30.76 28.49 28.78 28.26 26.39 26.21 26.19 24.11 24.36 24.38
t NAE 0.036 0.018 0.020 0.054 0.028 0.025 0.068 0.032 0.0327 0.082 0.038 0.039
IQI 0.992 0.989 0.989 0.980 0.982 0.981 0.966 0.966 0.965 0.943 0.947 0.947
Metric/
Optimizatio ACO DE HS ACO DE HS ACO DE HS ACO DE HS
Le n
n PSNR 32.79 32.49 33.48 29.87 28.27 30.15 27.6 25.62 27.39 25.98 24.75 26.12
a
NAE 0.04 0.017 0.017 0.062 0.026 0.025 0.080 0.032 0.033 0.097 0.046 0.045
IQI 0.995 0.992 0.996 0.988 0.984 0.990 0.98 0.970 0.975 0.970 0.962 0.977
Noise density(σ) 10 20 30 40
Input image-1
Input image-2
AUTHORS PROFILE
Abstract— The relationship between color and music as part of for researchers. It was this thought that made none other than
the complex system consisting of visual and auditory domain has Sir Issac Newton curious enough to propose such a
not yet been systematically investigated. As both color and music correspondence in his book ‘Opticks’ back in eighteenth
had derived and evolved their forms from the nature that we, century. Since then, researchers have attempted to identify
humans, perceive through our senses and as both the forms are systematic links between music and color. Perhaps the most
processed in the same part of the human body, i.e., the brain, it
will not be a wildly invalid assumption that they share a
direct connection comes from the fascinating phenomenon of
similarity in perception. Needless to say that color and music music–color synesthesia. Studies also show that non-
both have strong impact on emotion and feelings & also a few synesthetic people also have music-to-color associations. So
studies have been reported in literature to explore causal why and how this music-color association works in human? It
relationship between color and emotion. This work reports a is strongly suspected that emotion plays a key role in
neuro-cognitive study on response of brain to different color mediating these two stimuli in the brain. That is to say, both
stimuli. Red, Green, Blue: three primary colors utilizing the color and music has similar emotional qualities that inspire
signals of electro-encephalograms and multi-fractal methodology arousal in a similar manner. Why so? The reason is: music and
to access the degree of complexity with the help of quantitative color have been found to instigate emotional arousal time and
parameter. In this study the correlation between emotional
arousal and the effect of audio and visual stimuli has been
again. The strong relation between music and emotion has
studied. This investigation explores the problem from a new been repeatedly reported in various studies. Music has been
perspective. 15 participants were asked to hear 6 different music shown to affect the emotional state across age, culture and
pieces (each of 30 second duration). The type of emotion elicited language boundaries. The mood a song induces is so reliable
by different music pieces were identified by the participants from that music is often used as a mood-inducer in psychological
a given collection of possible emotional responses. Then they are studies. Similarly, association of color with emotions is also
asked to assign a color associating the emotion from a given color reported in previous literatures. Not only this, the complexity
wheel (structured according to Munsell color system). Each color, in bio-signals using color as a visual stimulus has been found
associated with a particular music piece, is a mixture of specific to be higher than that of music. In light of these, this study is
Red, Green and Blue values (RGB triplet) and has a specific HEX
number (hexadecimal representation), which is recorded for each
conducted to find whether a quantitative correlation can be
response. Then, the musical pieces used were further zoomed found between music, color and emotion since both the stimuli
with the help of fractal technique to identify different emotions is connected to emotion in a consistent manner.
related to music in a quantitative measure. Here, to analyze the This study provides new data in regard to association of color
complexity of the sound signal (which are non-stationary and to emotion. Main outcomes may be summarized as follows:
scale varying in nature), we have used Multifractal detrended the weighted average of emotion ‘Joy’ is higher in clip 1
fluctuation analysis (MFDFA), which is capable of determining whereas, the emotion ‘anxiety’ is prevalent in clip 3 and 4.
multifractal scaling behavior of non-stationary time series. Also, participants reported ‘romantic’ emotion in clip 6. The
Hence, with the data collected, we can correlate color, emotion respective color choices made by participants show a
and music quantitatively.
particular trend in these music pieces. The average values of
Keywords-music, color, MFDFA, RGB triplet, Hexadecimal Green and Red are found to be higher in case of emotions- Joy
representation and Anxiety respectively. Again, the clip 6 (romantic) has an
average higher Red value than Green or Blue. But, comparing
I. INTRODUCTION it with the clip corresponding anxiety, we found that even
The correlation between color and music with effect causing though the Red value is high in both, anxiety is far more
emotional arousal has always been an integral part of interest pronounced and easily associated with high Red values. Same
result was found in both clip 1 and 2 (‘joy’ and ‘devotion’), well as socio-cultural background. The choice included both
both of which are associated with high green values, but musician and non musicians as well. The experimental set up
emotion joy has higher association with Green. was established in a laboratory of Jadavpur University in CV
During the next part of the experiment, i.e., analyzing the
Raman Centre for Physics and Music. The audio clips were
sound signals with MFDFA, the above result showed
consistency. Comparing multifractal width of both signals played for 20 seconds via a standard music player in a room of
having high Red values (clip 3+4 and clip 6) it has been found ambient temperature. Each clip was played with a span of gap
that clips corresponding to anxiety have relatively higher of around 2 minutes and the Munsell Color scheme was
complexity than the one corresponding to romantic emotion preceded by a grey landscape before starting each experiment.
(as was seen in case of average Red values for these clips). The time of the experiments is a normal working hour of any
Similarly, higher green value (between joy and devotion) gave weekday. The participants were expected to be mentally
higher multifractal width (indicating greater complexity).
poised, stable while labeling the emotions and choosing the
This research, although is in a very nascent form, attempts
to identify the correlation between color and music having colors so that the emotions can be identified neutrally.
emotion as the mediator. The psychological connection
between color and emotion has been identified using basic
Munsell color scheme (with RGB triplets and hexadecimal B. Nonlinear chaos based assessment of music samples
representation). Previous works used a limited number of color
variations, thus restricting the emotional component
considerably. Using a color wheel and array of color hues and
The collected music clips were analyzed using a robust
saturation, this work, on the other hand, gives freedom to
analysis method called Multifractal Detrended Fluctuation
associate color to emotion. Also, unlike previous works, this
includes the latest chaos based tools in non-linear methodology Analysis (MFDFA) based on chaos and fractals. Nature is
to verify the psychological data with the analysis of the non- essentially example of most of the complex phenomena,
stationary music stimuli, hence making the results more nonlinear in character - recent research has advocated the
reliable and rigorous in form. This study hopes to stir the nonlinearity of manmade complex systems like music. The
prevailing ideas and initiate further research in cross-modal data series has been exhaustibly studied confirming the
participation in the brain. nonlinearity of brain functions. In view of above, we have opt
the most vigorous approach proposed so far. The multifractal
II. EXPERIMENTAL DETAILS spectrum identifies the deviations in fractal structure within
This investigation explores the problem from a new time periods with large and small fluctuations. The
perspective. 15 participants were asked to hear 6 different multifractals are fundamentally more complex and describe
music pieces (each of 30 second duration). The type of time series featured by very irregular dynamics, with sudden
emotion elicited by different music pieces were identified by and intense bursts of high-frequency fluctuations [6]. MFDFA
the participants from a given collection of possible emotional technique has been widely applied in various fields ranging
responses. Then they are asked to assign a color associating from stock market to biomedical fields for prognosis of
the emotion from a given color wheel (structured according to diseases [7] [8].
Munsell color system). Each color, associated with a particular
music piece, is a mixture of specific Red, Green and Blue In [4], another extensive search shows application of different
values (RGB triplet) and has a specific HEX number variants of MFDFA technique in order to investigate various
(hexadecimal representation), which is recorded for each time series. The analysis shows that the calculated singularity
response. Then, the musical pieces used were further zoomed spectra are very sensitive to the order of the detrending
with the help of fractal technique to identify different
emotions related to music in a quantitative measure. Here, to polynomial used within the MFDFA method. The relation
analyze the complexity of the sound signal (which are non- between the width of the multifractal spectrum and the order
stationary and scale varying in nature), we have used of the polynomial used in calculation is evident. Furthermore,
multifractal detrended fluctuation analysis (MFDFA), which is type of this relation itself depends on the kind of analyzed
capable of determining multifractal scaling behavior of non- signal. Such an analysis can give us some extra information
stationary time series. Hence, with the data collected, we can about the correlative structure of the time series being studied.
correlate color, emotion and music quantitatively.
In reference [5], electroencephalography (EEG) was
performed on 10 participants using a simple acoustical stimuli
A. Self Responses i.e. a tanpura drone. Non-linear analysis in the form of
The specimen six pieces of music were standardized with a set Multifractal Detrended Fluctuation Analysis (MFDFA) was
of emotions varying in the range of joy, sorrow, serenity, carried out on the extracted alpha and theta time series data
anger, heroic, romantic, devotion, anxiety and freedom. 15 from the EEG time series to study the variation of their
different people comprising of both male and female people complexity. It was found that in all the frontal electrodes alpha
were selected for the test. They belong to varied educational as as well as theta complexity increases as is evident from the
increase of multifractal spectral width. This study is entirely Table 3: proportion of RGB coefficients of the colors for each audio clip (for
the chosen subject)
new and gives interesting data regarding neural activation of
the alpha and theta brain rhythms while listening to simple
acoustical stimuli. The importance of this study lies in the R 0.103 0.16 0.448 0.73 0.13 0.27
context of emotion quantification using multifractal spectral
G 0.434 0.38 0.283 0.11 0.36 0.36
width as a parameter as well as in the field of cognitive music
therapy. B 0.463 0.46 0.269 0.16 0.5 0.37
The extensive research in this paper reports a neuro-cognitive Figure 1: plot of RGB values in six different audio clips (for the chosen subject)
study on response of brain to different color stimuli. Red, Color RED
Green, Blue: three primary colors utilizing the signals of
electro-encephalograms and multi-fractal methodology to
access the degree of complexity with the help of quantitative
parameter. The main interesting findings are summarized in R
the result and discussion.
300
III. RESULTS AND DISCUSSIONS
200
As explained in the experimental setup, each subject reacted to
the different music stimuli by choosing the proper emotion(s) 100 R
in the given table by putting 1 as mark of reaction. The
choices of their color indexing the appropriate RGB values are 0
associated with every subject and every clip. The following 1 2 3 4 5 6
tables show one such specimen data of one subject out of the Color GREEN
15 dataset. The choice of subject is random, unspecific and
regardless to reaction. The percentage of the RGB constituents
have also been calculated and given below.
G
Table 1: choice of emotions in different audio clips (for the chosen subject)
300
250
CLIP CLIP CLIP CLIP CLIP CLIP
EMOTION 1 2 3 4 5 6 200
150
JOY 1 1 G
100
SORROW 50
ANGER 0
HEROIC 1 1 2 3 4 5 6
SERENITY 1 1
ANXIETY 1 1 300
B
OTHER
200
Table 2: values of RGB corresponding to the color chosen for different audio 100
clips (for the chosen subject) B
0
COLOR CLIP 1 CLIP 2 CLIP 3 CLIP 4 CLIP 5 CLIP 6 1 2 3 4 5 6
R 52 87 220 240 63 185
G 219 205 139 36 171 246
B 234 244 132 53 238 253
The HEX code associated with each chosen color has resulted
in the mentioned RGB values as tabulated in table 2. The
variation of weighted average for clip 2
proportion of their contribution is also identified in Table 3. It
is to e mentioned that the data set reported in the literature is a Series1
specimen data set which is acconpanied with 15such another
data set whose contribution in totality lead to the tablesand
figures following. We know that Red,Green and Blue are the 6.5
three primary colors which can result in any other color by
proportionate ratio of mixing. In the further execution, each 3.5
2 2.5
clip has been marked by the level of their emotion snad 1 0.5 0.5 0
0 0
associated values of RGB for 15 diferrent subjects. The 1 2 3 4 5 6 7 8 9 10
weighted average of different emotions has been found out
from their which has been listed below in Table 4. The series
of figures below that indicate the clip wise variation of the
weighted average as per the leeled emotions. This study
indicate an approach to identify quantified parameters from an variation of weighted average for clip 3
qualitative impression of different emotions (such as Joy or
Anxiety or Serenity). Series1
4
cli 3
p
jo sorr ang her roma devot seren anxi othe 2
y ow er oic ntic ion ity ety rs 1
no 0 0 0 0 0
9. 1 2 3 4 5 6 7 8 9 10
1 0 0 2.33 2.33 0 0.5 0.5 1
33
5
0.
1.83 0 1 5.5 0 0 0
5
6.83 4
83
3. 2 2
6 1.5 0 0.5 6 1 3.5 0 0
5
0 0 0 0 0
1 2 3 4 5 6 7 8 9 10
Figure 2: plot of the weighted averages for different emotions in respective
clips
variation of weighted average for clip 1 variation of weighted average for clip 5
6.83
5 5.5
1.83 1
2.33 2.33 0.83 0 0 0 0
1 0.5 0.5 1 1 2 3 4 5 6 7 8 9 10
0 0 0
1 2 3 4 5 6 7 8 9 10
Series1
G
45
Average value
40
6 6 35
30
3.5 3.5
25
1.5 20
1
0.5
0 0 0 15
1 2 3 4 5 6 7 8 9 10 10
5
0
From the tables and datas recorded for the six different music 1 2 3 4 5 6
clips for sixteen subjects, the average values of RGB can be G 42.5 40.56 23.38 17.44 35.31 34
further investigated clipwise. These datas are tabulated in
Table 5 and their variations are shown in Figure 3. In the
analysis part, we will certainly see that the dominance of any
color (say Red) indicates any specific emotion or a group of
associated/adjoint emotions to be pronounced in certain
specific audio clips. And the complexity or spectral
dimension( multifractal widths) of those audio clips has been B
obtained by nonlinear technique, mfdfa, as mentioned in the
Average value
60
literature.The values of complexities associated with each clip
has been tabulated with associated emotion and dominant 50
color asper their weighted average and average values in the
table below. 40
Table 5: Clip wise average values of RGB
30
clip no R G B 20
1 25.9 42.5 31.5
2 31 40.5625 28.1875
10
3 50.8125 23.375 19.375 0
4 56.25 17.4375 19.8125 1 2 3 4 5 6
5 29.0625 35.3125 34.1875 B 31.5 28.19 19.38 19.81 34.19 56.88
6 34.625 34 56.875
Table 6: Clip wise indentified complexity in relation with emotion and color
Figure 3: clip wise variation of average values of R, G, B
60
50 clip2 0.428 DEVOTION GREEN
40
clip3 1.031 ANXIETY RED
30
20 clip4 0.458 ANXIETY RED
10
0 clip5 0.471 SERENITY GREEN/BLUE
1 2 3 4 5 6
R 25.9 31 50.81 56.25 29.06 34.63 Clip6 0.383 ROMANTIC RED
1) For author/s of only one affiliation (Heading 3): To [15] Clark, D. M., & Teasdale, J. D. (1985). Constraints on the effects of
change the default, adjust the template as follows. mood on memory. Journal of Personality and Social Psychology, 48, 1595-
1608.
a) Selection (Heading 4): Highlight all author and
affiliation lines. [16] Collins, M. (1929). A case of synaesthesia. Journal of General
b) Change number of columns: Select the Columns icon Psychology, 2, 12-27. Collier, G. L. (1996). Affective synesthesia: Extracting
emotion space from simple perceptual stimuli. Motivation and Emotion, 20, 1-
from the MS Word Standard toolbar and then select “1 32.
Column” from the selection palette.
c) Deletion: Delete the author and affiliation lines for the [17] Cutsforth, T. D. (1925). The role of emotion in a synaesthetic subject.
The American Journal of Psychology, 36, 527-543.
second affiliation.
She has published many of her research papers in different peer reviewed
[18] Dalla Bella, S., Peretz, I., Rousseau, L., & Gosselin, N. (2001). A journals and international conferences.
developmental study of the affective value of tempo and mode in music. Souparno Ray is a research fellow in Jadavpur University and has
Cognition, 80, B1-B10. published papers in many conferences and journals.
.
Dipak Ghosh is a professor emeritus in Jadavpur University. HE worked
AUTHORS PROFILE as a professor in the department of Physics. Served as the head of the
department, Dean of Science respectively. Has supervised more than
Chndrima Roy is working as a assitant professor in the department of
thirty PhDs and has published coutless research papers of international
Electronics and Communication engineering in Heritage Institute of
level of contribution in multiple areas.
Technology. She has worked a bit in the research area of Quantitative
Feedback Theory and optiization. Her current research interest includes
analysis of brain functiong and nonlinearity usiing chaos based techniques.
commonly used ones. A detailed survey on head pose Each of the step involved in estimating VFOA is explained in
estimation in computer vision is given in [15]. People have the following sections. The participants in Image 1 are
worked on extracting head pose from close-up images numbered as highlighted in the Fig 4. Person annotated as 3 is
capturing frontal face [1] or distant images which near frontal the person who is speaking. Hence he is the VFOA in this
[16]-[18]. Number of people covered in an image is inversely case.
proportional to the distance of the camera. We get few high
resolution images in close-up images or many low resolution
heads from a distant camera making it difficult to extract the
head orientation. A single top view camera can capture more
participants in a given environment at its highest resolution. In
this paper, it is proposed to estimate the VFOA from an image
captured by a single top-view camera using basic
segmentation algorithm.
Image 1 Image 2
II. ASSUMPTIONS AND DATABASE DEVELOPMENT
The speaker is detected in the images captured from a single
top-view camera. Dataset was not readily available for testing
the algorithm. A set of images were captured for testing the
algorithm. To prove the concept the algorithm is developed so
that it works under certain assumptions. The assumptions and
the dataset development are discussed in the following
sections. Image 3 Image 4
A. Assumptions
• The image is captured with the camera axis perpendicular
to the ground, in normal indoor daylight conditions
• More than two participants in each of the image
• Both male and female participants with black hair and
without any bags
• Heads of two people are clearly separated Image 5 Image 6
• The floor is not dark
• All participants are looking at one participant without
tilting their head
B. Database Generation
A number of images were captured as per the assumptions.
The images were taken in medium lighting conditions. The
scene here is that of a college balcony. The image was clicked
from the top maintaining the camera axis almost perpendicular Image 7 Image 8
to the ground. Most of the assumptions are satisfied in the
image captured. A total of ten images were captured with six
people in each of the image as shown in Fig 2. Five of them
are wearing light colored dress and one was wearing a blue
dress.
Many pixels of the head region in different images are expected. The six head objects are shown in Fig 6(b)
observed to arrive at the intensity range(thresholds) for each of highlighting the centroid of one of the participant.
the color components. RGB values of one pixel corresponding
to each participant is shown in Fig 5. Based on the
X: 315 Y: 72
observations the thresholds are chosen for red(8-44), RGB: 21, 24, 39
Start
Image input
X: 432 Y: 353
Head Segmentation X: 68 Y: 322
RGB: 19, 29, 38
RGB: 12, 22, 34
X: 253 Y: 423
RGB: 12, 22, 32
Compute the Orientation
of Head Segments
Fig. 5. RGB Intensities in Image 1
Speaker
Stop a b
Fig. 6. Segmented Image(a) and Head Detection(b) in Image 1
TABLE 2
Orientation of Head through Centroid - Image 1 COORDINATES OF THE INTERSECTIONS
187 301
195 337
X: 151.4
Y: 121.9 198 352
223 338
251 340
255 319
260 298
285 341
X: 65.4
Y: 307.5
296 296
302 296
304 291
IV. CONCLUSIONS
An algorithm is developed to detect the visual focus of
attention of the participants in a group. The person who is
speaking is normally assumed as the VFOA in this work. The
head segmentation algorithm used is based on color and area
of the head objects. The VFOA detection accuracy achieved is
70%. This is very encouraging enough to try with robust
segmentation algorithms and extend the work for videos in
Fig. 8. Intersection Points for Image 1 real-time.
The orientation varies depending on the shape of the head,
D. Estimating the Visual Focus of Attention hair style and the angle with which the image is captured.
The centroids and intersections are points in a 2D plane. Based Thus, more robust algorithms to segment the head are to be
on the intersections, the speaker can be estimated as follows. developed and tested with a larger dataset under different
Let the centroids and intersections be represented as conditions.
where 1 6 1 11 for Image 1. The distance
from centroid to intersection is computed using
distance formula. The distance between any two points
, , is given by
Abstract— Face recognition system is one of the esteemed passage of time. Automatic face recognition has demanding
research areas in pattern recognition and computer vision as long tasks in pattern recognition (PR) and artificial intelligence (AI)
as its major challenges. A few challenges in recognizing faces are [19].
blur, illumination, and varied expressions. Blur is natural while The following section II is about literature surveys on
taking photographs using cameras, mobile phones, etc. Blur can
be uniform and non-uniform. Usually non-uniform blur happens
illumination and expression in face recognition. In Section III,
in images taken using handheld image devices. Distinguishing or system architecture and functions of each module has been
handling a blurred image in a face recognition system is generally discussed. Section IV is about experiments on different
tough. Under varying lighting conditions, it is challenging to databases and results analysis. In Section V, conclusion and
identify the person correctly. Diversified facial expressions such future work have been summarized.
as happiness, sad, surprise, fear, anger changes or deforms the
faces from normal images. Identifying faces with facial
expressions is also a challenging task, due to the deformation II. ILLUMINATION AND EXPRESSION IN FACE
caused by the facial expressions. To solve these issues, a pre- RECOGNITION
processing step was carried out after which Blur and
Illumination-Robust Face recognition (BIRFR) algorithm was Vageeswaran, Mitra, and Chellappa (2013), provoked by
performed. The test image and training images with facial the problem of remote face recognition, the issue of
expression are transformed to neutral face using Facial identifying blurred and poorly illuminated faces is addressed.
expression removal (FER) operation. Every training image is The set of all images obtained by blurring a given image is a
transformed based on the optimal Transformation Spread convex set given by the convex hull of shifted versions of the
Function (TSF), and illumination coefficients. Local Binary image. Depending on this set-theoretic characterization, a
Pattern (LBP) features extracted from test image and blur-robust face recognition algorithm DRBF is suggested. In
transformed training image is used for classification. this algorithm we can easily incorporate existing knowledge
on the type of blur as constraints. Taking the low-dimensional
Keywords- Blur; Blur and Illumination- Robust Face
Recognition(BIRFR); Facial expression removal (FER); linear subspace model for illumination, the set of all images
Transformation Spread Function (TSF); Local Binary Pattern obtained from a given image is then shown by blurring and
(LBP). changing its illumination conditions is a bi-convex set. Again,
based on this set-theoretic characterization, a blur and
I. INTRODUCTION illumination robust algorithm IRBF is suggested. Combining a
Face recognition is one of the renowned research areas in discriminative learning based approach like SVM would be a
pattern recognition and computer vision considering its very promising direction for future work.
numerous practical uses in the area of biometrics, information Patel, Wu, Biswas, Phillips, and Chellappa (2012)
security, access control and surveillance system. Several other proposed a face recognition algorithm based on dictionary
applications of face recognition are found in areas such as learning methods that are robust to changes in lighting and
content-based image retrieval, video coding, video pose is proposed. This entails using a relighting approach
conferencing, crowd surveillance, and intelligent human– based on a robust albedo estimation. Different experiments on
computer interfaces. popular face recognition datasets have shown that the method
Recognizing a person or friend is an easy task for human is efficient and can perform importantly better than many
beings. One can easily recognize a person from his biometric competitive face recognition algorithms. Learning
characteristics but in computer vision recognizing a person is discriminative dictionaries is that it can tremendously increase
one of the most challenging tasks. Human faces are complex the overall computational complexity which can make the
and have no rigid structure. A person’s face changes with the real-time processing very difficult. Discriminative methods are
sensitive to noise. It is an interesting topic for future work to
develop and investigate the correctness of discriminative efficient for this purpose. Then, the FLMs were generated
dictionary learning algorithm. This algorithm is potent to pose, based on the proposed method, and finally face recognition is
expression and illumination variations. performed by iterative scoring classification.
Tai, Tan, and Brown (2011), together they have introduced Promising results were observed to handle the face pose in
two contributions to motion deblurring. The first is control and non-control (real-world) situations for face
formulation of the motion blur as an amalgamation of the recognition. It was demonstrated that efficiency of the
scene that has undergone a projective motion path. While a proposed method for pose-invariant face recognition was
straight-forward representation, this formulation is not used in improved in comparison to the state of the art approaches.
image deblurring. The advantages of this motion blur model
are that it is kernel-free and offers a compact representation
for spatially varying blur due to projective motion. In addition, III SYSTEM DESIGN
this kernel free formulation is intuitive with regards to the A. Problem Definition
physical phenomena causing the blur. The second contribution
is an extension to the Richardson- Lucy (RL) deblurring The scheduled work systematically addresses face
algorithm to incorporate motion blur model in a correction recognition under non-uniform motion blur and the merged
algorithm. The basic algorithm, as well as details on effects of blur, illumination, and expression. For this, a
incorporating state-of-the-art regularization, has been outlined. preprocessing step for expression removal step was carried out
A fundamental limitation is that the high-frequency details that after which Blur and Illumination-Robust Face Recognition
have been lost during the motion blur process cannot be algorithm. In Blur and Illumination-Robust Face Recognition
retrieved. The algorithm can only recover the “hidden” algorithm, features are extracted using LBP from the face
particulars that remain inside the motion blur images. Another image [2].
limitation is that this approach does not deal with moving or B. Proposed Solution
deformable objects or scene with significant depth variation. A
For a set of train images gcs and a test image, p, the
pre-processing step to isolate moving objects or depth layers
identity of the test image is to be found. The test image may be
out from background is necessary to deal with this limitation.
blurred, illuminated, along with varied expressions. In test
Other limitations include the problem of pixel colour
image, p, expressions are neutralized using Facial Expression
saturations and severe image noise.
removal (FER). The matrix Ac for each training (gallery) face
Ronen Basri and David W. Jacobs (2003) have proposed
is generated. The test image, p can be stated as the convex
that the set of all Lambertian reflectance functions obtained
consolidation of the columns of one of these matrices. For
with arbitrary distant light sources lies close to a 9D linear
recognition task, the optimal TSF and illumination coefficients
subspace. 9D space can be directly computed from a model, as
[Tc, αc,i] for each training image are computed [2]. Using
low-degree polynomial functions of its scaled surface normals.
facial expression removal, expressions are removed for
It gives us a new and effective way of understanding the
transformed image (blurred and illuminated).
effects of Lambertian reflectance as that of a low-pass filter on
lighting. C. System Architecture
Chao-Kuei Hsieh, Shang-Hong Lai, and Yung-Chang The diagram (Figure 1.) shows architecture for blur,
Chen (2009) described a new algorithm for expression illumination and expression invariant face recognition. A
invariant face recognition with one neutral face image per simple pre-processing technique for removing expressions
class in the training dataset has been proposed. The basic idea from test image and transformed image were done to form a
is to combine the advantages of the feature point labeling in reconstructed face image using wavelet transform. Wavelet
the model-based algorithms and the flexibility of the optical transform are now used to handle such variations.
flow computation to estimate the geometric deformation for A. Blur Invariant Face Recognition
expressive face images. If there had been 1, 2 … C face classes, then each class c
The computation time is greatly reduced in our system by contains gc, which denotes the train images of that class. Also,
using a standard neutral face image in the optical flow the Blurred test image, p belongs to any of the C classes. For
computation and warping process. The constrained optical each training face, transformed images have been formed and
flow warping algorithm significantly improves the recognition these images forms the column of the matrix Ac. The test
rate for face recognition from a single expressive face image image’s identity is obtained using reconstruction error rc in (1).
when only one training neutral image for each subject is The identity of p is found with the one having minimum rc [2].
available. rc= minT || p - AcT||22 + β || T ||1, subject to T ≥ 0 (1)
Ali Moeini and Hossein Moeini (2015), a novel approach Even though all the pixels contain equal weights, every
was proposed for real-world face recognition under pose and region in the face does not convey equal amount of
expression variations from only a single frontal image in the information. For this, a weighting matrix, W has been
gallery which was very rapid and real-time. To handle the introduced. This value was introduced to make algorithms
pose in face recognition, the Feature Library Matrix robust to variations like misalignment. For training the
(FLM)+Probabilistic Facial Expression Recognition Generic weights, training and test image folders of the dataset have
Elastic Model (PFER-GEM) method is proposed that is been used. The test image folder was blurred with Gaussian
algorithm was lesser to BIRFR by 36.986%. In BIRFR, the is less to BIRFR by 4.66%. Compared to other algorithms,
recognition rate obtained was 81.986%. BIRFR's recognition rate was higher and was found to be
86.67%.
TABLE II BIRFR AND OTHER ALGORITHMS FOR CROPPED
YALE DATASET D. BIEFR and Other Algorithms
ALGORITHM RECOGNITION Comparisons between BIEFR and other algorithms like
RATE (%) 2DGFD, 2D-DWT (2 Dimension- Discrete Wavelet
Transform), CT-WLD (Contour Transform-Weber Local
BIRFR 81.986 descriptor) were made for Yale face dataset. Whereas, for
JAFFE dataset CS, SRC, FLLEPCA, HO-SVD, and
DFR 42.14
Eigenfaces algorithms were compared with BIEFR.
IRBF 48.57
1.) JAFFE Database
SRC 19.29
CLDA 21 Pradeep Nagesh and et. al., [CS-Compressive Sensing]
viewed the different images of the same subject as an
ensemble of inter-correlated signals and assumed that changes
1.) Complete linear discrimination analysis (CLDA) due to variation in expressions are sparse with respect to the
whole image. They exploited this sparsity using distributed
Linear discrimination analysis (LDA) is focused at two compressive sensing theory, which enabled them to grossly
different kinds of classification problem, with the purpose of represent the training images of a given subject by only two
selecting the appropriate projection lines to find projection feature images: one that captures the integrated (common)
direction which is at maximum differentiated as two types of features of the face, and the other that captures the different
data points. In Although the LDA algorithm attains good expressions in all training samples.
classification effect in image recognition, two difficulties
arise: one is the high dimension vectors operations that grows TABLE IV BIEFR AND OTHER ALGORITHMS FOR
up computational complexity, and the other is that the within- JAFFE DATASET
class scatter matrix is always singular. In order to address ALGORITHM RECOGNITION
these problems, Fisherface [7] algorithm was proposed which RATE (%)
uses Principal Component Analysis (PCA) for dimension BIEFR 82
reduction before LDA. Nevertheless, the down side of PCA is CS [18] 89.94
the accompanied loss of some important identification SRC [5] [18] 90.1
information. Lastly, they applied CLDA algorithm for the FLLEPCA [2] [10] 93.93
second time feature extraction. By doing so, the proposed
HO-SVD [10] [11] 92.96
procedure can extract more features of identification ability
Eigenfaces 86
therefore. CLDA, Fisherface [7] [12].
In [5], they have used CS (Pradeep Nagesh and Baoxin Li
2.) Yale face dataset (2009)) algorithm and obtained 89.94%. They have used SRC
algorithm for expression invariant face recognition and
TABLE III BIRFR AND OTHER ALGORITHMS FOR obtained 90.1% in [5].
YALE DATASET FLLEPCA (Fusion of Locally Linear Embedding and
ALGORITHM RECOGNITION Principal Component Analysis) algorithm showed a result of
RATE (%) 93.93% in [3] [10]. In [10], HO-SVD (Higher Order Single
BIRFR 87.88 Valued Decomposition) showed a recognition rate of 92.96%.
In [11], Eigenfaces a recognition rate of 86% was obtained by
CLDA 85.19 Hua-Chun et. al.
[12]
2.) Yale Face Dataset
Fisherface 84.21
[12]
In [15], they have used, two-dimensional Gabor Fisher
PCA [23] 82 discriminant (2DGFD) and they have got 70.8% recognition
rate. In two-dimensional Discrete Wavelet Transform(2D-
DWT), the recognition obtained was 82.5%. In [10], Contour
In CLDA, the recognition rate obtained was 85.19% which Transform-Weber Local Descriptor (CT-WLD) has been used
was less to BIRFR by 1.23%. In Fisherface, the recognition and recognition rate obtained was 95.23%. Eigenfaces
rate obtained was 84.21% and was lesser to BIRFR by 2.21%. algorithm has been used in [10] and accuracy obtained was
In PCA [23], the recognition rate was found to be 82% which 86%.
TABLE V BIEFR AND OTHER ALGORITHMS FOR 4. Ali Moeini and Hossein Moeini, “Real-World and Rapid Face
Recognition Toward Pose and Expression Variations via Feature
YALE FACE DATASET Library Matrix”, IEEE Transactions On Information Forensics And
Security, vol. 10, no. 5, May 2015.
ALGORITHM RECOGNITION 5. Andrew Wagner, John Wright, Arvind Ganesh, Zihan Zhou, Hossein
RATE (%) Mobahi, and Yi Ma, “Towards a Practical Face Recognition System:
Robust Alignment and Illumination by Sparse Representation”.
BIEFR 67.996 6. AnubhaPearline.S, Hemalatha.M, “Face Recognition Under Varying
2DGFD [15] [10] 70.8 Blur in an Unconstrained Environment”, IJRET: International Journal
2D-DWT [19] [10] 82.5 of Research in Engineering and Technology eISSN: 2179-1103 |
pISSN: 2181-7168, Volume: 05 Issue: 04, Apr-2016.
CT-WLD [10] 95.23 7. Bellhumer. P. N, Hespanha.J, and Kriegman.D, “Eigenfaces vs.
Eigenfaces [10] 86 fisherfaces: Recognition using class specific linear projection”, IEEE
Transactions on Pattern Analysis and Machine Intelligence, Special
V CONCLUSION AND FUTURE WORK Issue on Face Recognition, vol. 19, no. 19, pp.711—712, Jul. 1997.
8. Chao-Kuei Hsieh, Shang-Hong Lai, and Yung-Chang Chen,
A face recognition system for unconstrained environment “Expression-Invariant Face Recognition with Constrained Optical Flow
was developed using BIEFR algorithm. In this algorithm, LBP Warping”, IEEE Transactions On Multimedia, vol. 11, no. 4., June,
features were extracted for the blurred, illuminated, expression 2009.
variated probe image. Every image in the gallery set was 9. A.W. Galli, G.T.Heydt, P.F. Ribeiro, “Exploring the Power of Wavelet
Analysis”, IEEE, Oct. 1996.
transformed using optimal TSF and their LBP features were 10. Hemprasad Y. Patil, Ashwin G. Kothari and Kishor M. Bhurchandi,
extracted. A simple pre-processing step, FER was carried out “Expression Invariant Face Recognition using Contourlet Transform”,
and the reconstructed face images have been used for further Image Processing Theory, Tools and Applications, IEEE, 2014.
processing. LBP features of transformed image and blurred 11. Hua-Chun, and Z. Yu-Jin., “Expression-independent face recognition
based on higher-order singular value decomposition”, International
probe image were compared to find the best match. BIRFR Conference on Machine Learning and Cybernetics, proceedings,
and BIEFR algorithms were implemented and their results Kunming-China, July 2008.
have been discussed in chapter 4 for the three datasets 12. Kong Rui, Zhang Bing, “An Effective New Algorithm for Face
CroppedYale, Yale face database and JAFFE. BIEFR gives Recognition”, International Conference on Computer Science and
Intelligent Communication, (CSIC), pp. no. 390-393, 2015.
good results even when the probe and gallery set have images 13. Michael J. Jones and Paul Viola, “Fast Multi-view Face Detection”,
of various illuminations. It was observed that for BIRFR Mitsubishi Electric Research Laboratories, 2003.
algorithm, when CroppedYale was used, the recognition rate 14. Michael J. Lyons, Julien Budynek, & Shigeru Akamatsu, “Coding
obtained was 81.986% and for Yale face dataset, 87.88%. It Facial Expressions with Gabor Wavelets”, 3rd IEEE International
Conference on Automatic Face and Gesture Recognition, pp. 120-114,
was observed that for BIEFR algorithm, while JAFFE was Apr.,1998.
used, the recognition rate noticed was 82% and for Yale face 15. Mutelo.R.M, W. L.Woo.W.L, and S. S. Dlay, “Discriminant analysis of
dataset, 67.996%. The system works effortlessly and is robust the two-dimensional Gabor features for face recognition”, Computer
to conditions like blur, illumination and expressions. The Vision, IET,2(2): 19-49, 2008.
16. Patel.V.M, Wu.T, Biswas.S, Phillips.P.J, and Chellappa.R,
results were improved when expression was removed. “Dictionary- based face recognition under variable lighting and pose”,
The issues of face recognition were not completely solved IEEE Trans. Inf. Forensics Security, vol. 7, no. 3, pp. 954–965, June,
using BRFR, few problems such as illumination, expression 2012.
variations were combined to form a better face recognition 17. Paul Viola And Michael J. Jones, “Robust Real-Time Face Detection”,
International Journal of Computer Vision , vol. 57, no. 2, pp. 97-154,
system using BIRFR and BIEFR algorithms. Yet, certain 2004.
problems like occlusion, pose variations, images with make-up 18. Pradeep Nagesh and Baoxin Li, Dept. of Computer Science &
were not handled and it still stands as a stumbling block to an Engineering, Arizona State University, Tempe, AZ 85287, USA, “A
efficient FR system. As future work, this FR system can be Compressive Sensing Approach for Expression-Invariant Face
Recognition”, pp. no. 1518-1514, 2009.
extended to other challenges like pose, occlusion, and make- 19. Ronen Basri and David W. Jacobs, “Lambertian Reflectance and Linear
up. Subspaces”, IEEE Transactions On Pattern Analysis And Machine
Intelligence, vol. 14, no. 2, February 2003.
20. Tai.Y.W, Tan.P, and Brown M. S., “Richardson-Lucy deblurring for
scenes under a projective motion path”, IEEE Trans. Pattern Anal.
Mach. Intell., vol. 33, no. 8, pp. 1003–1018, Aug. 2011.
REFERENCES 21. Vageeswaran.P, Mitra.K, and Chellappa.R (), “Blur and illumination
1. A. Abbas, M. I. Khalil, S.Abdel-Hay, H. M. Fahmy, “Expression and robust face recognition via set-theoretic characterization”, IEEE
Illumination Invariant Preprocessing Technique for Face Recognition”, Transactions On Image Processing, vol. 22, no. 4, pp. 962–972, Apr.,
Computer and System Engineering Department, Faculty of 2013.
Engineering, Ain Shams University, Cairo, Egypt, IEEE, pp. 59-64, 22. Viola.P and Jones.M, “Rapid Object Detection using a Boosted
2008. Cascade of Simple Features”, in Proceeding of Computer Vision and
2. Abhijith Punnappurath and Ambasamudram Narayanan Rajagopalan, Pattern Recognition, , pp. 511-518, 2001.
“Face Recognition Across Non-Uniform Motion Blur, Illumination, and 23. Yan.H, Wang.P, Chen.W.D, Liu.J, “Face Recognition Based on Gabor
Pose”, IEEE Transactions On Image Processing, Vol. 13, No. 7, pp. Wavelet Transform and Modular 2DPCA”, College of Computer
2067- 2082, Jul. 2015. Science Chongqing University of Technology Chongqing, China.
3. Abusham. E, Ngo.D, and Teoh. A, “Fusion of Locally Linear
Embedding and Principal Component Analysis for Face Recognition
(FLLEPCA)”, In Third International Conference on Advances in
Pattern Recognition, ICAPR 2005, Bath, UK, August, 2005.
AUTHORS PROFILE
Abstract— Medical image transmission is substantial in tumor. Generally fuzzy models typically use thresholding
information technologies in order to offer healthcare services at techniques to improve the non-enhanced tumors regions.
distant places. The features of the medical images may be In [6], G. Ulutas et al have proposed a secret image sharing
changed deliberately as the communication might yield through with reversible capabilities. In the proposed work, LSB is used
internet. Before considering the patient diagnostic assessments, for embedding the information in original image in which the
the doctor has to validate the reliability of region of interest in
the received image in order to avoid misdiagnosis. A novel
secret information is entrenched in the LSB plane of the
framework has been proposed to offer a secured healthcare original image. Conversely, it is simple to sense the presence
solution which comprises of Segmentation of region of interest of hidden information and also have little embedding capacity.
(ROI), steganography and encryption technique. This framework In [7], author has discussed Pixel value differencing
also ensures robustness for the embedded information in non- steganography in which these information bits are entrenched
region of interest and recovers ROI perfectly for diagnosis. In the into the pixel pair and correspondingly have inferior
proposed method, the medical image is segmented into ROI and embedding capacity with higher security along with additional
NROI regions using Bhattacharya coefficient segmentation computational time. Linjie Guo et al [8] has proposed an
technique. Patient medical information like blood pressure level, uniform embedding approach for efficient JPEG
sugar level, etc., is embedded into the non-region of interest.
Finally the entire image is encrypted using Logistic map
steganography in which the DCT coefficients are almost
encryption technique in order to ensure the security. laplacian distributed with small magnitude. As a result, most
Experimental outcomes prove that the proposed framework of the coefficients are altered due to the (no shrinkage) nsf5
affords robustness in terms of image quality, security and embedding strategy which happens around bin zero. In [9], Ki-
reliability which helps to alleviate misdiagnosis at the Physician Hyun Jung et al have proposed an enhanced embedding
end in a telemedicine application scenario. technique for exploiting modification direction, which
produces a high capacity and good PSNR compared with other
methods relating to the EMD. The idea behind the proposed
Keywords- Bhattacharya coefficient segmentation, EMD data hiding method is that each secret digit in a (2n+1)-ary
Data hiding technique, Logistic Map Encryption notational system can be carried by one cover pixel. By using
one pixel for cover data, the method achieves a capacity
I. INTRODUCTION
double that of the EMD method.
Tele health care services allows broadcast of medical In [10], the author has proposed 3D baker map encryption
information like image, data between patient and the doctors. technique. In this, it requires huge space for key and the
For treating the patients, medical images help doctors to sensitivity of the key is more. In [11], the authors have
determine the appropriate diagnostic measures [1, 2]. Medical proposed an advanced encryption scheme that yields improved
information plays a vital role in providing critical health care security because of the use of multiple chaotic signals.
services [3]. It is also important to check the validity of Additional common feature in chaos-based communication
medical information as it would have got altered in the algorithms is the combination of encryption and
communication media. When this scenario comes into picture, synchronization. Due to the external errors, the sensitivity of
data integrity of the region of interest (ROI) is the prime the synchronization is maximized. In [12], the author has
requirement in medical application for proper diagnosis. proposed a novel two-channel communication scheme using
Segmentation for MRI images plays a substantial role in chaotic systems. It involves all chaotic states and can be
medical practice. Automatic brain tumor segmentation endures preferred to yield robust sensitivity to the encryption error and
a challenging task in addition with computational complexity. consequently guarantee a high level of security.
MRI images are generally inferred by the doctor in terms of The designed framework is organized as follows: Section I
visual perception [4]. Statistical pattern recognition based surveys the introduction and existing techniques in
techniques [5] fails, moderately because huge distortions segmentation, data embedding and encryption. Section II
happen in the intracranial nerves due to the growth of the brain offers the proposed framework. Next in section III,
experimental outcomes and discussion are presented. Finally distinguished on a input image (I), when compared with a
the section IV provides conclusion drawn from the outcomes benchmark image (R) as in Fig.2 Subsequently, determine the
of the work. axis on an image the leftward partially works as the input
image I, and the rightward partially offers as the benchmark
II. PROPOSED FRAMEWORK image.
image (I) or benchmark image (R), and the superscripts B. Modified EMD Embedding Process
symbolize whether the histogram is calculated inside the Steganography technique entrenches the patient medical
region A(s) or inside the region B(s). For instance, PA(s) information into a cover image. In this, the entrenching
signifies the normalized intensity histogram of trial image I in process is achieved using Exploited Modification Direction
A(s). 〈 X, Y〉 denotes the amount of individual element (EMD) Algorithm. It involves each secret digit in a (2z+1) ary
between two trajectories X and Y. The innermost product notational system to be conceded by a cover pixel. By using
between square roots of two normalized histograms is known one pixel for cover data, the method attains a capacity twice
as Bhattacharya coefficient (BC) (shown in Eqn.2) which is a that of the EMD method. For a pixel value, gi is the cover
real number between 0 and 1 that calculates the relationship image data (153) as shown in fig.5 the function value is
among the two histograms. The bhattacharya coefficient value defined a
is 1, when the two standardized histograms are identical eg. a = (bi+ u) mod (2z + 1) (3)
E(S) = 1 - 0.8 = 0.2. Consider if 0.2 value is positive, the As an example, let the new pixel value of bi’ is 152 as shown
tumor is in right lobe while when the bhattacharya coefficient in fig.6 is obtained by using the eqn (4).
value is 0 the histograms are completely different such that the bi’ = bi + u (4)
value is negative, which denotes that the tumor is in left lobe In the extraction method, a secret digit c is calculated using
respectively. eqn (5)
BC ( a , b ) = ∑ a ( s ) + b ( s ) (2) c = bi’ mod (2z + 1) (5)
i
Once the upper histograms match very well, there is the
expectation of high score whereas the lower histograms have
huge mismatch. On the other hand, a low value of E(s) denotes
a low correlation between upper histograms, and a high
correlation between lower histograms. The Fig.3 depicts that
the detection of brain tumor for six different MRI images
using Bhattacharya coefficient based segmentation and fig.4
shows that the result of segmented image and extraction of
ROI in MRI images. For analysis, the data set has been taken Fig.5: An example of embedding method
from oasis-brains.org.
by using Pixels Mapping Tables (PMT).Two approaches are the SSIM value for region growing algorithm attains the value
involved in this algorithm explicitly pixel replacement and of 0.9824, whereas the proposed technique offers a high value
pixel scrambling approach. In pixel replacement approach we of SSIM (average of 0.99) which is clearly indicated in table I.
need to modify the values of the pixels, whereas in pixel Table II. illustrates that, the performance of the proposed
scrambling the pixels position has to modify. In this work, work is better than the conventional embedding algorithms in
pixel replacement approach has been used (i.e.) pixel (i, j) = terms of payload. Two distinct embedding algorithms namely
PMT1 (index) where index = mod (pixel (i, j) + shift, pixel LSB, nsf5 were compared with the proposed work for
value) and shift = C* random (logistic) with two Pixel different payloads, to analyze the performance in terms of the
Mapping Table (PMT) that is created by using logistic map significant metrics employed in Table I. The accomplished
(Key1). Pixel mapping table has been employed by which it results indicate that the proposed work achieves a substantial
contains the pixel values ranging from 0 to 255 in the enhancement in all the parameters consistently. This is owing
shambled order with the size 256 x1. to the fact that enhanced exploiting modification direction
technique associates different pixel clusters of the cover
III. EXPERIMENTAL OUTCOMES AND DISCUSSION image to symbolize more embedding directions with less
The proposed outcomes are accomplished with two sorts of pixel changes than that of the EMD method. By selecting the
imaging modalities like CT and MRI images. Two distinct appropriate combination of pixel groups, the embedding
segmentation algorithms namely region growing, watershed efficiency and the perceptional quality of the stego image is
were compared with proposed approach for five different brain enriched. From the obtained result, it is inferred that when the
tumor images. For all segmentation approaches, the payload is 0.01 bpac, there is a dramatic increase in PSNR
performance metrics such as Mean Square Error (MSE), Peak value from 62.9 dB to 66.07 dB and when the payload is 0.05
Signal-to Noise Ratio (PSNR) and Structural Similarity Index bpac, there is an improvement in PSNR value from 55.81 dB
(SSIM) were computed for comparison. The accomplished to 62.12 dB. Secondly, the MSE value reduces from 0.0548 to
outcomes indicate that the projected work attains a substantial 0.0160 and from 0.402 to 0.0179 when the payload is 0.01
enhancement in all the parameters habitually at the receiver and 0.05 respectively. Finally, the SSIM value for nsf5 attains
side. This is owing to the fact that it is fast, automated and a the value of 0.9856, whereas the proposed framework offers a
perfect segmentation algorithm that evades these sorts of high value of SSIM (average of 0.99) which is emphasized in
shortcomings by localizing a rectangular box, around the brain the table II. From the facts stated, the investigational
tumor in an MRI image. From the Table I. it is concluded that significances proved that the embedding efficiency and the
there is a dramatic rise in PSNR value from 57.794 dB to perceptional quality of the EMD algorithm are enriched.
63.107db, MSE value reduces from 0.0348 to 0.0160, SSIM
value also increases from 0.9856 to 0.99 for MRI images when
compared with conventional segmentation algorithms. Finally,
S.No Payload
LSB nsf5 PROPOSED
(bpac)
PSNR SSIM MSE PSNR SSIM MSE PSNR SSIM MSE
1 0.01 51.2 0.81692 0.0891 62.9 0.9856 0.0548 66.07 0.9965 0.0160
2 0.02 48.2 0.81682 0.0926 59.81 0.9776 0.0431 65.91 0.9937 0.0166
3 0.03 46.4 0.81671 0.0996 58.23 0.9721 0.0322 64.89 0.9921 0.0167
4 0.04 45.2 0.81664 0.1295 56.87 0.9692 0.479 63.89 0.9906 0.0173
5 0.05 44.2 0.81653 0.1958 55.81 0.9683 0.402 62.12 0.9892 0.0179
IV. CONCLUSION
This proposal aims to put forth a secured telemedicine [5] N. Moon, E. Bulitt, K. Leemput and G. Gerig, “ Model-based brain and
tumor segmentation”, Proceedings of ICPR, vol. 1, pp. 528-531,
framework suitable for a health care scenario. In this proposed Quebec, 2002.
work segmentation, steganography and encryption algorithm [6] G. Ulutas, M. Ulutas and V. V. Nabiyev, “Secret image sharing with
are collectively used to conceal patient diagnostics reversible capabilities,” Int. J. Internet Technology and Secured
information in an image. This framework comprises of Transactions, Vol. 4, no. 1, pp.1–11, 2012.
segmentation algorithm in order to detect the tumor pixels in [7] K Suresh Babu, K B Raja, Kiran Kumar K, Manjula Devi T H,
an image. When the scenario of telemedicine applications Venugopal K R, and L M Patnaik, “Authentication of Secret
Information in Image Steganography,” IEEE conference TENCON ,pp.
comes into picture, data integrity of the region of interest 1-6, 2008.
(ROI) is the prime requirement in medical application for [8] Linjie Guo, Jiangqun Ni and Yun Qing Shi,” Uniform Embedding for
proper diagnosis. An unsupervised automatic method for brain Efficient JPEG Steganography”, IEEE transactions on information
tumor segmentation on MRI images is employed to obtain the forensics and security, vol. 9, no. 5, may 2014.
ROI part in order to detect the tampered region within or to [9] Ki-Hyun Jung, Kee-Young Yoo , ”Improved Exploiting Modification
Direction Method by Modulus Operation”, International Journal of
alter its semantics, the use of multimedia hashes turns out to Signal Processing, Image Processing and Pattern Vol. 2, No.1, March,
be an effective solution to offer proof of legitimacy in (ROI). 2009
Next the patient critical information is embedded into the [10] Yaobin mao, guanrong chen, shiguo lian,” A novel fast image
image using EMD algorithm to uphold the quality of the encryption scheme based on 3-d chaotic baker maps”, International
image. During the transmission of data through an unsecured Journal of bifurcation and chaos in applied sciences and engineering,
volume 14, issue 10, October 2002.
channel the privacy of patient data should be defended as per
[11] T. yang, c. w. wu, and l. o. chua, “cryptography based on chaotic
Health Insurance Portability and Accountability Act (HIPAA), systems,” IEEE trans. circuits syst. I, vol. 44, pp. 469–472, may 1997.
under Telemedicine Legal Policy. Finally, an encryption [12] Zhong-ping jiang ,”a note on chaotic secure communication systems”,
technique is employed using logistic map before the actual IEEE transactions on circuits and systems-I: fundamental theory and
transmission. In the last state, the encrypted stego image is applications, vol. 49, no. 1, January 2002.
transmitted over the channel to analysis the robustness of the [13] Vallathan, Balachandran, Jayanthi: “Provisioning of enhanced medical
image. Similarly at the receiver, the image is decrypted and data security and quality for Telemedicine applications”, International
segmented into ROI and NROI. Next, the concealed data is Conference on Recent Trends in Engineering and Material Sciences 17-
extracted from the embedded image and verified for data 19 March 2016.
integrity. The sequence of modules in the proposed framework [14] Jianping Fan, Guihua Zeng , Mathurin Body , Mohand-Said Hacid,”
affords a better image quality by means of PSNR, MSE and Seeded region growing: an extensive and comparative study”, Pattern
Recognition Letters 26 (2005) pp.1139–1156.
SSIM, which helps to alleviate misdiagnosis at the Physician
[15] Andr´e Bleau and L. Joshua Leon,” Watershed-Based Segmentation and
end in telemedicine applications. Region Merging”, Computer Vision and Image Understanding, pp.317–
370, 2000.
[16] Monika Xess, S. Akila Agnes, ”Analysis of Image Segmentation
REFERENCES Methods Based on Performance Evaluation Parameters”, International
Journal of Computational Engineering Research, Vol, 04, Issue, 3.
[1] Das, S., Kundu, M.K.: “Effective management of medical information [17] Jayalaxmi S. Gonal, Vinayadatt V. Kohir ,” Automatic Detection and
through a novel blind watermarking technique”, J. Med. Syst., 2012, 36, Segmentation of Brain Tumors using Binary Morphological Level Sets
(5), pp. 3339–3351. with Bounding Box,” Proceedings of 2013 3rd International Conference
on Computer Engineering and Bioinformatics (ICCEB 2013), November
[2] Li, X.W., Kim, S.T.: ‘Optical 3D watermark based digital image 23-24, 2013 pp. 37 -43.
watermarking for telemedicine’, Opt. Lasers Eng., 2013, 51, (12), pp.
1310–1320.
[3] Deng, X., Chen, Z., Zeng, F., Zhang, Y., Mao, Y.: “Authentication and
recovery of medical diagnostic image using dual reversible digital
watermarking”, J. Nanosci. Nano technol., 2013, 13, (3), pp. 2099–2107.
[4] N. Ray, Russell Greiner, Albert Murtha, “Using Symmetry to Detect
Abnormalies inBrain MRI,” Computer Society of India
Communications, vol.31 no. 19, pp. 7-10, January 2008.
happen to be more promising and efficient wireless Internet Authors Chaminda et al., in their work [3] have enlisted certain
connectivity option for medical applications [3]. challenges involved in 3D video capturing to viewing in remote
devices especially focused towards medical domain. One of the
TABLE II. MEASURED SPEED OF VARIOUS ISPS AT RANDOM DAY challenges being network bandwidth capacity for
AND TIME INTERVALS.
transmitting/streaming good quality or high definition 3D
Sample Set 1 – Sample Set 2 – Sample Set 3 – video. The same work claims current 4G’s bandwidth
Service Provider Mbps Mbps Mbps availability can cater only partially towards 3D video streaming
requirements. To overcome this limitation, the authors of [3]
Down Up Down Up Down Up
link link link link link link suggest to utilize asymmetric 3D stereoscopic encoding
techniques and other related techniques.
ISP – 1 4G 18.16 0.60 16.16 0.54 19.34 0.68
These proposed techniques are good, but are still limited by
ISP – 1 3G 4.38 0.08 3.10 0.29 4.43 1.36
additional video processing algorithm and again limited by 4G
ISP – 1 ADSL 1.88 0.25 2.20 0.42 2.15 0.39 capacity & ubiquitous implementation of 4G networks.
ISP – 2 Fiber Net 0.47 0.53 0.89 0.79 0.76 0.62 To overcome these limitations and to reduce computational
ISP – 3 VDSL 12.28 0.84 14.46 0.95 11.34 0.79 load on battery operated mobile hand held devices, we propose
a novel idea of splitting the streaming process of right and left
Stereoscopic video [4-16] encoding is one of the fundamental channel into at least two available Internet connections say one
technique in 3D video encoding schemes, there are numerous is 4G and the other is WiFi network. By utilizing WiFi network
efforts put in by research in efficiently encoding stereoscopic which is connected to Internet through ADSL links do not have
3D content and transmission over Internet and mobile high upload bandwidth capacity by itself, hence combination of
networks. H.264/MVC is one of the widely accepted encoding both WiFi/ADSL 4G is deemed to give high bandwidth
standard formats for 3D video streaming applications, and very capacity. 3D video processing is a challenging task [7], there
suitable for low latency and high speed networks [17-25]. are variations in video processing and encoding techniques and
each algorithm or standard has its own advantages and
There has been attempts to successful transmit/stream 3D limitations. In general H.264 is one of the major encoding and
Video for medical related applications over 4G networks [3]. decoding industry video standard.
Tele operation of medical assistance by expert doctors is reality
now and advent of 3D transmission & viewing patient’s It is evident that a stereoscopic 3D video consumes twice the
condition has given tremendous boost to this segment. bandwidth when compared to a standard video transmission.
H.264/MVC [23] happens to be the most preferred multi view
coding technique for commercial 3D applications and devices.
An alternate approach of capturing, streaming and viewing of encoder is the key entity which calculates number of RTP
stereoscopic in corporate older H.264 AVC, where independent sessions required based on number of networks available and
of right channels and left are encoded individually. This is upload capacity in each network. TSM communicates to Rx
referred to as Simulcast [7]. Futuristic 3D video HEVC Session Manager (RSM) to open equal number of RTP
techniques will be based on H.264/MVC. receiving sockets. Interleaved CS-DON (I-C) packet based
H.264/MVC encodes both right and left view concurrently, mode [26] of MST in NAL is configured in the proposed
resulting in two interdependent Bit Streams (BS), followed by architecture. This mode enables interleaving and there by
a multiplexer. Which interleaves frames of each channel overcomes the limitation of low latency networks. Another
culminating into a single Transport stream (TS)[7]. Further the advantage of using this mode is to utilize the feature of
TS is packetized by Network Abstraction Layer (NAL) is Cross-Session Decoding Order Number (CS-DON) [26], this
various formats to suit the network requirement [3,21,26]. technique facilitates easy and efficient decoding order of all
Real-time Transport Protocol (RTP)[27-29] running over User RTP session packets.
Datagram Protocol(UDP) is most widely used approach for
streaming audio and video data. In the proposed architecture captured stereoscopic 3D
video is transmitted to Cloud Aggregation Server (CAS)
IV. ARCHITECTURE rather than transmitting to single or multiple viewers (using
multicast technique) directly. This approach gives flexibility
Proposed architecture exploits the possibility of Multi-
of storing captured video for further analysis and more
Session Transmission (MST) [2, 26] defined for H.264/MVC
importantly one cloud based viewing endpoint.
RTP sessions [28]. Transmit (Tx) & Receive (Rx) Session
Manager interfaces with NAL of both H.264/MVC encoder Round Trip Time (RTT) is used to estimate the upload
and decoder respectively to distribute and consolidate RTP speed capacity of each network connection available.
sessions on to multiple networks. Figure 3 depicts the Download speeds is not of matter of concern because almost
proposed approach pictorially. Tx Session Manager (TSM) at
all network types provide download speeds of 2Mbps or minimum of 40% of peak value for 20% or lesser overall
greater. Hence this work does not focus on this aspect. duration. Similarly 3G/4G networks maintained peak speeds
for duration 85% or more and minimum speed almost 90% or
V. SIMULATION AND RESULTS less compared to peak speed for 15% or lesser duration. Based
Comparative analyzes on the effect of network speed and on this study we have empirically developed these weighted
video streaming bit rates is carried out in this simulation. window functions. There are 100 samples and randomly one
Packet buffering trends and requirement of multiple networks
is analyzed along with maximum and minimum latency
encountered. GNU/Octave is used for simulation. 12 minutes
of video transmission is simulated and measurement samples
are taken at each second, thereby making the sample set equal
to 720 measurements. 1000 iterations of simulation were run
before concluding the results.
It is observed from our measurements (listed in Table–2),
that none of the single network’s upload speed is viable to
upload 480p HQ 3D video. In this regard we have considered
two networks one ADSL and other a 3G or 4G network. Two
networks is a reasonable set because majority of mobile
devices can simultaneously connect to WiFi and 3G/4G
networks. Since WiFi connectivity is symmetric and its speed
is relatively higher to any of the Internet connectivity speeds,
its effect on simulation is of less consequence and hence
ignored.
Bit rate of 3072 kbps for 480p HQ stereoscopic video has
been derived from Table–1. It’s implied that, if consolidated
Figure 5. Speed Probability Random Weight Function used for ADSL
upload speed is equal to or greater than 3072 kbps, throughout Network
the video streaming duration, only then the video is transmitted
without any delay or buffering. But in a practical network
scenario this is seldom true, and network speed is always a
dependent variable of various facts including network load,
distance between transmitter and receiver, environmental
condition, and power optimization settings.
TABLE III. SIMULATION TEST RESULTS OF RTP PACKET BUFFERING AND TIME TO EMPTY.
Peak ADSL Peak 3G/4G Mean ADSL Mean 3G/4G Consolidated Time to Empty
Balance in
Test Upload Rate Upload Rate Upload Rate Upload Rate Mean Upload or Latency in
Buffer in kbp
in kbps in kbps in kbps in kbps Rate in kbps seconds
1 1050 2500 950 2132 3082 -7889.5 -3
2 1050 2500 954 2127 3082 -7624 -3
3 1050 2500 958 2122 3080 -6102 -2
4 800 2700 729 2345 3074 -2070 -1
5 420 1360 381 1153 1535 1.10638e+006 720
Where:
PUS = Peak Upload Speed in Kbps
SPRW = Speed Probability Random Weight per second
Various available ISPs could be grouped as a super set of
(1) as;
AISP = {ISP1, ISP2, ISP3, …, ISPn} ….(2)
Where AISP is Available ISP’s set.
Considering above, one can compute number of similar
ISPs required as:
Similar ISPs required = Required 3D Video Upload
Speed in Kbps divided by sum of product of each ISP’s
Peak Upload Speed and Speed Probability Random Weight
Where
R3DV is the required 3D Video Upload Speed in Kbps
is the peak Upload Speed
is the Speed Probability Random Weight
n is the number of ISPs available.
If the resulting value is greater than 1, it means that there
is deficit of upload speed and one or more ISPs have to be
added. If the resulting value is less than 1, then available
ISPs are more than required upload speed.
VI. CONCLUSION
This work proposes a unique solution of routing
stereoscopic 3D video stream encoded using H.264/MVC
Figure 10. Test5 Results algorithm over multiple network paths using multiple RTP
sessions. The need for such a solution is justified through
simulation and practical measurements. During trials it was
observed that upload speeds of at-least 3 ISPs measured
individually were not sufficient enough to transmit high
quality stereoscopic 3D video. Consolidated peak upload
speeds of either individual ISP or multiple ISP’s have to be
at-least 15% or higher than the encoded video bit rate. Hence,
two or more Internet connections are foreseen to successfully
transmit 3D video without any latency or jitter.
REFERENCE
[1] Vishwa Kiran S, Ramesh Prasad, Thriveni J, Venugopal K R and L M
Figure 11. Time to Empty buffer or latency variation over 1000 simulation Patnaik, “Mobile Cloud Computing for Medical Applications”, 11th
iterations. IEEE India Conference on Emerging Trends and Innovation (INDICON
2014), December 2014, Pune, India.
Based on the simulation results we propose the following [2] Martin, Jim, Rahul Amin, Ahmed Eltawil, and Amr Hussien.
equation (3) to compute the number of similar ISP connections "Limitations of 4G wireless systems." In Proceedings of Virginia Tech
to fulfill the upload speed requirements combination of each Wireless Symposium (Blacksburg, VA). 2011.
ISP’s Peak Upload Speed and it’s Speed Probability Random [3] Hewage, Chaminda TER, Maria G. Martini and Nabeel Khan. "3D
Medical Video Transmission over 4G Networks", Proceedings of the 4th
Weight function is considered as a set. International Symposium on Applied Sciences in Biomedical and
Communication Technologies. ACM, 2011
ISP = {PUS, SPRW} ….(1)
[4] Karim, H A, Hewage, C T, Worrall, Kondoz, A.M., "Scalable Multiple
Description Video Coding for Stereoscopic 3D", IEEE Transactions on
Consumer Electronics, Vol. 54, No.2, pp.745-752, 2008
[5] Tseng, Belle L and Dimitris Anastassiou. "Compatible Video Coding of [26] Y K Wang, R Even, T kristensen, Tandberg, R Jesup,"RFC-6148: RTP
Stereoscopic Sequences using MPEG-2's Scalability and Interlaced Payload Format for H.264 Video".
Structure", International Workshop on HDTV, Vol. 94, 1994. [27] Y K Wang, T Schier1, R. Skupin and P Yue, “RTP Payload Format for
[6] Hewage et al, "Comparison of Stereo Video Coding Support in MPEG-4 MVC Video”, Retrived from http://datatracker.ietf.org/drafts/current/
MAC, H. 264/AVC and H. 264/SVC", Proc. of IET Visual Information [28] Teruhiko Suzuki, Miska M Hannuksela, Ying Chen, Shinobu Hattori,
Engineering-VIE07, 2007 “Study Text of Working Draft 5 of ISO/IEC 14496-10:2012/DAM 2
[7] Merkle, H Brust, K Dix, K Muller and Wiegand, "Stereo Video MVC extensions for inclusion of depth maps”, 2012
Compression for Mobile 3D Services", IEEE 3D TV Conference: The [29] Wenger, Y K Wang, Schier, RTP Payload format for Scalable Video
True Vision- Capture, Transmission and Display of 3D Video, 2009 Coding, Internet Engineering Task Force, 2011,
[8] Balasko, Hrvoje. "Comparison of Compression Algorithms for High
Definition and Super High Definition Video Signals", 2009
[9] Vetro A, Tourapis, A M, Muller and Chen, "3D-TV Content Storage and
Transmission", IEEE Transactions on Broadcasting, Vol. 57, No.2, pp.
384-394, 2011.
[10] Merkle, Wang Y, Muller, Smolic and Wiegand, "Video Plus Depth
Compression for Mobile 3D Services", Proc. IEEE 3DTV Conference,
Germany, 2009.
[11] Jin, L., Boev, A., Gotchev, A., & Egiazarian, “3D-DCT based Multi-
Scale Full Reference Quality Metric for Stereoscopic Video”, Technical
Article.
[12] Zund F, Pritch Y, Sorkine-Hornung A, Mangold S, and Gross T,
"Content-Aware Compression using Saliency-Driven Image
Retargeting", IEEE International Conference on Image Processing , pp.
1845-1849, 2013.
[13] Schwarz H, Marpe D and Wiegand, "Overview of the scalable video
coding extension of the H. 264/AVC standard", IEEE Transactions on
circuits and systems for video technology, Vol. 17, No. 9, pp.1103-1120,
2007.
[14] Stockhammer, Thomas, Miska M. Hannuksela, and Thomas Wiegand.
"H. 264/AVC in Wireless Environments", IEEE transactions on circuits
and systems for video technology, Vol. 13, No. 7, pp. 657-673, 2003
[15] Kovacs, Peter Tamas, et al, "Overview of the Applicability of H.
264/MVC for Real-Time Light-Field Applications", 3DTV-Conference:
The True Vision-Capture, Transmission and Display of 3D Video
(3DTV-CON), 2014.
[16] Mamatha R B, Keshaveni N, “Compartive Sutdy of video Compression
Techiques h.264/AVC”, International Journal of Advanced Research in
Comptuer Science and Sofware Engineering, Vol. 4, Issue 11, 2014.
[17] Aman Jassal, "H.265/HEVC Video Transmission over 4G Cellular
Networks", The University of British Columbia, 2016
[18] Wenger, Stephan. "H. 264/avc over ip." IEEE Transactions on Circuits
and Systems for Video Technolog, Vol. 13, No. 7, pp. 645-656, 2003.
[19] Stockhammer, Thomas, Miska M. Hannuksela, and Thomas Wiegand.
"H. 264/AVC in Wireless Environments", ." IEEE Transactions on
Circuits and Systems for Video Technolog, Vol. 13, No. 7, pp. 657-673,
2003
[20] Micallef, Brian W., Carl J. Debono, and Reuben A. Farrugia. "Error
Concealment Techniques for H. 264/MVC Encoded Sequences." IEEE
Proc. of Int. Conf. of Electrotechnical and Computer Science (ERK),
Portoroz, Slovenia, 2010
[21] Kordelas, Athanasios, Tasos Dagiuklas, and Ilias Politis. "On the
performance of H. 264/MVC over lossy IP-based networks." Signal
Processing Conference (EUSIPCO), 2012 Proceedings of the 20th
European. IEEE, 2012.
[22] Micallef, Brian W., and Carl J. Debono. "An Analysis on the Effect of
Transmission Errors in Real-Time H. 264-MVC Bit-Streams", 15th
IEEE Mediterranean Electrotechnical Conference(IEEE MELECON),
2010.
[23] Liu Yanwei et al, "A Novel Rate Control Technique for Multiview
Video Plus Depth based 3D Video Coding", IEEE Transactions on
Broadcasting, Vol. 57, No. 2, pp. 562-571, 2011
[24] Mor Liu Zhao, et al. "Experimental evaluation of H. 264/Multiview
Video Coding over IP Networks", arXiv preprint alg-geom/9705028,
1997.
[25] Seo Kwang-deok, et al, "A practical RTP packetization scheme for SVC
video transport over IP networks," Etri Journal, Vol. 32, No.2, pp. 281-
291, 2010.
ABSTRACT: Speech is the most important bio these modalities, facial expressions and speech are
signal that human being can produce and known to be more effective for expression of
perceive. Speaking is the process of converting emotions. Speech Emotion Recognition (SER) can
find several applications such as call centres
discrete phonemes into continuous acoustic management, commercial products, life-support
signal. Speech is the most natural and desirable systems, virtual guides, customer service, lie
method of human communication as a source of detectors, conference room research, emotional
information to communicate with intentions and speech synthesis, art, entertainment and others. The
emotions. it is a time continuous signal main goal of emotional speech analysis/recognition
containing information about message, speaker is to identify the different basic emotional
states(Primary emotions) and to categorize them
attitude, language accent , dialect and emotion
like positive(non negative) and negative
many psychological studies suggests only 10% emotions.[3]
of human life is completely unemotional the rest
involves emotional. This paper aims to various Emotion speech analysis is carried out at
speech features used for the analysis of speaker various levels like segmental, sub segmental and
utterance of emotional speech of human being supra segmental. The performance analysis of
emotion recognition is mainly depends on the
which are analyzed by using speech analysing
speaker and the phonetic information. Emotional
software’s(PRATT) and MATLAB. The features are broadly classified as spectral features
emotional analysis helps in the study of what and prosodic features. Most emotional speech can
emotions are and how the speech features be attributed to the larger segments known as supra
properties change for different emotional state segmental features or prosodic features. The
of a human being for different languages as prosodic features are rhythmic, intonational and
subject. speaking rate properties in speech. prosodic
features mainly estimated over the uttered speech
Keywords: Emotion, Human Communication, Speech sentence in the form of long term statistics.[5] [6]
Processing, speech features,
Prosody features are resultants of vocal
fold activity and the spectrum peaks which are
I.INTRODUCTION: influenced by the vocal tract. The prosodic features
gives the utterance level characteristics. Prosodic
Speech analysis helps in phonetic features include mean and variance of fundamental
description, linguistic variability, Text To Speech frequency and energy. The micro prosody features
development (TTS),Automatic Speech Recognition of speech are jitter, shimmer and Harmonics to
(ASR), Speaker identification and for speech noise ratio (HNR)and noise to harmonic ratio
pathology and rehabilitation. Emotional speech (NHR).the performance evaluation of emotional
research (analysis/recognition/synthesis) is a speech considered is carried out.
multidisciplinary field of research with large
contributions from psychology, acoustics, Indian languages:
linguistics, medicine and computer science Most of the Indian language (except a few
engineering. such as English and Urdu) share a common
phonetic base, i.e., they share a common set of
During previous years, many researchers speech sounds, and in addition possess a few more
have worked on the recognition of nonverbal sounds individually. This common phonetic base
information, and have especially focused on consists of about 50 phones, including 15 vowels
emotion recognition. Many kinds of physiological and 35 consonants. While all these languages share
characteristics are used to extract emotions, such as a common phonetic base, some of the languages
voice, facial expressions, hand gestures, body like Hindi, Marathi and Nepali also share a
movements, heart beat and blood pressure. Among common script called Devanagari. Languages such
(c)
(c)
(d)
(d)
mean
6000 formant
frequency
5000
(hertz) f1
mean
(a)
4000 formant
frequency
3000 (hertz) f2
mean
2000 formant
frequency
1000 (hertz) f3
mean
formant
0 frequency
(hertz) f4
(a)
(b)
(b)
(d)
(c)
anger 1.3053 946.06 2508.7 3420.34 4848.95 220.21 2.58 118.59 10.528 0.956
fear 1.8592 844.45 2380.8 3650.25 5037.43 200.5 2.45 122.87 10.188 1.021
happy 1.4604 860.19 2501.7 3369.91 4472.48 169.68 3.079 182.3 12.047 1.141
neutral 1.6038 782.18 2435.5 3550.51 4985.1 170.78 2.366 138.97 9.688 0.962
TABLE 2.SAMPLE 2.
TABLE 3 SAMPLE 3.
TABLE 4 SAMPLE 4.
2
1.8
1.6
1.4
1.2 Sample 1 time (seconds)
1 sample 2 time (seconds)
0.8 sample 3 time (seconds)
0.6 sample 4 time (seconds)
0.4
0.2
0
anger fear happy neutral
250
200
150 mean pitch (hz) sample 1
mean pitch (hz) sample 2
100 mean pitch (hz) sample 3
mean pitch (hz) sample 4
50
0
anger fear happy neutral
Fig 5. Graph showing Mean pitch(Hz) variation for the emotions considered for all the four samples.
emotion Pitch Pitch Pitch Pitch Intensity Intensity Transmission Voice quality
mean range variance contour mean range duration
anger high broader large declines increases high longer Breathy/moderate
ly timbre
fear increas high Very inclines increases high less Danger/afraid
es high
happy high high high raising high high faster Pleasure/anxious
Neutral normal normal normal normal normal normal normal Calm & weak
ABSTRACT: The paper presents formant frequency exploring frequency modulated, amplitude modulated and
estimation for English vowels. Auto-Regressive (AR) model time-modulated carriers (example: resonance movements,
and LPC (Linear Predictive Coding) based estimation of harmonics, noise, pitch, power, duration). The objective is to
formant frequency has been carried out here for different convey information on words, speaker identity, accent,
recordings. The authors propose to use this autoregressive speech style and emotion. The entire gamut of information
model and LPC for utterances made by speakers of south India
on different occasions and in various recording conditions.
is basically conveyed in the large of the traditional telephone
Samples of speeches have recorded consuming ice cold water bandwidth of 4 kHz. Speech energy 4 kHz reflects audio
and having an instant check of its effect and again later after 5 quality and sensation. [2]
minutes. Formant frequencies obtained are compared to those
with recorded under normal conditions. Vowel utterances II. Speech Production.
made by speakers have been recorded 20 times, variations in
the results, for different conditions on intra speaker basis. The Speech is means of communication. It has the distinct
authors propose to investigate those and also the variability as feature as a signal that carries a message or information. It is
a parameter for the speech recognition process. The entire known to be an acoustic waveform that carries a message or
process has been done using MATLAB.
information and the acoustic waveform that can carry
Keywords: Auto regressive model, Formant frequency, LPC, temporal information from the speaker to the listener.
Signal processing. Efficiency underlies acoustic transmission and reception of
any speech. But this is applicable only for transmission over
I. Introduction. a short distance. There is a spread of radiated acoustic
Speech is the most accepted and convenient means of energy at frequencies that are used by the vocal tract and
communication which is well known and recognized. The ear. But this gets reduced with intensity rapidly. [3] [13]
narrow concept of speech is that it is just a sequence of
sounds punctuated by abrupt changes happening from one to Even on occasions when the source is able to produce
another or some signals that are ignored and go into oblivion substantial volume of acoustic power, there is a support of
soon after utterance. It is much more than that. It is a unique only a fraction thereof by the medium without any
signal that conveys information of linguistic and non distortion, while the rest of the it gets squandered in air dust
linguistic type. Such information conveyed by speech is for particles molecular disturbance. It is also results in getting
knowledge of multiple levels speech signals typify (typify) over aero molecular viscosity. The ambient acoustic noise
information bearing that come up as a function of a single places a restriction or limit on the sensitivity of the ear.
independent variable like time . Speech is not just an Physiological noses to play same role in and around the ear
information signal. It is something beyond that. It is actually drum. Voluntary, formalized motions of the respiratory and
a complex wave and acoustic output arising as a result of the masticators apparatus have the speech as the acoustic end
speaker’s effort.[1] product. The closed loop has the ability to develop, control,
maintain and correct it. Acoustic feedback of the hearing
Speech analysis is synonymous with feature extraction of mechanisms, and the kine-sthate feedback of the speech
speech. Speech sounds are sensations of air pressure musculature too have a role here. The central neurons
variations produced by exhaled air and later modulated and system organizes and coordinates information from the
shaped by vibration of glottal cords and the resonance of senses which is then used for directing the function as also
vocal tract during the time air is pushed out through the lips for delivering the descend, linguistically dependant, vocal
and the nose. Speech is signal with information galore articulator motion and acoustic speed. [4][7]
Fig 2 shows the simplified model of speech production for frequency of 22,100Hz in MAT LAB and the samples were
providing idea of linear prediction is intimately linked to the normalized to have amplitude normalization by considering
basic speech synthesis model in which the sampled speech English vowels by Indian English speakers (a,e,i,o,u). [1]
signal was shown as modeled as the output of a linear,
slowly time-varying system excited by either quasi-periodic
Speech Pre- Hamming
impulses (during voiced speech), or random noise (during
unvoiced speech). The linear prediction method provides a signal emphasis window
robust reliable and accurate method for estimating the
parameters that characterize the linear, time-varying system.
AR
an all pole system function describes the linear system over
intervals of short duration. A time varying digital filter with MODEL
known steady state system function represents the
composite spectrum effects of radiation, vocal tract and
glottal excitation.[11] Formant Vocal tract
estimation transfer function
E. Forward-Backward (FB) auto regressive model for
Formant Estimation.
Fig 3. Block diagram for default AR model formant
Formant Frequency Estimation Algorithm for the speech estimation
analysis based on Forward-Backward (FB) in auto
regressive model is shown in Figure 3. Figure depicts the
estimation process for formant frequency estimation. By the III. RESULTS & DISUSSION.
model of speech production, it is analyzed that the speech
encounters a spectral tilt of –20dB/decade, hence to match
this, pre-emphasis filter is used to flatten the spectrum and
boost the higher frequencies, followed by Hamming
window, by obtaining the AR Model and Vocal Tract
transfer function. From the frequency spectrum of the Vocal
Tract transfer function, the formants are extracted. [10]
F. Speech Database:
Figure 4 shows the recorded speech sample for the vowel /a/
in all the three different conditions it can be clearly noticed
the shift in the speech transition time in the figure. Similarly
all the other four vowels were recorded and observed the
variation in the transition time for the three different
conditions.
900
800
700
600
500
400
300 normal
200
100 ice cold
(b) 0
five min
ar_fb
ar_fb
ar_fb
ar_fb
ar_fb
lpc
lpc
lpc
lpc
lpc
a e i o u
f1 (Hz)
Fig 5 (a)
2500
2000
1500
1000
normal
500
ice cold
0
ar_fb
ar_fb
ar_fb
ar_fb
ar_fb
lpc
lpc
lpc
lpc
lpc
five min
(c) a e i o u
TABLE 1: Mean Formant Frequency f1 for the all the vowels at three various conditions.
Mean f1 (Hz)
a e i o u
lpc ar_fb lpc ar_fb lpc ar_fb lpc ar_fb lpc ar_fb
normal 646.03636 612.18209 290.24804 291.32943 672.04429 647.53608 485.78492 495.06655 439.85147 426.10873
ice cold 474.5618 460.34875 460.41055 482.99677 854.464 830.62063 483.53912 469.64243 807.65435 760.19553
five min 534.13713 514.05292 356.49856 351.10009 737.86742 687.52753 515.97832 487.47977 772.58747 736.92003
TABLE 2: Mean Formant Frequency f2 for the all the vowels at three various conditions.
Mean f2 (Hz)
Vowel /a/ /e/ /i/ /o/ /u/
lpc ar_fb lpc ar_fb lpc ar_fb lpc ar_fb lpc ar_fb
normal 1620.3607 1517.1281 1760.0784 1702.7908 1775.2124 1657.1006 972.49609 968.4073 1496.825 1436.3321
ice cold 1785.251 1689.3834 2134.8212 2078.1335 1734.2823 1664.4509 965.34183 934.61768 1658.2033 1541.877
five min 1743.4963 1736.4056 1976.5384 1984.8818 1791.2401 1657.8592 1066.8032 1020.8186 1641.6667 1545.9573
3000
4000
2500 3500
2000 3000
2500
1500 2000
normal 1500 normal
1000
ice cold 1000 ice cold
500 500
five min five min
0 0
ar_fb
ar_fb
ar_fb
ar_fb
ar_fb
lpc
lpc
lpc
lpc
lpc
ar_fb
ar_fb
ar_fb
ar_fb
ar_fb
lpc
lpc
lpc
lpc
lpc
a e i o u a e i o u
f3 (Hz) f4 (Hz)
TABLE 3: Mean Formant Frequency f3 for the all the vowels at three various conditions
Mean f3 (Hz)
Vowel /a/ /e/ /i/ /o/ /u/
lpc ar_fb lpc ar_fb lpc ar_fb lpc ar_fb lpc ar_fb
normal 2338.4056 2144.2687 2349.8735 2292.883 2442.0181 2306.1086 1888.4018 1739.9349 2227.0988 2080.3389
ice cold 2213.7282 2097.3323 2602.2587 2512.4837 2253.6379 2138.7426 2355.1313 1632.054 2480.365 2246.105
five min 2335.5912 2265.3495 2475.2071 2382.0561 2472.7629 2271.0133 2374.9275 1664.5661 2386.4126 2178.9476
TABLE 4: Mean Formant Frequency f4 for the all the vowels at three various conditions
Mean f4 (Hz)
Vowel /a/ /e/ /i/ /o/ /u/
lpc ar_fb lpc ar_fb lpc ar_fb lpc ar_fb lpc ar_fb
normal 3348.7498 2970.6265 3074.6989 2878.1543 3169.7349 2961.7258 2690.452 2511.3353 2862.833 2674.8642
ice cold 2970.271 2751.231 3280.729 3137.3099 3179.0703 2952.5652 3376.3344 2566.8026 3360.7236 3028.9773
five min 3145.7905 2993.7702 3164.7015 3042.5215 3276.1846 3037.5383 3459.3933 2694.705 3415.4331 3071.0632
4000
a ar_fb
3500
a lpc
a lpc
3000
e ar_fb
2500 e lpc
e lpc
2000
i ar_fb
1500 i lpc
i lpc
1000 o ar_fb
o lpc
500
o lpc
0 u ar_fb
f1 f2 f3 f4 f1 f2 f3 f4 f1 f2 f3 f4 u lpc
normal ice cold after five minutes
Fig 6.summarized formants representation for the vowel samples and three different conditions
Dr.M.B.Manjunatha., currently
analysis for kannada Phonetics, in ITSI-TEEE,vol -3 Issue-1, pages 16-19
working as Principal at Akshaya
(2015).
Institute of Technology, Tumkuru,
with total teaching experience of
[3]. Shiva Prasad K.M., Anil Kumar C., KodandaRamaiah G N.,
21years. Obtained Doctorate in
M.B.Manjunatha., Speaker based vocal tract shape estimation for kannada
Speech processing from Magadh
vowels, in IEEE pg 1-6 2015, DOI: 10.1109/EESCO.2015.7253942.
University. Served as placement
officer, HOD, Vice principal and
[4]. Shiva Prasad K.M., Anil Kumar C., M.B.Manjunatha.,
also acted as institutional incharge
KodandaRamaiah G N., Various front end tools for digital speech
for NAAC,ISO,AICTE & NBA in various institutions. Successfully
Processing, in IEEE pg 905-911, 2015, Print ISBN: 978-9-3805-4415-1.
guided more than 45 UG projects & 30 PG projects and currently
guiding for 5 research scholar, having 50 natonal publications and 30
[5]. Anil Kumar C., Shiva Prasad K.M., M.B.Manjunatha.,
international publications.
KodandaRamaiah G N., Vocal Tract shape estimation of Vowels & C-V-V-
C for diversified Indian English Speakers, in IEEE pg 1-7 2015, DOI:
10.1109/EESCO.2015.7253941.
[11] Powen Ru, Taishih Chi, and Shihab Shamma, “The synergy between
speech production and perception”, J. Acoust. Soc. Am. , January 2003 J.
Makhoul, “Linear Prediction: A Tutorial Review,” Proc. IEEE, Vol. 63,pp
561-580,1975.
AUTHORS PROFILE
A. Physical Model distances [13], [14]) and few other approaches are
established on specific radiation sources and detection
This is one of the early techniques for enhancing hardware [15], [16]. In this method the image formation
degraded imaged due to bad weather. There are several process is modeled considering polarization effects of light
physics based models for fog removal. We shall review a scattering in haze. Further, this modeling is then used to
few models; beginning with the scattering models the improve the foggy scene while fetching information about
Dichromatic Atmospheric Scattering model [11] and the atmospheric properties and scene structure. This technique
Contrast or Monochrome model [12] which describes the incorporates the datum that even a partial polarization of
colors and contrasts of a scene captured under bad (foggy) the airlight could be made use in post-processing of hazy
weather conditions. According to the dichromatic image to remove scattering effects. The two unknowns of
atmospheric scattering model [11] the color of a scene
an image namely, object radiance (without haze) and the
point in fog or haze captured by a color camera, is given
airlight , are estimated from two images captured at
by a linear combination of direction of airlight( in foggy nearly unknown orientations of the polarizing filter. The
condition) and the direction of the color of the scene polarization filter is oriented parallel to the plane of
point in a unblemished and clear day. incidence to measure the intensity
= ̂ + , ( ) (6)
∥ = /2 + ∥ , where ∥ = ,
= , = 1− , ≡ ( ⊥ − ∥)/
= + 1− (2) From Equations (6) and (7), the airlight of any point can be
From both of the models we can infer that the contrast and estimated as
color of a scene point degrade exponentially with its depth ∥ ⊥
(8)
from the observer. These models are used in the design of = , where ∥ and ⊥ are estimated intensities
inter-active techniques like Dichromatic Color Transfer
and Depth Heuristics for image deweathering using simple and the unpolarized image as
inputs from the user. In Dichromatic Color Transfer a bad Total = ∥ + ⊥ (9)
pixel is replaced by a best matching good pixel using the
input by the user: The attenuation is estimated as
= ι + (3) (10)
=1−
Here, the color of the pixel is replaced by dehazed color ∞
i. One limitation of the color transfer method is that it Hence, an estimate for the scene radiance in absence of
can be used only to restore color images since all colors in atmospheric scattering is given by
the selected degraded region of an image might not have Total − (11)
equivalent colors in the selected good color region. In =
Depth Heuristics technique the input is an approximate −
minimum and maximum distances and interpolate for
points in between. is hence the dehazed image.
= min + ( max − min), C. The Cost Estimation Method (4)
The effects of dissimilar densities of hazy weather The common limitations of polarization method is that it
(moderate, heavy, etc.) can be produced by different values is not suitable for dynamic scenes in which the changes are
of the scattering coefficient. Thus, by changing more rapid than the filter rotation for finding the maximum
continuously, the clear day radiance at each pixel can be and minimum DOP(Degree of Polarization) and needs user
estimates progressively as, interaction. Whereas this method requires only a single
input image and neither requires neither any user
= [ − (1 − )] (5) interactions nor any geometrical information of the input
image. An overview of the method is as follows: For an
B. Polarization Method input image, the atmospheric light L1 is estimated using
Image enhancement methods proposed earlier to this ( ) = ∞ ( ) ( )
+ ∞(1 − ( )
) (12)
method require prior information about the scene (e.g.,
(The first term in the equation is the direct attenuation and 4. The contrast limited histogram of the contextual region is
the second term is the airlight), then from this the light given by the set of rules:
chromaticity is obtained as
if HCR(i) > NCL; HNCR(i) = NCL;
∞c (13)
c = else if HCR(i) + Nacp ≥ NCL; HNCR(i) = NCL;
∞r + ∞g + ∞b
else HCR(i) = HCR(i) + Nacp
The light color of the input image is removed using the 5. The distributed pixel is given by
light chromaticity. Later, the smoothness cost and data cost
for every pixel are computed. Corresponding equations
used are:
=
1 (14)
( )
c( )= ( ) ( ) + ( ) 1
1 E. Dark Channel Prior
where and are color vectors, and The dark channel prior [6] [9] method for enhancement
of foggy image is based on the statistics of haze-free
are both scalar values. outdoor images. The dark pixels of the local regions (not
([ ]x ) (15) having the sky region), have very low intensity in at least
( x | x) = edges
one color (RGB) channel, whereas in a hazy image, the
intensity of these dark pixels in that channel is high and is
where is the data cost, [ ]x is obtained by plugging mainly contributed by the airlight. The dark channel prior
every value of x into Equation 14, is a constant to is given by the following statistical observation:
normalize edges and x is a small patch centered at location
. These data and smoothness costs build up complete ( )= min ( min ( ( ))), (19)
∈ , , ∈ ( )
Markov random fields (MRFs) [20] that is optimized using
the inference methods like graph-cuts or belief where is a color channel of an image and ( ) is a
propagation, and produce the estimated values of the local patch centered at . According to observation, it is
airlight. Based on this estimation, the direct attenuation of seen that apart from the sky region, the intensity of is
the scene with enhanced visibility is computed. This low and inclines to be zero when is a haze-free image.
method improves the image visibility by only enhancing With a haze image represented by Equation 12 and taking
the contrast and not the original colors of input image. the operation in the local patch on this equation, we
have:
D. CLAHE
min ( ( )) = ̃ ( ) min ( ) + (1 − ̃ ( )) (20)
Contrast Limited Adaptive Histogram Equalization ∈ ( ) ∈ ( )
(CLAHE)is a non-model-based algorithm. Pizer et al.[17] and this min operation is performed on three color channels
proposed the Contrast Limited Adaptive Histogram independently. By taking the minimum among these three
Equalization (CLAHE) method. CLAHE limits the noise and considering the fact that is always positive, we
enhancement by establishing a maximum value. The have:
CLAHE technique put on histogram equalization to a
contextual region such that every pixel of original image is min( min ( ( ))/ )=0 (21)
∈ ( )
in the center of the contextual region. In this technique, the
pixels of the clipped region of the original histogram are Using this the transmission ̃ is estimated from Eq.20 and
redistributed to every gray level. The CLAHE method is given by
involves the following steps: ̃ ( ) = 1 − min( min ( ( ))/ ) (22)
∈ ( )
1. Compute average number of pixels
Consequently, accurate estimation of the haze’s
× (16) transmission is obtained by these dark pixels. Therefore,
=
this method is substantially valid and is capable in handling
distant objects even for the heavy haze image. Also, the
2. Calculate the actual clip-limit method doesn’t depend on significant variance on
= × (17) transmission or surface shading in the input image. The
dark channel prior may be invalid whenever the scene
is the minimum multiple of average pixels in each object is characteristically alike to the airlight in a large
gray level of the contextual region. local region and there is no shadow casting on the object.
3. The number of pixels distributed averagely into each F. Wiener Filtering
gray level is The Wiener filter [21] is the significant method for
= ∑ / , where ∑
removal of blur in hazy/foggy images due to linear motion
is the total number of clipped pixels (18) or unfocused optics. Linear motion in a photograph causes
Blurring and results in poor sampling. Every pixel in a
digital image represents the intensity of a single fixed point where the normalized term
in front of the camera. Unluckily, a given pixel is mixture
of intensities from points along the line of the camera’s = (∥ ( ) − ( ) ∥) (∥ − ∥) (24)
motion, whenever the shutter speed is very slow while the ∈
camera is in motion. This is a 2D analogy to
ensures that the filter preserves image energy
( , ) = ( , ). ( , )
where is the original input image to be filtered;
where is the Fourier transform of an “ideal” version of a are the coordinates of the current pixel to be filtered;
given image, and is the blurring function. Ideally one is the window centered in ; is the range kernel for
smoothing differences in intensities. is the spatial kernel
could reverse-engineer a , or estimate, if and are
known. This technique is inverse filtering. The 2-d Fourier for smoothing differences in coordinates. The weight is
transform of H for motion is a series of functions in assigned using the spatial closeness and the intensity
parallel on a line perpendicular to the direction of motion; difference. Let (i,j) be the location of a pixel to be denoised
and the 2-d Fourier transform of for focus blurring is the in an image and let its neighboring pixel be located at
sombrero function. ( , ), then, the weight assigned for the neighboring pixel
at ( , ) to denoise the chosen pixel ( , ) is given by:
However, in the reality, there are two problems with
( ) ( ) ∥ (, ) ( , )∥
reverse-engineer method. First, is not known accurately. ( )
The blurring function for a given situation can be guessed; (, , , )=
whereas a lot of trial and error is required for determination
of a good blurring function. Second, there is failure in Where and are smoothing parameters and ( , ) and
inverse filtering under some situations as the function ( , ) are the intensity of pixels ( , ) and ( , )
goes to 0 at some values of and . Also, noise in real respectively.
pictures becomes amplified such that it destroys all After calculating the weights normalize them.
attempts at reconstruction of . Hence, the preeminent
way to solve the second problem is to use Wiener filtering. ∑ , ( , )∗ ( , , , ) (26)
(, )=
This tool solves an estimate for according to the ∑ , (, , , )
following equation (derived from a least squares method):
where is the denoised intensity of pixel ( , ) . This
( , ) bilateral filter is used in [8] to refine the airlight map
( , ) = | ( , )| ×
| ( , )| × ( , ) + ( , ) estimated as a result of histogram equalization over foggy
image. Once airlight map is obtained, image is restored
where is a constant chosen to optimize the estimate. using Koschmieders law. Later Histogram stretching of
Wiener filters are the most common deblurring technique output image is performed to get final defoggy image.
used because it mathematically returns the best results.
Whereas, in other filtering like median filtering the media H. Color Attenuation Prior
function can be accurately estimated and can be used to Color attenuation prior [10] is a simple yet dominant
recover the clear picture based on dark channel prior, but method for haze removal from a single hazy image. In a
with a distortion due to disposal with the dark channel and Hazy image, whenever there is variation in haze
relatively large depth of image area is easily confused with concentration, the saturation and brightness of a pixel vary
area of the sky. Hence, details of the object might be erased sharply. When the color of the scene declines due to the
and great contrast occurs to the brightness level of image. influence of haze/fog, its saturation decreases sharply and
Therefore, in order to overcome these disadvantages, at the same time there is increase in its brightness resulting
Wiener filtering based on dark colors processing is used in in high value of difference. This difference between the
[7] to defog images with best results. saturation and the brightness is exploited in the estimation
G. Bilateral Filtering of the haze concentration. Subsequently, this increase in
concentration of haze along with the change of the scene
A bilateral filter is a smoothing filter for images. Also it depth leads to an assumption that the depth of the scene is
is non-linear and edge-preserving and noise-reducing. In a positively correlated with the concentration of the haze,
bilateral filter, the intensity value of every pixel in an hence:
image is replaced by the weighted average of intensity
values of its nearby pixels, where the weights are based on ( )∝ ( )∝ ( )− ( ) (27)
Gaussian distribution. Significantly, these weights depend Here represents the scene depth, represents the
on radiometric differences such as range differences (e.g.,
concentration of the haze, represents the brightness of the
color intensity, depth distance, etc.) as well as on Euclidean
scene and represents the saturation. The above statistics
distance of pixels. Hence, sharp edges are preserved by
is viewed as color attenuation prior. A linear model i.e., a
methodically looping through each pixel and correcting
more accurate expression is as follows:
weights to the adjacent pixels consequently. The bilateral
filter is defined as ( ) = + ( ) + ( ) + ( )
( )= ∑ ∈ ( ) (∥ ( ) − ( ) ∥) (∥ − ∥), (23) where is the position within the image, , , are the
unknown linear coefficients, ( ) is a random variable
Dichromatic Color Transfer, Deweathering using Depth a cost function was developed in the framework of Markov
Heuristics and Restoration using Planar Depth segments, random fields, which could be proficiently optimized by
for adding and deletion of effects from a single image. various techniques, such as graph-cuts or belief
Inference: The color and contrast restoration on several propagation. Methodology: Firstly the direct attenuation
images captured under poor weather conditions were for a selected patch is computed using Eq.(12) ,then the
demonstrated effectively. Furthermore, the method data cost is computed using Eq.(14) .Later, the smoothness
demonstrates an example of adding weather effects to cost is computed using Eq.(15) .With these two costs a
given images. Hence, this interactive method for image comprehensive graph in term of MRF is found, later
enhancement can be used as easy-to use plugins in various inference in MRFs with the number of labels is done using
image processing software. The method can also be used to the Graph-cut algorithm. Lastly, the direct attenuation for
add weather effects to images. the entire image from the estimated airlight is computed
using Eq.(12) Inference: This method requires no
3. Sarit Shwartz et al.[3]. The primary part of this paper is geometrical information of the input image, and can be
separation of airlight and blind estimation. Technique used: applied for gray as well as color images. Hence the
Physics based. Methodology: This method makes use of proposed method is dynamic method that requires only a
mathematical tools established in the field of Blind Source single input image. Further, the method could be applied to
Separation (BSS), which is also known as Independent improve visibility in under-water [18] or other muddled
Component Analysis (ICA). Every scene recovery needs a media with same optical model. The proposed method can
subtraction of the airlight. Specifically, this could be be beneficial for applications like outdoor surveillance
accomplished by studying polarization-filtered images. systems, intelligent vehicle systems, remote sensing
Nevertheless, the salvage of degrade image requires systems, graphics editors, etc..
parameters of the airlight as input. Hence, the parameter p
(degree of polarization)is obtained by blind estimation, 5. Zhiyuan Xu et al.(2009)[5] proposed a Contrast Limited
which results in separation of A(Airlight) from D(Direct Adaptive Histogram Equalization (CLAHE)-based method
transmission) is as follows: to enhance the images degraded by fog. Technique used:
CLAHE. Methodology: This method establishes a
Given maximum value to clip the histogram and redistributes the
= , where = is the transmittance of clipped pixels equally to each gray-level. It can limit the
the atmosphere, noise while enhancing the image contrast. In this method,
firstly, the original image is converted from RGB to
Lobject is the object radiance and HSI(Hue Saturation Intensity). Secondly, the intensity
= (1 − ), component of the HSI image is processed by CLAHE.
Finally, the HSI image is converted back to RGB image.
we have the image radiance Inference: This method doesn’t require any prior weather
= + information, so the method can be applied to the image
captured in the real foggy conditions. The experimental
The degree of polarization of the airlight is defined as results show that there is a significant improvement in
image contrast enhancement of fog degraded images. In
− comparison with other methods, this method is more
=
simple and faster. In addition, the effect of this method is
It follows that satisfactory.
(1 − ) (1 + ) 6. K. He et al. (2010)[6] has proposed a simple yet effective
= + , = + image prior which is the dark channel prior to eliminate
2 2 2 2
fog/haze from a given single input image. Technique used:
Inference: The methodology in this paper is for blind Dark Channel Prior. Methodology: The dark channel prior
recovery of the parameter p (degree of polarization) which is a kind of prior which is based on statistics of haze-free
is needed for separating the airlight from the (clear) outdoor images. It is grounded on a significant
measurements; hence contrast of the image is recovered observation that maximum local patches in haze/fog-free
with neither user interaction nor existence of the sky in the images (outdoor) contain few pixels with very low
frame. As a result eliminates the need for user interaction intensities in at least one color channel. This prior could be
and other conditions needed for image dehazing. Further, used with any haze imaging model to estimate directly the
the work could be protracted to establishment of blind thickness of the haze and recuperate a high quality
attenuation estimation. This work can be extended to other haze/fog free image. The dark channel prior for an image J
scattering modalities (e.g., underwater photography). is given by the Eq.19.
4. Robby T. Tan [4] proposed a method that is based on two 7. Yanjuan Shuaiet al.[7] has presented an image haze/fog
elementary observations: first, images with superior removal system by wiener filtering based on dark channel
visibility i.e., clear-day images have more contrast when prior aiming at color distortion problem for some large
compared with images afflicted due to bad weather; white bright area in the image due to use of image haze
second, airlight tends to be smooth whenever its variation removal by dark channel prior. Technique used: Wiener
depends on the distance of objects to the viewer. Technique filtering based on dark channel prior. Methodology: The
used: Cost function. Depending on these two observations, algorithm mainly estimates the median function in the use
of the media filtering method based on the dark channel, to Proposed algorithm can be used as a pre-processing step
make the media function more accurate and combine with for numerous computer vision algorithms which are based
the wiener filtering closer; so that the image restoration on feature information (for example, tracking,
problem becomes an optimization problem, and by segmentation, object detection).
minimizing mean-square error a clearer, fogless image is
obtained finally. Inference: Experimental results show that V. FUTURE SCOPE
the proposed algorithm can make the image more detailed,
the contour smoother and the whole image clearer. In Nearly all the prevailing dehazing algorithms are based
particular, this algorithm can recover the contrast of a huge on the persistent assumption; hence a supple model is
white area fog image. The algorithm recompenses for the greatly preferred. Only limited approaches [19] have been
deficiency of dark channel prior algorithm there by implemented by using haze density analysis. Such
expanding the use of dark channel prior algorithm and also approaches could be applied in video dehazing for driving
condenses the running time of the image algorithm. The safety assistance device. More focus of fog removal
steps involved for dehazing an image using wiener filtering algorithms should be produce information on features for
based on dark channel prior: tracking, object detection, segmentation and recognition.
Methods have been implemented for extracting only
Step 1: Input the fog image ( ) objects of interest at foreground from degraded images,
Step 2: Use dark colors method to estimate the value A of now the implementation can be extended for extracting
the sky brightness background images and their depths with respect to the
Step 3: Use median filtering method to estimate the value t seen. Nevertheless work can be focused on CLAHE
of the media function method along with color attenuation prior to reduce the
Step 4: and are substituted into Eq.(30) to get the image noise in dehazed image while retaining the contrast and
′( ) through dark colors color of the image.
( )− (30)
( )=
max ( ( ), ̂ VI. CONCLUSION
Images taken in bad weather (for example foggy or
Step 5: The image ′( ) is processed by wiener filtering. hazy) usually lose contrast and reliability. This is because
Step 6: Output the image of the fact that light is absorbed and scattered by particles
( ) = ′( ) & water droplets in the atmosphere during the process of
8. Tripathi A.K. et al.[8] proposed an system using bilateral propagation. Besides, even automatic systems such as
filter for the estimation of airlight and scene contrast Visual Attention Modeling, which toughly depend on the
recovery. Technique used: Bilateral filter. By quantitative description of the input images, fail to work normally since
and qualitative analysis of the proposed algorithm it can be input images are degraded. Consequently, the techniques
verified that it performs better than that of prior state of for image haze removal must be improved there by
the art algorithms like Dark Channel Prior. Methodology: benefiting many image understanding and computer vision
Steps of the proposed for removal algorithm: applications for example Image/video retrieval, Remote
Sensing, Image Classification and video analysis &
Step 1: Input the image . recognition. Subsequently concentration of the haze varies
Step 2: Perform pre-processing by Histogram equalization from place to place and it is tough to detect in a hazy
over the foggy image. This results in better estimation of image, thus image dehazing is a thought-provoking and
airlight . stimulating task.
Assist. Prof, Kanwalvir Singh Dhindsa, B.B.S.B.Engg.College, Fatehgarh Sahib (Punjab), India
Dr. Jamal Ahmad Dargham, School of Engineering and Information Technology, Universiti Malaysia Sabah
Mr. Nitin Bhatia, DAV College, India
Dr. Dhavachelvan Ponnurangam, Pondicherry Central University, India
Dr. Mohd Faizal Abdollah, University of Technical Malaysia, Malaysia
Assist. Prof. Sonal Chawla, Panjab University, India
Dr. Abdul Wahid, AKG Engg. College, Ghaziabad, India
Mr. Arash Habibi Lashkari, University of Malaya (UM), Malaysia
Mr. Md. Rajibul Islam, Ibnu Sina Institute, University Technology Malaysia
Professor Dr. Sabu M. Thampi, .B.S Institute of Technology for Women, Kerala University, India
Mr. Noor Muhammed Nayeem, Université Lumière Lyon 2, 69007 Lyon, France
Dr. Himanshu Aggarwal, Department of Computer Engineering, Punjabi University, India
Prof R. Naidoo, Dept of Mathematics/Center for Advanced Computer Modelling, Durban University of Technology,
Durban,South Africa
Prof. Mydhili K Nair, Visweswaraiah Technological University, Bangalore, India
M. Prabu, Adhiyamaan College of Engineering/Anna University, India
Mr. Swakkhar Shatabda, United International University, Bangladesh
Dr. Abdur Rashid Khan, ICIT, Gomal University, Dera Ismail Khan, Pakistan
Mr. H. Abdul Shabeer, I-Nautix Technologies,Chennai, India
Dr. M. Aramudhan, Perunthalaivar Kamarajar Institute of Engineering and Technology, India
Dr. M. P. Thapliyal, Department of Computer Science, HNB Garhwal University (Central University), India
Dr. Shahaboddin Shamshirband, Islamic Azad University, Iran
Mr. Zeashan Hameed Khan, Université de Grenoble, France
Prof. Anil K Ahlawat, Ajay Kumar Garg Engineering College, Ghaziabad, UP Technical University, Lucknow
Mr. Longe Olumide Babatope, University Of Ibadan, Nigeria
Associate Prof. Raman Maini, University College of Engineering, Punjabi University, India
Dr. Maslin Masrom, University Technology Malaysia, Malaysia
Sudipta Chattopadhyay, Jadavpur University, Kolkata, India
Dr. Dang Tuan NGUYEN, University of Information Technology, Vietnam National University - Ho Chi Minh City
Dr. Mary Lourde R., BITS-PILANI Dubai , UAE
Dr. Abdul Aziz, University of Central Punjab, Pakistan
Mr. Karan Singh, Gautam Budtha University, India
Mr. Avinash Pokhriyal, Uttar Pradesh Technical University, Lucknow, India
Associate Prof Dr Zuraini Ismail, University Technology Malaysia, Malaysia
Assistant Prof. Yasser M. Alginahi, Taibah University, Madinah Munawwarrah, KSA
Mr. Dakshina Ranjan Kisku, West Bengal University of Technology, India
Mr. Raman Kumar, Dr B R Ambedkar National Institute of Technology, Jalandhar, Punjab, India
Associate Prof. Samir B. Patel, Institute of Technology, Nirma University, India
Dr. M.Munir Ahamed Rabbani, B. S. Abdur Rahman University, India
Asst. Prof. Koushik Majumder, West Bengal University of Technology, India
Dr. Alex Pappachen James, Queensland Micro-nanotechnology center, Griffith University, Australia
Assistant Prof. S. Hariharan, B.S. Abdur Rahman University, India
Asst Prof. Jasmine. K. S, R.V.College of Engineering, India
Mr Naushad Ali Mamode Khan, Ministry of Education and Human Resources, Mauritius
Prof. Mahesh Goyani, G H Patel Collge of Engg. & Tech, V.V.N, Anand, Gujarat, India
Dr. Mana Mohammed, University of Tlemcen, Algeria
Prof. Jatinder Singh, Universal Institutiion of Engg. & Tech. CHD, India
Mrs. M. Anandhavalli Gauthaman, Sikkim Manipal Institute of Technology, Majitar, East Sikkim
Dr. Bin Guo, Institute Telecom SudParis, France
Mrs. Maleika Mehr Nigar Mohamed Heenaye-Mamode Khan, University of Mauritius
Prof. Pijush Biswas, RCC Institute of Information Technology, India
Mr. V. Bala Dhandayuthapani, Mekelle University, Ethiopia
Dr. Irfan Syamsuddin, State Polytechnic of Ujung Pandang, Indonesia
Mr. Kavi Kumar Khedo, University of Mauritius, Mauritius
Mr. Ravi Chandiran, Zagro Singapore Pte Ltd. Singapore
Mr. Milindkumar V. Sarode, Jawaharlal Darda Institute of Engineering and Technology, India
Dr. Shamimul Qamar, KSJ Institute of Engineering & Technology, India
Dr. C. Arun, Anna University, India
Assist. Prof. M.N.Birje, Basaveshwar Engineering College, India
Prof. Hamid Reza Naji, Department of Computer Enigneering, Shahid Beheshti University, Tehran, Iran
Assist. Prof. Debasis Giri, Department of Computer Science and Engineering, Haldia Institute of Technology
Subhabrata Barman, Haldia Institute of Technology, West Bengal
Mr. M. I. Lali, COMSATS Institute of Information Technology, Islamabad, Pakistan
Dr. Feroz Khan, Central Institute of Medicinal and Aromatic Plants, Lucknow, India
Mr. R. Nagendran, Institute of Technology, Coimbatore, Tamilnadu, India
Mr. Amnach Khawne, King Mongkut’s Institute of Technology Ladkrabang, Ladkrabang, Bangkok, Thailand
Dr. P. Chakrabarti, Sir Padampat Singhania University, Udaipur, India
Mr. Nafiz Imtiaz Bin Hamid, Islamic University of Technology (IUT), Bangladesh.
Shahab-A. Shamshirband, Islamic Azad University, Chalous, Iran
Prof. B. Priestly Shan, Anna Univeristy, Tamilnadu, India
Venkatramreddy Velma, Dept. of Bioinformatics, University of Mississippi Medical Center, Jackson MS USA
Akshi Kumar, Dept. of Computer Engineering, Delhi Technological University, India
Dr. Umesh Kumar Singh, Vikram University, Ujjain, India
Mr. Serguei A. Mokhov, Concordia University, Canada
Mr. Lai Khin Wee, Universiti Teknologi Malaysia, Malaysia
Dr. Awadhesh Kumar Sharma, Madan Mohan Malviya Engineering College, India
Mr. Syed R. Rizvi, Analytical Services & Materials, Inc., USA
Dr. S. Karthik, SNS Collegeof Technology, India
Mr. Syed Qasim Bukhari, CIMET (Universidad de Granada), Spain
Mr. A.D.Potgantwar, Pune University, India
Dr. Himanshu Aggarwal, Punjabi University, India
Mr. Rajesh Ramachandran, Naipunya Institute of Management and Information Technology, India
Dr. K.L. Shunmuganathan, R.M.K Engg College , Kavaraipettai ,Chennai
Dr. Prasant Kumar Pattnaik, KIST, India.
Dr. Ch. Aswani Kumar, VIT University, India
Mr. Ijaz Ali Shoukat, King Saud University, Riyadh KSA
Mr. Arun Kumar, Sir Padam Pat Singhania University, Udaipur, Rajasthan
Mr. Muhammad Imran Khan, Universiti Teknologi PETRONAS, Malaysia
Dr. Natarajan Meghanathan, Jackson State University, Jackson, MS, USA
Mr. Mohd Zaki Bin Mas'ud, Universiti Teknikal Malaysia Melaka (UTeM), Malaysia
Prof. Dr. R. Geetharamani, Dept. of Computer Science and Eng., Rajalakshmi Engineering College, India
Dr. Smita Rajpal, Institute of Technology and Management, Gurgaon, India
Dr. S. Abdul Khader Jilani, University of Tabuk, Tabuk, Saudi Arabia
Mr. Syed Jamal Haider Zaidi, Bahria University, Pakistan
Assist. Prof. Nisheeth Joshi, Apaji Institute, Banasthali University, Rajasthan, India
Associate Prof. Kunwar S. Vaisla, VCT Kumaon Engineering College, India
Prof Anupam Choudhary, Bhilai School Of Engg.,Bhilai (C.G.),India
Mr. Divya Prakash Shrivastava, Al Jabal Al garbi University, Zawya, Libya
Associate Prof. Dr. V. Radha, Avinashilingam Deemed university for women, Coimbatore.
Dr. Kasarapu Ramani, JNT University, Anantapur, India
Dr. Anuraag Awasthi, Jayoti Vidyapeeth Womens University, India
Dr. C G Ravichandran, R V S College of Engineering and Technology, India
Dr. Mohamed A. Deriche, King Fahd University of Petroleum and Minerals, Saudi Arabia
Mr. Abbas Karimi, Universiti Putra Malaysia, Malaysia
Mr. Amit Kumar, Jaypee University of Engg. and Tech., India
Dr. Nikolai Stoianov, Defense Institute, Bulgaria
Assist. Prof. S. Ranichandra, KSR College of Arts and Science, Tiruchencode
Mr. T.K.P. Rajagopal, Diamond Horse International Pvt Ltd, India
Dr. Md. Ekramul Hamid, Rajshahi University, Bangladesh
Mr. Hemanta Kumar Kalita , TATA Consultancy Services (TCS), India
Dr. Messaouda Azzouzi, Ziane Achour University of Djelfa, Algeria
Prof. (Dr.) Juan Jose Martinez Castillo, "Gran Mariscal de Ayacucho" University and Acantelys research Group,
Venezuela
Dr. Jatinderkumar R. Saini, Narmada College of Computer Application, India
Dr. Babak Bashari Rad, University Technology of Malaysia, Malaysia
Dr. Nighat Mir, Effat University, Saudi Arabia
Prof. (Dr.) G.M.Nasira, Sasurie College of Engineering, India
Mr. Varun Mittal, Gemalto Pte Ltd, Singapore
Assist. Prof. Mrs P. Banumathi, Kathir College Of Engineering, Coimbatore
Assist. Prof. Quan Yuan, University of Wisconsin-Stevens Point, US
Dr. Pranam Paul, Narula Institute of Technology, Agarpara, West Bengal, India
Assist. Prof. J. Ramkumar, V.L.B Janakiammal college of Arts & Science, India
Mr. P. Sivakumar, Anna university, Chennai, India
Mr. Md. Humayun Kabir Biswas, King Khalid University, Kingdom of Saudi Arabia
Mr. Mayank Singh, J.P. Institute of Engg & Technology, Meerut, India
HJ. Kamaruzaman Jusoff, Universiti Putra Malaysia
Mr. Nikhil Patrick Lobo, CADES, India
Dr. Amit Wason, Rayat-Bahra Institute of Engineering & Boi-Technology, India
Dr. Rajesh Shrivastava, Govt. Benazir Science & Commerce College, Bhopal, India
Assist. Prof. Vishal Bharti, DCE, Gurgaon
Mrs. Sunita Bansal, Birla Institute of Technology & Science, India
Dr. R. Sudhakar, Dr.Mahalingam college of Engineering and Technology, India
Dr. Amit Kumar Garg, Shri Mata Vaishno Devi University, Katra(J&K), India
Assist. Prof. Raj Gaurang Tiwari, AZAD Institute of Engineering and Technology, India
Mr. Hamed Taherdoost, Tehran, Iran
Mr. Amin Daneshmand Malayeri, YRC, IAU, Malayer Branch, Iran
Mr. Shantanu Pal, University of Calcutta, India
Dr. Terry H. Walcott, E-Promag Consultancy Group, United Kingdom
Dr. Ezekiel U OKIKE, University of Ibadan, Nigeria
Mr. P. Mahalingam, Caledonian College of Engineering, Oman
Dr. Mahmoud M. A. Abd Ellatif, Mansoura University, Egypt
Mr. Souleymane Balla-Arabé, Xi’an University of Electronic Science and Technology, China
Mr. Mahabub Alam, Rajshahi University of Engineering and Technology, Bangladesh
Mr. Sathyapraksh P., S.K.P Engineering College, India
Dr. N. Karthikeyan, SNS College of Engineering, Anna University, India
Dr. Binod Kumar, JSPM's, Jayawant Technical Campus, Pune, India
Assoc. Prof. Dinesh Goyal, Suresh Gyan Vihar University, India
Mr. Md. Abdul Ahad, K L University, India
Mr. Vikas Bajpai, The LNM IIT, India
Dr. Manish Kumar Anand, Salesforce (R & D Analytics), San Francisco, USA
Assist. Prof. Dheeraj Murari, Kumaon Engineering College, India
Assoc. Prof. Dr. A. Muthukumaravel, VELS University, Chennai
Mr. A. Siles Balasingh, St.Joseph University in Tanzania, Tanzania
Mr. Ravindra Daga Badgujar, R C Patel Institute of Technology, India
Dr. Preeti Khanna, SVKM’s NMIMS, School of Business Management, India
Mr. Kumar Dayanand, Cambridge Institute of Technology, India
Dr. Syed Asif Ali, SMI University Karachi, Pakistan
Prof. Pallvi Pandit, Himachal Pradeh University, India
Mr. Ricardo Verschueren, University of Gloucestershire, UK
Assist. Prof. Mamta Juneja, University Institute of Engineering and Technology, Panjab University, India
Assoc. Prof. P. Surendra Varma, NRI Institute of Technology, JNTU Kakinada, India
Assist. Prof. Gaurav Shrivastava, RGPV / SVITS Indore, India
Dr. S. Sumathi, Anna University, India
Assist. Prof. Ankita M. Kapadia, Charotar University of Science and Technology, India
Mr. Deepak Kumar, Indian Institute of Technology (BHU), India
Dr. Dr. Rajan Gupta, GGSIP University, New Delhi, India
Assist. Prof M. Anand Kumar, Karpagam University, Coimbatore, India
Mr. Mr Arshad Mansoor, Pakistan Aeronautical Complex
Mr. Kapil Kumar Gupta, Ansal Institute of Technology and Management, India
Dr. Neeraj Tomer, SINE International Institute of Technology, Jaipur, India
Assist. Prof. Trunal J. Patel, C.G.Patel Institute of Technology, Uka Tarsadia University, Bardoli, Surat
Mr. Sivakumar, Codework solutions, India
Mr. Mohammad Sadegh Mirzaei, PGNR Company, Iran
Dr. Gerard G. Dumancas, Oklahoma Medical Research Foundation, USA
Mr. Varadala Sridhar, Varadhaman College Engineering College, Affiliated To JNTU, Hyderabad
Assist. Prof. Manoj Dhawan, SVITS, Indore
Assoc. Prof. Chitreshh Banerjee, Suresh Gyan Vihar University, Jaipur, India
Dr. S. Santhi, SCSVMV University, India
Mr. Davood Mohammadi Souran, Ministry of Energy of Iran, Iran
Mr. Shamim Ahmed, Bangladesh University of Business and Technology, Bangladesh
Mr. Sandeep Reddivari, Mississippi State University, USA
Assoc. Prof. Ousmane Thiare, Gaston Berger University, Senegal
Dr. Hazra Imran, Athabasca University, Canada
Dr. Setu Kumar Chaturvedi, Technocrats Institute of Technology, Bhopal, India
Mr. Mohd Dilshad Ansari, Jaypee University of Information Technology, India
Ms. Jaspreet Kaur, Distance Education LPU, India
Dr. D. Nagarajan, Salalah College of Technology, Sultanate of Oman
Dr. K.V.N.R.Sai Krishna, S.V.R.M. College, India
Mr. Himanshu Pareek, Center for Development of Advanced Computing (CDAC), India
Mr. Khaldi Amine, Badji Mokhtar University, Algeria
Mr. Mohammad Sadegh Mirzaei, Scientific Applied University, Iran
Assist. Prof. Khyati Chaudhary, Ram-eesh Institute of Engg. & Technology, India
Mr. Sanjay Agal, Pacific College of Engineering Udaipur, India
Mr. Abdul Mateen Ansari, King Khalid University, Saudi Arabia
Dr. H.S. Behera, Veer Surendra Sai University of Technology (VSSUT), India
Dr. Shrikant Tiwari, Shri Shankaracharya Group of Institutions (SSGI), India
Prof. Ganesh B. Regulwar, Shri Shankarprasad Agnihotri College of Engg, India
Prof. Pinnamaneni Bhanu Prasad, Matrix vision GmbH, Germany
Dr. Shrikant Tiwari, Shri Shankaracharya Technical Campus (SSTC), India
Dr. Siddesh G.K., : Dayananada Sagar College of Engineering, Bangalore, India
Dr. Nadir Bouchama, CERIST Research Center, Algeria
Dr. R. Sathishkumar, Sri Venkateswara College of Engineering, India
Assistant Prof (Dr.) Mohamed Moussaoui, Abdelmalek Essaadi University, Morocco
Dr. S. Malathi, Panimalar Engineering College, Chennai, India
Dr. V. Subedha, Panimalar Institute of Technology, Chennai, India
Dr. Prashant Panse, Swami Vivekanand College of Engineering, Indore, India
Dr. Hamza Aldabbas, Al-Balqa’a Applied University, Jordan
Dr. G. Rasitha Banu, Vel's University, Chennai
Dr. V. D. Ambeth Kumar, Panimalar Engineering College, Chennai
Prof. Anuranjan Misra, Bhagwant Institute of Technology, Ghaziabad, India
Ms. U. Sinthuja, PSG college of arts &science, India
Dr. Ehsan Saradar Torshizi, Urmia University, Iran
Dr. Shamneesh Sharma, APG Shimla University, Shimla (H.P.), India
Assistant Prof. A. S. Syed Navaz, Muthayammal College of Arts & Science, India
Assistant Prof. Ranjit Panigrahi, Sikkim Manipal Institute of Technology, Majitar, Sikkim
Dr. Khaled Eskaf, Arab Academy for Science ,Technology & Maritime Transportation, Egypt
Dr. Nishant Gupta, University of Jammu, India
Assistant Prof. Nagarajan Sankaran, Annamalai University, Chidambaram, Tamilnadu, India
Assistant Prof.Tribikram Pradhan, Manipal Institute of Technology, India
Dr. Nasser Lotfi, Eastern Mediterranean University, Northern Cyprus
Dr. R. Manavalan, K S Rangasamy college of Arts and Science, Tamilnadu, India
Assistant Prof. P. Krishna Sankar, K S Rangasamy college of Arts and Science, Tamilnadu, India
Dr. Rahul Malik, Cisco Systems, USA
Dr. S. C. Lingareddy, ALPHA College of Engineering, India
Assistant Prof. Mohammed Shuaib, Interal University, Lucknow, India
Dr. Sachin Yele, Sanghvi Institute of Management & Science, India
Dr. T. Thambidurai, Sun Univercell, Singapore
Prof. Anandkumar Telang, BKIT, India
Assistant Prof. R. Poorvadevi, SCSVMV University, India
Dr Uttam Mande, Gitam University, India
Dr. Poornima Girish Naik, Shahu Institute of Business Education and Research (SIBER), India
Prof. Md. Abu Kausar, Jaipur National University, Jaipur, India
Dr. Mohammed Zuber, AISECT University, India
Prof. Kalum Priyanath Udagepola, King Abdulaziz University, Saudi Arabia
Dr. K. R. Ananth, Velalar College of Engineering and Technology, India
Assistant Prof. Sanjay Sharma, Roorkee Engineering & Management Institute Shamli (U.P), India
Assistant Prof. Panem Charan Arur, Priyadarshini Institute of Technology, India
Dr. Ashwak Mahmood muhsen alabaichi, Karbala University / College of Science, Iraq
Dr. Urmila Shrawankar, G H Raisoni College of Engineering, Nagpur (MS), India
Dr. Krishan Kumar Paliwal, Panipat Institute of Engineering & Technology, India
Dr. Mukesh Negi, Tech Mahindra, India
Dr. Anuj Kumar Singh, Amity University Gurgaon, India
Dr. Babar Shah, Gyeongsang National University, South Korea
Assistant Prof. Jayprakash Upadhyay, SRI-TECH Jabalpur, India
Assistant Prof. Varadala Sridhar, Vidya Jyothi Institute of Technology, India
Assistant Prof. Parameshachari B D, KSIT, Bangalore, India
Assistant Prof. Ankit Garg, Amity University, Haryana, India
Assistant Prof. Rajashe Karappa, SDMCET, Karnataka, India
Assistant Prof. Varun Jasuja, GNIT, India
Assistant Prof. Sonal Honale, Abha Gaikwad Patil College of Engineering Nagpur, India
Dr. Pooja Choudhary, CT Group of Institutions, NIT Jalandhar, India
Dr. Faouzi Hidoussi, UHL Batna, Algeria
Dr. Naseer Ali Husieen, Wasit University, Iraq
Assistant Prof. Vinod Kumar Shukla, Amity University, Dubai
Dr. Ahmed Farouk Metwaly, K L University
Mr. Mohammed Noaman Murad, Cihan University, Iraq
Dr. Suxing Liu, Arkansas State University, USA
Dr. M. Gomathi, Velalar College of Engineering and Technology, India
Assistant Prof. Sumardiono, College PGRI Blitar, Indonesia
Dr. Latika Kharb, Jagan Institute of Management Studies (JIMS), Delhi, India
Associate Prof. S. Raja, Pauls College of Engineering and Technology, Tamilnadu, India
Assistant Prof. Seyed Reza Pakize, Shahid Sani High School, Iran
Dr. Thiyagu Nagaraj, University-INOU, India
Assistant Prof. Noreen Sarai, Harare Institute of Technology, Zimbabwe
Assistant Prof. Gajanand Sharma, Suresh Gyan Vihar University Jaipur, Rajasthan, India
Assistant Prof. Mapari Vikas Prakash, Siddhant COE, Sudumbare, Pune, India
Dr. Devesh Katiyar, Shri Ramswaroop Memorial University, India
Dr. Shenshen Liang, University of California, Santa Cruz, US
Assistant Prof. Mohammad Abu Omar, Limkokwing University of Creative Technology- Malaysia
Mr. Snehasis Banerjee, Tata Consultancy Services, India
Assistant Prof. Kibona Lusekelo, Ruaha Catholic University (RUCU), Tanzania
Assistant Prof. Adib Kabir Chowdhury, University College Technology Sarawak, Malaysia
Dr. Ying Yang, Computer Science Department, Yale University, USA
Dr. Vinay Shukla, Institute Of Technology & Management, India
Dr. Liviu Octavian Mafteiu-Scai, West University of Timisoara, Romania
Assistant Prof. Rana Khudhair Abbas Ahmed, Al-Rafidain University College, Iraq
Assistant Prof. Nitin A. Naik, S.R.T.M. University, India
Dr. Timothy Powers, University of Hertfordshire, UK
Dr. S. Prasath, Bharathiar University, Erode, India
Dr. Ritu Shrivastava, SIRTS Bhopal, India
Prof. Rohit Shrivastava, Mittal Institute of Technology, Bhopal, India
Dr. Gianina Mihai, Dunarea de Jos" University of Galati, Romania
Assistant Prof. Ms. T. Kalai Selvi, Erode Sengunthar Engineering College, India
Assistant Prof. Ms. C. Kavitha, Erode Sengunthar Engineering College, India
Assistant Prof. K. Sinivasamoorthi, Erode Sengunthar Engineering College, India
Assistant Prof. Mallikarjun C Sarsamba Bheemnna Khandre Institute Technology, Bhalki, India
Assistant Prof. Vishwanath Chikaraddi, Veermata Jijabai technological Institute (Central Technological Institute), India
Assistant Prof. Dr. Ikvinderpal Singh, Trai Shatabdi GGS Khalsa College, India
Assistant Prof. Mohammed Noaman Murad, Cihan University, Iraq
Professor Yousef Farhaoui, Moulay Ismail University, Errachidia, Morocco
Dr. Parul Verma, Amity University, India
Professor Yousef Farhaoui, Moulay Ismail University, Errachidia, Morocco
Assistant Prof. Madhavi Dhingra, Amity University, Madhya Pradesh, India
Assistant Prof.. G. Selvavinayagam, SNS College of Technology, Coimbatore, India
Assistant Prof. Madhavi Dhingra, Amity University, MP, India
Professor Kartheesan Log, Anna University, Chennai
Professor Vasudeva Acharya, Shri Madhwa vadiraja Institute of Technology, India
Dr. Asif Iqbal Hajamydeen, Management & Science University, Malaysia
Assistant Prof., Mahendra Singh Meena, Amity University Haryana
Assistant Professor Manjeet Kaur, Amity University Haryana
Dr. Mohamed Abd El-Basset Matwalli, Zagazig University, Egypt
Dr. Ramani Kannan, Universiti Teknologi PETRONAS, Malaysia
Assistant Prof. S. Jagadeesan Subramaniam, Anna University, India
Assistant Prof. Dharmendra Choudhary, Tripura University, India
Assistant Prof. Deepika Vodnala, SR Engineering College, India
Dr. Kai Cong, Intel Corporation & Computer Science Department, Portland State University, USA
Dr. Kailas R Patil, Vishwakarma Institute of Information Technology (VIIT), India
Dr. Omar A. Alzubi, Faculty of IT / Al-Balqa Applied University, Jordan
Assistant Prof. Kareemullah Shaik, Nimra Institute of Science and Technology, India
Assistant Prof. Chirag Modi, NIT Goa
Dr. R. Ramkumar, Nandha Arts And Science College, India
Dr. Priyadharshini Vydhialingam, Harathiar University, India
Dr. P. S. Jagadeesh Kumar, DBIT, Bangalore, Karnataka
Dr. Vikas Thada, AMITY University, Pachgaon
Dr. T. A. Ashok Kumar, Institute of Management, Christ University, Bangalore
Dr. Shaheera Rashwan, Informatics Research Institute
Dr. S. Preetha Gunasekar, Bharathiyar University, India
Asst Professor Sameer Dev Sharma, Uttaranchal University, Dehradun
Dr. Zhihan lv, Chinese Academy of Science, China
Dr. Ikvinderpal Singh, Trai Shatabdi GGS Khalsa College, Amritsar
Dr. Umar Ruhi, University of Ottawa, Canada
Dr. Jasmin Cosic, University of Bihac, Bosnia and Herzegovina
Dr. Homam Reda El-Taj, University of Tabuk, Kingdom of Saudi Arabia
Dr. Mostafa Ghobaei Arani, Islamic Azad University, Iran
Dr. Ayyasamy Ayyanar, Annamalai University, India
Dr. Selvakumar Manickam, Universiti Sains Malaysia, Malaysia
Dr. Murali Krishna Namana, GITAM University, India
Dr. Smriti Agrawal, Chaitanya Bharathi Institute of Technology, Hyderabad, India
Professor Vimalathithan Rathinasabapathy, Karpagam College Of Engineering, India
IJCSIS 2016
ISSN: 1947-5500
http://sites.google.com/site/ijcsis/
International Journal Computer Science and Information Security, IJCSIS, is the premier
scholarly venue in the areas of computer science and security issues. IJCSIS 2011 will provide a high
profile, leading edge platform for researchers and engineers alike to publish state-of-the-art research in the
respective fields of information technology and communication security. The journal will feature a diverse
mixture of publication articles including core and applied computer science related topics.
Authors are solicited to contribute to the special issue by submitting articles that illustrate research results,
projects, surveying works and industrial experiences that describe significant advances in the following
areas, but are not limited to. Submissions may span a broad range of topics, e.g.:
Track A: Security
Access control, Anonymity, Audit and audit reduction & Authentication and authorization, Applied
cryptography, Cryptanalysis, Digital Signatures, Biometric security, Boundary control devices,
Certification and accreditation, Cross-layer design for security, Security & Network Management, Data and
system integrity, Database security, Defensive information warfare, Denial of service protection, Intrusion
Detection, Anti-malware, Distributed systems security, Electronic commerce, E-mail security, Spam,
Phishing, E-mail fraud, Virus, worms, Trojan Protection, Grid security, Information hiding and
watermarking & Information survivability, Insider threat protection, Integrity
Intellectual property protection, Internet/Intranet Security, Key management and key recovery, Language-
based security, Mobile and wireless security, Mobile, Ad Hoc and Sensor Network Security, Monitoring
and surveillance, Multimedia security ,Operating system security, Peer-to-peer security, Performance
Evaluations of Protocols & Security Application, Privacy and data protection, Product evaluation criteria
and compliance, Risk evaluation and security certification, Risk/vulnerability assessment, Security &
Network Management, Security Models & protocols, Security threats & countermeasures (DDoS, MiM,
Session Hijacking, Replay attack etc,), Trusted computing, Ubiquitous Computing Security, Virtualization
security, VoIP security, Web 2.0 security, Submission Procedures, Active Defense Systems, Adaptive
Defense Systems, Benchmark, Analysis and Evaluation of Security Systems, Distributed Access Control
and Trust Management, Distributed Attack Systems and Mechanisms, Distributed Intrusion
Detection/Prevention Systems, Denial-of-Service Attacks and Countermeasures, High Performance
Security Systems, Identity Management and Authentication, Implementation, Deployment and
Management of Security Systems, Intelligent Defense Systems, Internet and Network Forensics, Large-
scale Attacks and Defense, RFID Security and Privacy, Security Architectures in Distributed Network
Systems, Security for Critical Infrastructures, Security for P2P systems and Grid Systems, Security in E-
Commerce, Security and Privacy in Wireless Networks, Secure Mobile Agents and Mobile Code, Security
Protocols, Security Simulation and Tools, Security Theory and Tools, Standards and Assurance Methods,
Trusted Computing, Viruses, Worms, and Other Malicious Code, World Wide Web Security, Novel and
emerging secure architecture, Study of attack strategies, attack modeling, Case studies and analysis of
actual attacks, Continuity of Operations during an attack, Key management, Trust management, Intrusion
detection techniques, Intrusion response, alarm management, and correlation analysis, Study of tradeoffs
between security and system performance, Intrusion tolerance systems, Secure protocols, Security in
wireless networks (e.g. mesh networks, sensor networks, etc.), Cryptography and Secure Communications,
Computer Forensics, Recovery and Healing, Security Visualization, Formal Methods in Security, Principles
for Designing a Secure Computing System, Autonomic Security, Internet Security, Security in Health Care
Systems, Security Solutions Using Reconfigurable Computing, Adaptive and Intelligent Defense Systems,
Authentication and Access control, Denial of service attacks and countermeasures, Identity, Route and
Location Anonymity schemes, Intrusion detection and prevention techniques, Cryptography, encryption
algorithms and Key management schemes, Secure routing schemes, Secure neighbor discovery and
localization, Trust establishment and maintenance, Confidentiality and data integrity, Security architectures,
deployments and solutions, Emerging threats to cloud-based services, Security model for new services,
Cloud-aware web service security, Information hiding in Cloud Computing, Securing distributed data
storage in cloud, Security, privacy and trust in mobile computing systems and applications, Middleware
security & Security features: middleware software is an asset on
its own and has to be protected, interaction between security-specific and other middleware features, e.g.,
context-awareness, Middleware-level security monitoring and measurement: metrics and mechanisms
for quantification and evaluation of security enforced by the middleware, Security co-design: trade-off and
co-design between application-based and middleware-based security, Policy-based management:
innovative support for policy-based definition and enforcement of security concerns, Identification and
authentication mechanisms: Means to capture application specific constraints in defining and enforcing
access control rules, Middleware-oriented security patterns: identification of patterns for sound, reusable
security, Security in aspect-based middleware: mechanisms for isolating and enforcing security aspects,
Security in agent-based platforms: protection for mobile code and platforms, Smart Devices: Biometrics,
National ID cards, Embedded Systems Security and TPMs, RFID Systems Security, Smart Card Security,
Pervasive Systems: Digital Rights Management (DRM) in pervasive environments, Intrusion Detection and
Information Filtering, Localization Systems Security (Tracking of People and Goods), Mobile Commerce
Security, Privacy Enhancing Technologies, Security Protocols (for Identification and Authentication,
Confidentiality and Privacy, and Integrity), Ubiquitous Networks: Ad Hoc Networks Security, Delay-
Tolerant Network Security, Domestic Network Security, Peer-to-Peer Networks Security, Security Issues
in Mobile and Ubiquitous Networks, Security of GSM/GPRS/UMTS Systems, Sensor Networks Security,
Vehicular Network Security, Wireless Communication Security: Bluetooth, NFC, WiFi, WiMAX,
WiMedia, others
This Track will emphasize the design, implementation, management and applications of computer
communications, networks and services. Topics of mostly theoretical nature are also welcome, provided
there is clear practical potential in applying the results of such work.
Broadband wireless technologies: LTE, WiMAX, WiRAN, HSDPA, HSUPA, Resource allocation and
interference management, Quality of service and scheduling methods, Capacity planning and dimensioning,
Cross-layer design and Physical layer based issue, Interworking architecture and interoperability, Relay
assisted and cooperative communications, Location and provisioning and mobility management, Call
admission and flow/congestion control, Performance optimization, Channel capacity modeling and analysis,
Middleware Issues: Event-based, publish/subscribe, and message-oriented middleware, Reconfigurable,
adaptable, and reflective middleware approaches, Middleware solutions for reliability, fault tolerance, and
quality-of-service, Scalability of middleware, Context-aware middleware, Autonomic and self-managing
middleware, Evaluation techniques for middleware solutions, Formal methods and tools for designing,
verifying, and evaluating, middleware, Software engineering techniques for middleware, Service oriented
middleware, Agent-based middleware, Security middleware, Network Applications: Network-based
automation, Cloud applications, Ubiquitous and pervasive applications, Collaborative applications, RFID
and sensor network applications, Mobile applications, Smart home applications, Infrastructure monitoring
and control applications, Remote health monitoring, GPS and location-based applications, Networked
vehicles applications, Alert applications, Embeded Computer System, Advanced Control Systems, and
Intelligent Control : Advanced control and measurement, computer and microprocessor-based control,
signal processing, estimation and identification techniques, application specific IC’s, nonlinear and
adaptive control, optimal and robot control, intelligent control, evolutionary computing, and intelligent
systems, instrumentation subject to critical conditions, automotive, marine and aero-space control and all
other control applications, Intelligent Control System, Wiring/Wireless Sensor, Signal Control System.
Sensors, Actuators and Systems Integration : Intelligent sensors and actuators, multisensor fusion, sensor
array and multi-channel processing, micro/nano technology, microsensors and microactuators,
instrumentation electronics, MEMS and system integration, wireless sensor, Network Sensor, Hybrid
Sensor, Distributed Sensor Networks. Signal and Image Processing : Digital signal processing theory,
methods, DSP implementation, speech processing, image and multidimensional signal processing, Image
analysis and processing, Image and Multimedia applications, Real-time multimedia signal processing,
Computer vision, Emerging signal processing areas, Remote Sensing, Signal processing in education.
Industrial Informatics: Industrial applications of neural networks, fuzzy algorithms, Neuro-Fuzzy
application, bioInformatics, real-time computer control, real-time information systems, human-machine
interfaces, CAD/CAM/CAT/CIM, virtual reality, industrial communications, flexible manufacturing
systems, industrial automated process, Data Storage Management, Harddisk control, Supply Chain
Management, Logistics applications, Power plant automation, Drives automation. Information Technology,
Management of Information System : Management information systems, Information Management,
Nursing information management, Information System, Information Technology and their application, Data
retrieval, Data Base Management, Decision analysis methods, Information processing, Operations research,
E-Business, E-Commerce, E-Government, Computer Business, Security and risk management, Medical
imaging, Biotechnology, Bio-Medicine, Computer-based information systems in health care, Changing
Access to Patient Information, Healthcare Management Information Technology.
Communication/Computer Network, Transportation Application : On-board diagnostics, Active safety
systems, Communication systems, Wireless technology, Communication application, Navigation and
Guidance, Vision-based applications, Speech interface, Sensor fusion, Networking theory and technologies,
Transportation information, Autonomous vehicle, Vehicle application of affective computing, Advance
Computing technology and their application : Broadband and intelligent networks, Data Mining, Data
fusion, Computational intelligence, Information and data security, Information indexing and retrieval,
Information processing, Information systems and applications, Internet applications and performances,
Knowledge based systems, Knowledge management, Software Engineering, Decision making, Mobile
networks and services, Network management and services, Neural Network, Fuzzy logics, Neuro-Fuzzy,
Expert approaches, Innovation Technology and Management : Innovation and product development,
Emerging advances in business and its applications, Creativity in Internet management and retailing, B2B
and B2C management, Electronic transceiver device for Retail Marketing Industries, Facilities planning
and management, Innovative pervasive computing applications, Programming paradigms for pervasive
systems, Software evolution and maintenance in pervasive systems, Middleware services and agent
technologies, Adaptive, autonomic and context-aware computing, Mobile/Wireless computing systems and
services in pervasive computing, Energy-efficient and green pervasive computing, Communication
architectures for pervasive computing, Ad hoc networks for pervasive communications, Pervasive
opportunistic communications and applications, Enabling technologies for pervasive systems (e.g., wireless
BAN, PAN), Positioning and tracking technologies, Sensors and RFID in pervasive systems, Multimodal
sensing and context for pervasive applications, Pervasive sensing, perception and semantic interpretation,
Smart devices and intelligent environments, Trust, security and privacy issues in pervasive systems, User
interfaces and interaction models, Virtual immersive communications, Wearable computers, Standards and
interfaces for pervasive computing environments, Social and economic models for pervasive systems,
Active and Programmable Networks, Ad Hoc & Sensor Network, Congestion and/or Flow Control, Content
Distribution, Grid Networking, High-speed Network Architectures, Internet Services and Applications,
Optical Networks, Mobile and Wireless Networks, Network Modeling and Simulation, Multicast,
Multimedia Communications, Network Control and Management, Network Protocols, Network
Performance, Network Measurement, Peer to Peer and Overlay Networks, Quality of Service and Quality
of Experience, Ubiquitous Networks, Crosscutting Themes – Internet Technologies, Infrastructure,
Services and Applications; Open Source Tools, Open Models and Architectures; Security, Privacy and
Trust; Navigation Systems, Location Based Services; Social Networks and Online Communities; ICT
Convergence, Digital Economy and Digital Divide, Neural Networks, Pattern Recognition, Computer
Vision, Advanced Computing Architectures and New Programming Models, Visualization and Virtual
Reality as Applied to Computational Science, Computer Architecture and Embedded Systems, Technology
in Education, Theoretical Computer Science, Computing Ethics, Computing Practices & Applications
Authors are invited to submit papers through e-mail ijcsiseditor@gmail.com. Submissions must be original
and should not have been published previously or be under consideration for publication while being
evaluated by IJCSIS. Before submission authors should carefully read over the journal's Author Guidelines,
which are located at http://sites.google.com/site/ijcsis/authors-notes .
© IJCSIS PUBLICATION 2016
ISSN 1947 5500
http://sites.google.com/site/ijcsis/