Pub Date : 2015-08-20DOI: 10.1109/IC3.2015.7346676
Prabhat Dansena, K. P. Kumar, R. Pal
Automatic extraction of important regions from a cheque image helps in automatic analysis of the cheque. It can be used for automated clearing of cheques, detection of frauds in the cheques, and so on. A novel approach of extracting important regions from a cheque image is proposed, in this paper, based on identification of lines. Experimental results demonstrate the success of the proposed approach.
{"title":"Line based extraction of important regions from a cheque image","authors":"Prabhat Dansena, K. P. Kumar, R. Pal","doi":"10.1109/IC3.2015.7346676","DOIUrl":"https://doi.org/10.1109/IC3.2015.7346676","url":null,"abstract":"Automatic extraction of important regions from a cheque image helps in automatic analysis of the cheque. It can be used for automated clearing of cheques, detection of frauds in the cheques, and so on. A novel approach of extracting important regions from a cheque image is proposed, in this paper, based on identification of lines. Experimental results demonstrate the success of the proposed approach.","PeriodicalId":217950,"journal":{"name":"2015 Eighth International Conference on Contemporary Computing (IC3)","volume":"98 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122641255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2015-08-20DOI: 10.1109/IC3.2015.7346664
Divyesh Patel, T. Srivastava
The field of Discrete Tomography (DT) deals with the reconstruction of 2D discrete images from a few number of their projections. The ideal problem of DT is to reconstruct a binary image from its horizontal and vertical projections. It turns out that this problem of DT is highly underdetermined and therefore it is inevitable to impose additional constraints to this problem. This paper uses the convexity property of binary images and the problem of reconstruction of h-convex binary images from its horizontal and vertical projections is considered here. This problem is transformed into two different optimization problems by defining two appropriate objective functions. Then two simulated annealing (SA) algorithms to solve the two optimization problems are developed. The SA algorithms are tested on various randomly generated test images. The algorithms are also tested on noisy images. Finally numerical results have been reported showing good reconstruction fidelity.
{"title":"Reconstructing h-convex binary images from its horizontal and vertical projections by simulated annealing","authors":"Divyesh Patel, T. Srivastava","doi":"10.1109/IC3.2015.7346664","DOIUrl":"https://doi.org/10.1109/IC3.2015.7346664","url":null,"abstract":"The field of Discrete Tomography (DT) deals with the reconstruction of 2D discrete images from a few number of their projections. The ideal problem of DT is to reconstruct a binary image from its horizontal and vertical projections. It turns out that this problem of DT is highly underdetermined and therefore it is inevitable to impose additional constraints to this problem. This paper uses the convexity property of binary images and the problem of reconstruction of h-convex binary images from its horizontal and vertical projections is considered here. This problem is transformed into two different optimization problems by defining two appropriate objective functions. Then two simulated annealing (SA) algorithms to solve the two optimization problems are developed. The SA algorithms are tested on various randomly generated test images. The algorithms are also tested on noisy images. Finally numerical results have been reported showing good reconstruction fidelity.","PeriodicalId":217950,"journal":{"name":"2015 Eighth International Conference on Contemporary Computing (IC3)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124597731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2015-08-20DOI: 10.1109/IC3.2015.7346653
Juan J. Noda Arencibia, C. Travieso-González, David Sánchez-Rodríguez, M. Dutta, Garima Vyas
This paper presents a new approach for the acoustic classification of frogs' calls using a novel fusion of features: Mel Frequency Cepstral Coefficients (MFCCs), Shannon entropy and syllable duration. First, the audio recordings of different frogs' species are segmented in syllables. For each syllable, each feature is extracted and the cepstral features (MFCC) are computed and evaluated separately as in previous works. Finally, the data fusion is used to train a multiclass Support Vector Machine (SVM) classifier. In our experiment, the results show that our novel feature fusion increase the classification accuracy; achieving an average of 94.21% ± 8,04 in 18 frog's species.
{"title":"Automatic classification of frogs calls based on fusion of features and SVM","authors":"Juan J. Noda Arencibia, C. Travieso-González, David Sánchez-Rodríguez, M. Dutta, Garima Vyas","doi":"10.1109/IC3.2015.7346653","DOIUrl":"https://doi.org/10.1109/IC3.2015.7346653","url":null,"abstract":"This paper presents a new approach for the acoustic classification of frogs' calls using a novel fusion of features: Mel Frequency Cepstral Coefficients (MFCCs), Shannon entropy and syllable duration. First, the audio recordings of different frogs' species are segmented in syllables. For each syllable, each feature is extracted and the cepstral features (MFCC) are computed and evaluated separately as in previous works. Finally, the data fusion is used to train a multiclass Support Vector Machine (SVM) classifier. In our experiment, the results show that our novel feature fusion increase the classification accuracy; achieving an average of 94.21% ± 8,04 in 18 frog's species.","PeriodicalId":217950,"journal":{"name":"2015 Eighth International Conference on Contemporary Computing (IC3)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117138341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2015-08-20DOI: 10.1109/IC3.2015.7346706
E. Katta, Anuja Arora
Cross Language Information Retrieval (CLIR) is a sub domain of Information Retrieval. It deals with retrieval of information in a specified language that is different from the language of user's query. In this paper, an improved English-Hindi based CLIR is proposed. There are various un-noticed domains in this broad research area that are required to be worked upon in order to improve the performance of an English-Hindi based CLIR. Not much research effort has been put up to improve the searching and ranking aspects of CLIR systems, especially in case of English-Hindi based CLIR. This paper focuses on applying algorithms like Naïve Bayes and particle swarm optimization in order to improve ranking and searching aspects of a CLIR system. We matched terms contained in documents to the query terms in same sequence as present in the search query to make our system more efficient. Along with this our approach also makes use of bilingual English-Hindi translator for query conversion in Hindi language. Further, we use Hindi query extension and synonym generation which helps in retrieving more relevant results in an English-Hindi based CLIR as compared to existing one. Both of these techniques applied to this improved approach gives user a change to choose more appropriate Hindi query than just by using the single translated query and hence improving overall performance.
{"title":"An improved approach to English-Hindi based Cross Language Information Retrieval system","authors":"E. Katta, Anuja Arora","doi":"10.1109/IC3.2015.7346706","DOIUrl":"https://doi.org/10.1109/IC3.2015.7346706","url":null,"abstract":"Cross Language Information Retrieval (CLIR) is a sub domain of Information Retrieval. It deals with retrieval of information in a specified language that is different from the language of user's query. In this paper, an improved English-Hindi based CLIR is proposed. There are various un-noticed domains in this broad research area that are required to be worked upon in order to improve the performance of an English-Hindi based CLIR. Not much research effort has been put up to improve the searching and ranking aspects of CLIR systems, especially in case of English-Hindi based CLIR. This paper focuses on applying algorithms like Naïve Bayes and particle swarm optimization in order to improve ranking and searching aspects of a CLIR system. We matched terms contained in documents to the query terms in same sequence as present in the search query to make our system more efficient. Along with this our approach also makes use of bilingual English-Hindi translator for query conversion in Hindi language. Further, we use Hindi query extension and synonym generation which helps in retrieving more relevant results in an English-Hindi based CLIR as compared to existing one. Both of these techniques applied to this improved approach gives user a change to choose more appropriate Hindi query than just by using the single translated query and hence improving overall performance.","PeriodicalId":217950,"journal":{"name":"2015 Eighth International Conference on Contemporary Computing (IC3)","volume":"os-12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127760755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2015-08-20DOI: 10.1109/IC3.2015.7346736
Pratik Ranjan, H. Om
The signature schemes are used to verify the authenticity of a signature and the corresponding documents. The undeniable signature schemes are challenge and response based interactive schemes, where the active participation of signer is compulsory. These schemes are used in private communication where the confidential deals and agreements take place as a legitimate signer cannot deny his signature. In this paper, we analyze the Thomas and Lal's braid group based zero-knowledge undeniable signature scheme and show that it is insecure against the man-in-the-middle and impersonation attacks. In addition, we propose an efficient undeniable signature scheme using the braid groups that provides secrecy and authenticity of a legitimate signer. Furthermore, we show that our scheme is secure against the above mentioned attacks.
{"title":"An efficient undeniable signature scheme using braid groups","authors":"Pratik Ranjan, H. Om","doi":"10.1109/IC3.2015.7346736","DOIUrl":"https://doi.org/10.1109/IC3.2015.7346736","url":null,"abstract":"The signature schemes are used to verify the authenticity of a signature and the corresponding documents. The undeniable signature schemes are challenge and response based interactive schemes, where the active participation of signer is compulsory. These schemes are used in private communication where the confidential deals and agreements take place as a legitimate signer cannot deny his signature. In this paper, we analyze the Thomas and Lal's braid group based zero-knowledge undeniable signature scheme and show that it is insecure against the man-in-the-middle and impersonation attacks. In addition, we propose an efficient undeniable signature scheme using the braid groups that provides secrecy and authenticity of a legitimate signer. Furthermore, we show that our scheme is secure against the above mentioned attacks.","PeriodicalId":217950,"journal":{"name":"2015 Eighth International Conference on Contemporary Computing (IC3)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132956689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2015-08-20DOI: 10.1109/IC3.2015.7346648
Chandresh Kumar Maurya, Durga Toshniwal, G. V. Venkoparao
Anomaly detection is an important task in many real world applications such as fraud detection, suspicious activity detection, health care monitoring etc. In this paper, we tackle this problem from supervised learning perspective in online learning setting. We maximize well known Gmean metric for class-imbalance learning in online learning framework. Specifically, we show that maximizing Gmean is equivalent to minimizing a convex surrogate loss function and based on that we propose novel online learning algorithm for anomaly detection. We then show, by extensive experiments, that the performance of the proposed algorithm with respect to sum metric is as good as a recently proposed Cost-Sensitive Online Classification(CSOC) algorithm for class-imbalance learning over various benchmarked data sets while keeping running time close to the perception algorithm. Our another conclusion is that other competitive online algorithms do not perform consistently over data sets of varying size. This shows the potential applicability of our proposed approach.
{"title":"Online anomaly detection via class-imbalance learning","authors":"Chandresh Kumar Maurya, Durga Toshniwal, G. V. Venkoparao","doi":"10.1109/IC3.2015.7346648","DOIUrl":"https://doi.org/10.1109/IC3.2015.7346648","url":null,"abstract":"Anomaly detection is an important task in many real world applications such as fraud detection, suspicious activity detection, health care monitoring etc. In this paper, we tackle this problem from supervised learning perspective in online learning setting. We maximize well known Gmean metric for class-imbalance learning in online learning framework. Specifically, we show that maximizing Gmean is equivalent to minimizing a convex surrogate loss function and based on that we propose novel online learning algorithm for anomaly detection. We then show, by extensive experiments, that the performance of the proposed algorithm with respect to sum metric is as good as a recently proposed Cost-Sensitive Online Classification(CSOC) algorithm for class-imbalance learning over various benchmarked data sets while keeping running time close to the perception algorithm. Our another conclusion is that other competitive online algorithms do not perform consistently over data sets of varying size. This shows the potential applicability of our proposed approach.","PeriodicalId":217950,"journal":{"name":"2015 Eighth International Conference on Contemporary Computing (IC3)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132915173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2015-08-20DOI: 10.1109/IC3.2015.7346643
Shailendra Tiwari, R. Srivastava
Bayesian statistical algorithm plays a significant role in the quality of the images produced by Emission Tomography like PET/SPECT, since they can provide an accurate system model. The major drawbacks associated with this algorithm include the problem of slow convergence, choice of optimum initial point and ill-posedness. To address these issues, in this paper a hybrid-cascaded framework for Median Root Prior (MRP) based reconstruction algorithm is proposed. This framework consists of breaking the reconstruction process into two parts viz. primary and secondary. During primary part, simultaneous algebraic reconstruction technique (SART) is applied to overcome the problems of slow convergence and initialization. It provides fast convergence and produce good reconstruction results with lesser number of iterations than other iterative methods. The task of primary part is to provide an enhanced image to secondary part to be used as an initial estimate for reconstruction process. The secondary part is a hybrid combination of two parts namely the reconstruction part and the prior part. The reconstruction is done using Median Root Prior (MRP) while Anisotropic Diffusion (AD) is used as prior to deal with ill-posedness. A comparative analysis of the proposed model with some other standard methods in literature is presented both qualitatively and quantitatively for a simulated phantom and a standard medical image test data. Using cascaded primary and secondary reconstruction steps, yields significant improvements in reconstructed image quality. It also accelerates the convergence and provides enhanced results using the projection data. The obtained result justifies the applicability of the proposed method.
{"title":"An efficient and modified median root prior based framework for PET/SPECT reconstruction algorithm","authors":"Shailendra Tiwari, R. Srivastava","doi":"10.1109/IC3.2015.7346643","DOIUrl":"https://doi.org/10.1109/IC3.2015.7346643","url":null,"abstract":"Bayesian statistical algorithm plays a significant role in the quality of the images produced by Emission Tomography like PET/SPECT, since they can provide an accurate system model. The major drawbacks associated with this algorithm include the problem of slow convergence, choice of optimum initial point and ill-posedness. To address these issues, in this paper a hybrid-cascaded framework for Median Root Prior (MRP) based reconstruction algorithm is proposed. This framework consists of breaking the reconstruction process into two parts viz. primary and secondary. During primary part, simultaneous algebraic reconstruction technique (SART) is applied to overcome the problems of slow convergence and initialization. It provides fast convergence and produce good reconstruction results with lesser number of iterations than other iterative methods. The task of primary part is to provide an enhanced image to secondary part to be used as an initial estimate for reconstruction process. The secondary part is a hybrid combination of two parts namely the reconstruction part and the prior part. The reconstruction is done using Median Root Prior (MRP) while Anisotropic Diffusion (AD) is used as prior to deal with ill-posedness. A comparative analysis of the proposed model with some other standard methods in literature is presented both qualitatively and quantitatively for a simulated phantom and a standard medical image test data. Using cascaded primary and secondary reconstruction steps, yields significant improvements in reconstructed image quality. It also accelerates the convergence and provides enhanced results using the projection data. The obtained result justifies the applicability of the proposed method.","PeriodicalId":217950,"journal":{"name":"2015 Eighth International Conference on Contemporary Computing (IC3)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133725086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2015-08-20DOI: 10.1109/IC3.2015.7346720
A. P. Shah, V. Neema, Shreeniwas Daulatabad
Dynamic CMOS logic circuits are used in modern VLSI circuits because of its high system performance and its performance is high due to higher speed over static CMOS circuit. However dynamic logic circuit has less noise immunity and increased leakage power dissipation. Increase in leakage current combine with reduced noise margin results in performance degradation of dynamic circuits. In this paper DOIND logic approach is proposed for domino logic which reduces the leakage current with minimum delay penalty. Simulation is performed at 70 nm technology node for a domino logic and DOIND logic buffer using tanner EDA tool. Simulation results shows that proposed DOIND approach decreases the leakage current 93.3%, static power 93.3% and static energy 86.66% at supply voltage 1.15V. Proposed circuit also improves dynamic power 60.78%, dynamic energy delay product (EDP) 62.18% and dynamic power delay product (PDP) 62.07% at 1.15V supply voltage.
{"title":"A novel leakage reduction DOIND approach for nanoscale domino logic circuits","authors":"A. P. Shah, V. Neema, Shreeniwas Daulatabad","doi":"10.1109/IC3.2015.7346720","DOIUrl":"https://doi.org/10.1109/IC3.2015.7346720","url":null,"abstract":"Dynamic CMOS logic circuits are used in modern VLSI circuits because of its high system performance and its performance is high due to higher speed over static CMOS circuit. However dynamic logic circuit has less noise immunity and increased leakage power dissipation. Increase in leakage current combine with reduced noise margin results in performance degradation of dynamic circuits. In this paper DOIND logic approach is proposed for domino logic which reduces the leakage current with minimum delay penalty. Simulation is performed at 70 nm technology node for a domino logic and DOIND logic buffer using tanner EDA tool. Simulation results shows that proposed DOIND approach decreases the leakage current 93.3%, static power 93.3% and static energy 86.66% at supply voltage 1.15V. Proposed circuit also improves dynamic power 60.78%, dynamic energy delay product (EDP) 62.18% and dynamic power delay product (PDP) 62.07% at 1.15V supply voltage.","PeriodicalId":217950,"journal":{"name":"2015 Eighth International Conference on Contemporary Computing (IC3)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130842147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2015-08-20DOI: 10.1109/IC3.2015.7346690
Arijul Haque, K. S. Rao
This work explores the spectral energies of neutral, sad and angry speech, and analyzes the potential of spectral energy modification to convert neutral speech to sad/angry speech. A method of modifying the spectral energy of neutral speech signals based on a filter bank implementation is proposed for the purpose of converting a given neutral speech to a target emotional speech. Since pitch plays a vital role in emotion expression, we modify the pitch contour first by using the method of Gaussian normalization. This is followed by modification of spectral energy using a method proposed in this paper. The expressiveness of the resultant speech is compared with speech obtained by modifying only the pitch contour, and we have observed improvements in expressiveness due to incorporation of proposed spectral energy modification. The method is found to be quite good for neutral to sad conversion. However, the quality of conversion to anger is not good, and the reasons behind this are analyzed.
{"title":"Analysis and modification of spectral energy for neutral to sad emotion conversion","authors":"Arijul Haque, K. S. Rao","doi":"10.1109/IC3.2015.7346690","DOIUrl":"https://doi.org/10.1109/IC3.2015.7346690","url":null,"abstract":"This work explores the spectral energies of neutral, sad and angry speech, and analyzes the potential of spectral energy modification to convert neutral speech to sad/angry speech. A method of modifying the spectral energy of neutral speech signals based on a filter bank implementation is proposed for the purpose of converting a given neutral speech to a target emotional speech. Since pitch plays a vital role in emotion expression, we modify the pitch contour first by using the method of Gaussian normalization. This is followed by modification of spectral energy using a method proposed in this paper. The expressiveness of the resultant speech is compared with speech obtained by modifying only the pitch contour, and we have observed improvements in expressiveness due to incorporation of proposed spectral energy modification. The method is found to be quite good for neutral to sad conversion. However, the quality of conversion to anger is not good, and the reasons behind this are analyzed.","PeriodicalId":217950,"journal":{"name":"2015 Eighth International Conference on Contemporary Computing (IC3)","volume":"148 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122919079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2015-08-20DOI: 10.1109/IC3.2015.7346705
P. Suja, P. KalyanKumarV., Shikha Tripathi
Emotions are characterized as responses to internal and external events of a person. Emotion recognition through facial expressions from videos plays a vital role in human computer interaction where the dynamic changes in face movements needs to be realized quickly. In this work, we propose a simple method, using the geometrical based approach for the recognition of six basic emotions in video sequences of BU-4DFE database. We have chosen optimum feature points out of the 83 feature points provided in the BU-4DFE database. A video expressing emotion will have frames containing neutral, onset, apex and offset of that emotion. We have dynamically identified the frame that is most expressive for an emotion (apex). The Euclidean distance between the feature points in apex and neutral frame is determined and their difference in corresponding neutral and the apex frame is calculated to form the feature vector. The feature vectors thus formed for all the emotions and subjects are given to Neural Networks (NN) and Support Vector Machine (SVM) with different kernels for classification. We have compared the accuracy obtained by NN & SVM. Our proposed method is simple, uses only two frames and yields good accuracy for BU-4DFE database. Very complex algorithms exist in literature using BU-4DFE database and our proposed simple method gives comparable results. It can be applied for real time implementation and kinesics in future.
{"title":"Dynamic facial emotion recognition from 4D video sequences","authors":"P. Suja, P. KalyanKumarV., Shikha Tripathi","doi":"10.1109/IC3.2015.7346705","DOIUrl":"https://doi.org/10.1109/IC3.2015.7346705","url":null,"abstract":"Emotions are characterized as responses to internal and external events of a person. Emotion recognition through facial expressions from videos plays a vital role in human computer interaction where the dynamic changes in face movements needs to be realized quickly. In this work, we propose a simple method, using the geometrical based approach for the recognition of six basic emotions in video sequences of BU-4DFE database. We have chosen optimum feature points out of the 83 feature points provided in the BU-4DFE database. A video expressing emotion will have frames containing neutral, onset, apex and offset of that emotion. We have dynamically identified the frame that is most expressive for an emotion (apex). The Euclidean distance between the feature points in apex and neutral frame is determined and their difference in corresponding neutral and the apex frame is calculated to form the feature vector. The feature vectors thus formed for all the emotions and subjects are given to Neural Networks (NN) and Support Vector Machine (SVM) with different kernels for classification. We have compared the accuracy obtained by NN & SVM. Our proposed method is simple, uses only two frames and yields good accuracy for BU-4DFE database. Very complex algorithms exist in literature using BU-4DFE database and our proposed simple method gives comparable results. It can be applied for real time implementation and kinesics in future.","PeriodicalId":217950,"journal":{"name":"2015 Eighth International Conference on Contemporary Computing (IC3)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115173525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}