Takahiro Hayashi, N. Ishii, Masato Ishimori, K. Abe
This paper reports improvement of indirect matching, which is a fast CBMIR (content-based music information retrieval) framework proposed in our previous study. Indirect matching achieves fast retrieval by combining offline search with representative queries and online quick similarity estimation based on the results of the offline search. We have found that the retrieval accuracy of indirect matching decreases when representative queries have little variation. This paper proposes a method for selecting representative queries having wide variation. To ensure wide variation between representative queries, the proposed method combines MDS (multi-dimensional scaling) and Ward's clustering. Experimental results have shown that the retrieval accuracy of indirect matching can be stabilized by the proposed method.
{"title":"Stability Improvement of Indirect Matching for Music Information Retrieval","authors":"Takahiro Hayashi, N. Ishii, Masato Ishimori, K. Abe","doi":"10.1109/ISM.2015.29","DOIUrl":"https://doi.org/10.1109/ISM.2015.29","url":null,"abstract":"This paper reports improvement of indirect matching, which is a fast CBMIR (content-based music information retrieval) framework proposed in our previous study. Indirect matching achieves fast retrieval by combining offline search with representative queries and online quick similarity estimation based on the results of the offline search. We have found that the retrieval accuracy of indirect matching decreases when representative queries have little variation. This paper proposes a method for selecting representative queries having wide variation. To ensure wide variation between representative queries, the proposed method combines MDS (multi-dimensional scaling) and Ward's clustering. Experimental results have shown that the retrieval accuracy of indirect matching can be stabilized by the proposed method.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"91 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126173949","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In this paper we present an open, secure and flexible architectural approach for integrating Content Decryption Module (CDM) and Encryption Media Extensions (EME) interface with Trusted Execution Environment (TEE). This provides security hardening for playback of the Digital Rights Management (DRM) protected content without any need for a dedicated secure processor. This article is not focused on specific TEE approaches, but rather considers that any TEE can be potentially integrated with CDM through the EME interface. Our work introduces the approaches for integrating CDM with Intel SGX and ARM TrustZone TEE technologies.
{"title":"Towards Security of Native DRM Execution in HTML5","authors":"D. Livshits, A. Mikityuk, S. Pham, A. Shabtai","doi":"10.1109/ISM.2015.48","DOIUrl":"https://doi.org/10.1109/ISM.2015.48","url":null,"abstract":"In this paper we present an open, secure and flexible architectural approach for integrating Content Decryption Module (CDM) and Encryption Media Extensions (EME) interface with Trusted Execution Environment (TEE). This provides security hardening for playback of the Digital Rights Management (DRM) protected content without any need for a dedicated secure processor. This article is not focused on specific TEE approaches, but rather considers that any TEE can be potentially integrated with CDM through the EME interface. Our work introduces the approaches for integrating CDM with Intel SGX and ARM TrustZone TEE technologies.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"98 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127500402","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Real time detection and tracking of multi vehicles in airborne videos is still a challenging problem due to the camera motion and low resolution. In this paper, a real time technique for simultaneously detecting, tracking and counting vehicles in airborne and stationary camera videos is proposed. First, feature points are extracted and tracked through video frames. A new strategy is used for removing the non-stationary background points by measuring the changes in the histogram of the pixels around each feature point with time. The obtained foreground features are clustered and grouped into separate trackable vehicles based on their motion properties. Experimental results performed on videos representing airborne and fixed cameras confirm the excellent properties of the proposed algorithm.
{"title":"A Novel Algorithm for Vehicle Detection and Tracking in Airborne Videos","authors":"M. Abdelwahab, M. Abdelwahab","doi":"10.1109/ISM.2015.77","DOIUrl":"https://doi.org/10.1109/ISM.2015.77","url":null,"abstract":"Real time detection and tracking of multi vehicles in airborne videos is still a challenging problem due to the camera motion and low resolution. In this paper, a real time technique for simultaneously detecting, tracking and counting vehicles in airborne and stationary camera videos is proposed. First, feature points are extracted and tracked through video frames. A new strategy is used for removing the non-stationary background points by measuring the changes in the histogram of the pixels around each feature point with time. The obtained foreground features are clustered and grouped into separate trackable vehicles based on their motion properties. Experimental results performed on videos representing airborne and fixed cameras confirm the excellent properties of the proposed algorithm.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134424083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Seethal Paluri, K. Kambhatla, M. Medley, J. Matyjas, Sunil Kumar
We present a real-time, priority-aware joint packet fragmentation and error protection scheme for transmitting H.264/AVC compressed video over Rayleigh fading channels. A priority is assigned to every video slice based on the predicted cumulative mean squared error (CMSE) contributed by its loss relative to all the slices of the group of pictures (GOP). Here, the CMSE of a slice, which is predicted by using our low-complexity generalized linear model, represents the degradation in video quality contributed by its loss. The slices of a frame belonging to the same priority are aggregated to form video packet(s). We simulate the fragment error rates (FERs) for a combination of different fragment sizes and rate compatible punctured convolutional (RCPC) code rates. These FERs are then used to determine the optimal fragment sizes and code rates for packets of each priority class by minimizing the expected normalized predicted CMSE of all the priority classes per GOP in H.264 video bitstream. We observed a significant improvement in the received video quality over the conventional and priority-agnostic packet fragmentation schemes.
{"title":"Priority-Aware Joint Packet Fragmentation and Error Protection Scheme for H.264 Video over Wireless Channels","authors":"Seethal Paluri, K. Kambhatla, M. Medley, J. Matyjas, Sunil Kumar","doi":"10.1109/ISM.2015.47","DOIUrl":"https://doi.org/10.1109/ISM.2015.47","url":null,"abstract":"We present a real-time, priority-aware joint packet fragmentation and error protection scheme for transmitting H.264/AVC compressed video over Rayleigh fading channels. A priority is assigned to every video slice based on the predicted cumulative mean squared error (CMSE) contributed by its loss relative to all the slices of the group of pictures (GOP). Here, the CMSE of a slice, which is predicted by using our low-complexity generalized linear model, represents the degradation in video quality contributed by its loss. The slices of a frame belonging to the same priority are aggregated to form video packet(s). We simulate the fragment error rates (FERs) for a combination of different fragment sizes and rate compatible punctured convolutional (RCPC) code rates. These FERs are then used to determine the optimal fragment sizes and code rates for packets of each priority class by minimizing the expected normalized predicted CMSE of all the priority classes per GOP in H.264 video bitstream. We observed a significant improvement in the received video quality over the conventional and priority-agnostic packet fragmentation schemes.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"159 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131644842","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The use of Voice over Internet Protocol (VoIP) is becoming ubiquitous due to the multiple shortcomings of traditional Public Switched Telephone Network (PSTN) systems. As a result, the development of secure key establishment protocols is becoming increasingly important. The Communications-Electronics Security Group (CESG), in response to this demand, has published new key agreement protocols for the Multimedia Internet KEYing (MIKEY) protocol to provide low-cost secure VoIP communications, supported on Identity-based Public-Key Cryptography (IDPKC). In the context of IDPKC, the identity of users is used to derive their public-keys, which eliminates the expenses of maintaining a Public-Key Infrastructure (PKI). However, IDPKC systems suffer from inefficient user revocation and key renewal. In this paper, we take advantage of the fact that users need to be connected to the Internet to communicate, for introducing a SEcurity Mediator (SEM), who possesses a share of the users' private-keys, and with whom the users must cooperate, to sign and decrypt cryptograms. By taking advantage of this sharing, we introduce mechanisms to provide immediate user revocation and key renewal.
由于传统的公共交换电话网(PSTN)系统存在诸多缺陷,VoIP (Voice over Internet Protocol)的使用变得越来越普遍。因此,开发安全的密钥建立协议变得越来越重要。为了响应这一需求,通信电子安全组(CESG)发布了多媒体互联网密钥(MIKEY)协议的新密钥协议协议,以提供基于身份的公钥加密(IDPKC)支持的低成本安全VoIP通信。在IDPKC上下文中,用户的身份用于派生其公钥,从而消除了维护公钥基础设施(PKI)的费用。然而,IDPKC系统存在低效的用户撤销和密钥更新问题。在本文中,我们利用用户需要连接到互联网进行通信的事实,引入了一个安全中介(SEM),它拥有用户私钥的共享,并且必须与用户合作,对密码进行签名和解密。通过利用这种共享,我们引入了提供即时用户撤销和密钥更新的机制。
{"title":"Featuring Immediate Revocation in Mikey-Sakke (FIRM)","authors":"P. Martins, L. Sousa, P. Chawan","doi":"10.1109/ISM.2015.101","DOIUrl":"https://doi.org/10.1109/ISM.2015.101","url":null,"abstract":"The use of Voice over Internet Protocol (VoIP) is becoming ubiquitous due to the multiple shortcomings of traditional Public Switched Telephone Network (PSTN) systems. As a result, the development of secure key establishment protocols is becoming increasingly important. The Communications-Electronics Security Group (CESG), in response to this demand, has published new key agreement protocols for the Multimedia Internet KEYing (MIKEY) protocol to provide low-cost secure VoIP communications, supported on Identity-based Public-Key Cryptography (IDPKC). In the context of IDPKC, the identity of users is used to derive their public-keys, which eliminates the expenses of maintaining a Public-Key Infrastructure (PKI). However, IDPKC systems suffer from inefficient user revocation and key renewal. In this paper, we take advantage of the fact that users need to be connected to the Internet to communicate, for introducing a SEcurity Mediator (SEM), who possesses a share of the users' private-keys, and with whom the users must cooperate, to sign and decrypt cryptograms. By taking advantage of this sharing, we introduce mechanisms to provide immediate user revocation and key renewal.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"421 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131676675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In this paper, we present the design and implementation of toy car localization and navigation system, which enables toy cars and "passengers" to learn and exchange their fine-grained locations in an indoor environment with the help of the projected light based localization technique. The projected light consists of a sequence of gray code images which assigns each pixel in the projection area a unique gray code to distinguish their coordination. The light sensors installed on a toy cars and a potential "passenger" receive the light streams from the projected light, based on which their locations are inferred. The toy car then utilizes A* algorithm to plan a route based on its location, its orientation, the target's location and the map of "roads". The fast speed of projected light based localization technique enables the toy car to adjust its own orientation while "driving" and keep itself on "roads". The toy car system demonstrates that the localization technique and the client-server architecture can benefit similar applications that require fine-grained location information of multiple objects simultaneously.
{"title":"Smart Toy Car Localization and Navigation Using Projected Light","authors":"Mingming Fan, Qiong Liu, Shang Ma, Patrick Chiu","doi":"10.1109/ISM.2015.113","DOIUrl":"https://doi.org/10.1109/ISM.2015.113","url":null,"abstract":"In this paper, we present the design and implementation of toy car localization and navigation system, which enables toy cars and \"passengers\" to learn and exchange their fine-grained locations in an indoor environment with the help of the projected light based localization technique. The projected light consists of a sequence of gray code images which assigns each pixel in the projection area a unique gray code to distinguish their coordination. The light sensors installed on a toy cars and a potential \"passenger\" receive the light streams from the projected light, based on which their locations are inferred. The toy car then utilizes A* algorithm to plan a route based on its location, its orientation, the target's location and the map of \"roads\". The fast speed of projected light based localization technique enables the toy car to adjust its own orientation while \"driving\" and keep itself on \"roads\". The toy car system demonstrates that the localization technique and the client-server architecture can benefit similar applications that require fine-grained location information of multiple objects simultaneously.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131826354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper presents a GPU based image registration algorithm that utilizes Hough Transform and Least Square Optimization to calculate the transformation between two images. In our approach, we calculate the transformation parameters of all possible combination solutions of matched feature points by exploiting parallel processing power of the GPU. We applied our algorithm on a variety of images including the problem of mosaic image generation. Experimental results show that our method is robust to the outliers (incorrect matches) and it can achieve very accurate registration (numeric and visual) results with much faster (up to 20 times) than CPU implementation.
{"title":"GPU Based Robust Image Registration for Composite Translational, Rotational and Scale Transformations","authors":"S. Dinç, R. S. Aygün, F. Fahimi","doi":"10.1109/ISM.2015.51","DOIUrl":"https://doi.org/10.1109/ISM.2015.51","url":null,"abstract":"This paper presents a GPU based image registration algorithm that utilizes Hough Transform and Least Square Optimization to calculate the transformation between two images. In our approach, we calculate the transformation parameters of all possible combination solutions of matched feature points by exploiting parallel processing power of the GPU. We applied our algorithm on a variety of images including the problem of mosaic image generation. Experimental results show that our method is robust to the outliers (incorrect matches) and it can achieve very accurate registration (numeric and visual) results with much faster (up to 20 times) than CPU implementation.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129229146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
C. Beecks, Marwan Hassani, Florian Obeloer, T. Seidl
One of the most fundamental challenges when accessing gestural patterns in 3D motion capture databases is the definition of spatiotemporal similarity. While distance-based similarity models such as the Gesture Matching Distance on gesture signatures are able to leverage the spatial and temporal characteristics of gestural patterns, their applicability to large 3D motion capture databases is limited due to their high computational complexity. To this end, we present a lower bound approximation of the Gesture Matching Distance that can be utilized in an optimal multi-step query processing architecture in order to provide efficient query processing. We investigate the performance in terms of accuracy and efficiency based on 3D motion capture databases and show that our approach is able to achieve an increase in efficiency of more than one order of magnitude with a negligible loss in accuracy.
{"title":"Efficient Query Processing in 3D Motion Capture Databases via Lower Bound Approximation of the Gesture Matching Distance","authors":"C. Beecks, Marwan Hassani, Florian Obeloer, T. Seidl","doi":"10.1109/ISM.2015.86","DOIUrl":"https://doi.org/10.1109/ISM.2015.86","url":null,"abstract":"One of the most fundamental challenges when accessing gestural patterns in 3D motion capture databases is the definition of spatiotemporal similarity. While distance-based similarity models such as the Gesture Matching Distance on gesture signatures are able to leverage the spatial and temporal characteristics of gestural patterns, their applicability to large 3D motion capture databases is limited due to their high computational complexity. To this end, we present a lower bound approximation of the Gesture Matching Distance that can be utilized in an optimal multi-step query processing architecture in order to provide efficient query processing. We investigate the performance in terms of accuracy and efficiency based on 3D motion capture databases and show that our approach is able to achieve an increase in efficiency of more than one order of magnitude with a negligible loss in accuracy.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131039909","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper proposes a concept of exploring virtual worlds using a pico projector. To demonstrate this concept, a game, called Gone Fishing, was created on top of the infrastructure vuSpot. Gone Fishing is a game where a mobile device projects a virtual seabed over a room. Each user owns an aquarium and fish, flora and decorations can be collected and exchanged. Fish must be pursued to be captured. Pico projectors are becoming popular and more applications are needed. We intend to create applications that use pico projectors and mobile devices to play games and explore spaces using vuSpot, an infrastructure adaptable to a space that uses existing video camera networks to support interactive experiences. With this game we aim at exploring the application of pico projectors in gaming, investigating new forms of interaction and social experiences using mobile devices and pico projectors.
{"title":"Using Pico Projectors to Explore Other Worlds","authors":"Rossana Santos, N. Correia","doi":"10.1109/ISM.2015.59","DOIUrl":"https://doi.org/10.1109/ISM.2015.59","url":null,"abstract":"This paper proposes a concept of exploring virtual worlds using a pico projector. To demonstrate this concept, a game, called Gone Fishing, was created on top of the infrastructure vuSpot. Gone Fishing is a game where a mobile device projects a virtual seabed over a room. Each user owns an aquarium and fish, flora and decorations can be collected and exchanged. Fish must be pursued to be captured. Pico projectors are becoming popular and more applications are needed. We intend to create applications that use pico projectors and mobile devices to play games and explore spaces using vuSpot, an infrastructure adaptable to a space that uses existing video camera networks to support interactive experiences. With this game we aim at exploring the application of pico projectors in gaming, investigating new forms of interaction and social experiences using mobile devices and pico projectors.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121368823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Luiz Olmes Carvalho, Lúcio F. D. Santos, Willian D. Oliveira, A. Traina, C. Traina
Near-duplicate image detection plays an important role in several real applications. Such task is usually achieved by applying a clustering algorithm followed by refinement steps, which is a computationally expensive process. In this paper we introduce a framework based on a novel similarity join operator, which is able both to replace and speed up the clustering step, whereas also releasing the need of further refinement processes. It is based on absolute and relative similarity ratios, ensuring that top ranked image pairs are in the final result. Experiments performed on real datasets shows that our proposal is up to three orders of magnitude faster than the best techniques in the literature, always returning a high-quality result set.
{"title":"Self Similarity Wide-Joins for Near-Duplicate Image Detection","authors":"Luiz Olmes Carvalho, Lúcio F. D. Santos, Willian D. Oliveira, A. Traina, C. Traina","doi":"10.1109/ISM.2015.114","DOIUrl":"https://doi.org/10.1109/ISM.2015.114","url":null,"abstract":"Near-duplicate image detection plays an important role in several real applications. Such task is usually achieved by applying a clustering algorithm followed by refinement steps, which is a computationally expensive process. In this paper we introduce a framework based on a novel similarity join operator, which is able both to replace and speed up the clustering step, whereas also releasing the need of further refinement processes. It is based on absolute and relative similarity ratios, ensuring that top ranked image pairs are in the final result. Experiments performed on real datasets shows that our proposal is up to three orders of magnitude faster than the best techniques in the literature, always returning a high-quality result set.","PeriodicalId":250353,"journal":{"name":"2015 IEEE International Symposium on Multimedia (ISM)","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126042664","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}