N. Asokan, Seamus Moloney, Philip Ginzboorg, Kari Kostiainen
The increasing popularity and variety of consumer multimedia devices is driving the need for networked homes. Yet setting up a secure wireless network is a daunting task for most ordinary users. Recently, there have been several proposals for easing this process. However, none of the proposals consider the problem of how to make it easy to manage visitor access. In this paper, we motivate the requirements for visitor management, show the shortcomings of the current easy setup proposals in this regard, and propose a new setup procedure that makes it easy to manage visitor access to wireless networks. Our contributions are twofold: first we present an approach to assigning categories to client devices at admission time so that selective revocation of clients based on those categories becomes possible. Then we present the idea of admission tickets, a flexible and secure way to delegate conditional access rights. We report the results and experience of prototyping of the proposed procedure using the HostAP framework.
{"title":"Visitor access management in personal wireless networks","authors":"N. Asokan, Seamus Moloney, Philip Ginzboorg, Kari Kostiainen","doi":"10.1109/ISM.2005.122","DOIUrl":"https://doi.org/10.1109/ISM.2005.122","url":null,"abstract":"The increasing popularity and variety of consumer multimedia devices is driving the need for networked homes. Yet setting up a secure wireless network is a daunting task for most ordinary users. Recently, there have been several proposals for easing this process. However, none of the proposals consider the problem of how to make it easy to manage visitor access. In this paper, we motivate the requirements for visitor management, show the shortcomings of the current easy setup proposals in this regard, and propose a new setup procedure that makes it easy to manage visitor access to wireless networks. Our contributions are twofold: first we present an approach to assigning categories to client devices at admission time so that selective revocation of clients based on those categories becomes possible. Then we present the idea of admission tickets, a flexible and secure way to delegate conditional access rights. We report the results and experience of prototyping of the proposed procedure using the HostAP framework.","PeriodicalId":322363,"journal":{"name":"Seventh IEEE International Symposium on Multimedia (ISM'05)","volume":"277 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126017234","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Nowadays, the amount of multimedia data is increasing rapidly, and hence, there is an increasing need for efficient methods to manage the multimedia content. This paper proposes a framework for the description and retrieval of multimedia data. The data are represented at both the syntactic (structure, metadata and low level features) and semantic (the meaning of the data) levels. We use the MPEG-7 standard, which provides a set of tools to describe multimedia content from different viewpoints, to represent the syntactic level. However, due to its XML schema based representation, MPEG-7 is not suitable to represent the semantic aspect of the data in a formal and concise way. Moreover, inferential mechanisms are not provided. To alleviate these limitations, we propose to extend MPEG-7 with a domain ontology, formalized using a logical formalism. Then, the semantic aspect of the data is described using the ontology's vocabulary, as a set of logical expressions. We enhance the ontology by a rules layer, to describe more complex constraints between domain concepts and relations. User's queries may concern the syntactic and/or semantic features. The syntactic constraints are expressed using XQuery language and evaluated using an XML query engine; whereas the semantic query constraints are expressed using a rules language and evaluated using a specific resolution mechanism.
{"title":"A logic based approach for the multimedia data representation and retrieval","authors":"Samira Hammiche, S. Benbernou, A. Vakali","doi":"10.1109/ISM.2005.11","DOIUrl":"https://doi.org/10.1109/ISM.2005.11","url":null,"abstract":"Nowadays, the amount of multimedia data is increasing rapidly, and hence, there is an increasing need for efficient methods to manage the multimedia content. This paper proposes a framework for the description and retrieval of multimedia data. The data are represented at both the syntactic (structure, metadata and low level features) and semantic (the meaning of the data) levels. We use the MPEG-7 standard, which provides a set of tools to describe multimedia content from different viewpoints, to represent the syntactic level. However, due to its XML schema based representation, MPEG-7 is not suitable to represent the semantic aspect of the data in a formal and concise way. Moreover, inferential mechanisms are not provided. To alleviate these limitations, we propose to extend MPEG-7 with a domain ontology, formalized using a logical formalism. Then, the semantic aspect of the data is described using the ontology's vocabulary, as a set of logical expressions. We enhance the ontology by a rules layer, to describe more complex constraints between domain concepts and relations. User's queries may concern the syntactic and/or semantic features. The syntactic constraints are expressed using XQuery language and evaluated using an XML query engine; whereas the semantic query constraints are expressed using a rules language and evaluated using a specific resolution mechanism.","PeriodicalId":322363,"journal":{"name":"Seventh IEEE International Symposium on Multimedia (ISM'05)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114504413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In this paper, we present a novel audio watermarking technique that utilizes the low frequency components (LFCs) of an audio signal to identify the location of the embedded watermarks. The embedding takes place by modifying the amplitude of selected samples determined by the LFCs of the audio signal. The amount of modification to the amplitude is determined by the amount of distortion detected by the human ear. This technique is blind where the decoder does not need the original audio file to extract the watermarks. In this technique, we use a novel data recovery scheme to recover any watermarks that were lost because of an intentional or unintentional attempt of watermark removal (attack). Experimental results show that this technique is highly robust against single and double attacks with watermark recovery rates greater than 90%.
{"title":"A novel audio watermarking technique based on low frequency components","authors":"Hamad Alaryani, Abdou Youssef","doi":"10.1109/ISM.2005.16","DOIUrl":"https://doi.org/10.1109/ISM.2005.16","url":null,"abstract":"In this paper, we present a novel audio watermarking technique that utilizes the low frequency components (LFCs) of an audio signal to identify the location of the embedded watermarks. The embedding takes place by modifying the amplitude of selected samples determined by the LFCs of the audio signal. The amount of modification to the amplitude is determined by the amount of distortion detected by the human ear. This technique is blind where the decoder does not need the original audio file to extract the watermarks. In this technique, we use a novel data recovery scheme to recover any watermarks that were lost because of an intentional or unintentional attempt of watermark removal (attack). Experimental results show that this technique is highly robust against single and double attacks with watermark recovery rates greater than 90%.","PeriodicalId":322363,"journal":{"name":"Seventh IEEE International Symposium on Multimedia (ISM'05)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125411027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper presents the current state of the art in Web3d multi-user technologies and more specifically networked virtual environment platforms that support the X3D standard, which are currently limited in number, and also the existing X3D browsers that enable these applications. Furthermore, this paper presents the architecture of a platform, called EVE, which was developed to support multi-user virtual environments and was initially based on VRML, as well as the implementation regarding the process of upgrading this platform in order to fully support the X3D standard, using the open-source Xj3D library. Finally, the benefits of shifting to the X3D standard are presented, and some of the problems that currently exist are discussed.
{"title":"Advances in X3D multi-user virtual environments","authors":"C. Bouras, Alexandros Panagopoulos, T. Tsiatsos","doi":"10.1109/ISM.2005.28","DOIUrl":"https://doi.org/10.1109/ISM.2005.28","url":null,"abstract":"This paper presents the current state of the art in Web3d multi-user technologies and more specifically networked virtual environment platforms that support the X3D standard, which are currently limited in number, and also the existing X3D browsers that enable these applications. Furthermore, this paper presents the architecture of a platform, called EVE, which was developed to support multi-user virtual environments and was initially based on VRML, as well as the implementation regarding the process of upgrading this platform in order to fully support the X3D standard, using the open-source Xj3D library. Finally, the benefits of shifting to the X3D standard are presented, and some of the problems that currently exist are discussed.","PeriodicalId":322363,"journal":{"name":"Seventh IEEE International Symposium on Multimedia (ISM'05)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116104071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The development of highly interactive multimedia applications is still a challenging and complex task. In addition to the application logic multimedia applications typically provide a sophisticated user interface with integrated media objects. As a consequence, the development process involves different experts for software design, user interface design, and media design. There is still a lack of concepts for a structured development process to integrate these requirements. In this paper we introduce the Multimedia Modeling Language (MML), a visual modeling language supporting the design process in multimedia application development. It is part of a model-driven development approach for multimedia applications. The language is oriented on well-established software engineering concepts, in particular UML 2.0. It integrates the results of two different research lines: application-oriented multimedia modeling and model-based user interface development. In this paper we describe the concepts of the language and present the modeling process with MML. In particular we show how MML aims to integrate the different developer roles in multimedia application design.
{"title":"MML: a language for modeling interactive multimedia applications","authors":"Andreas Pleuß","doi":"10.1109/ISM.2005.80","DOIUrl":"https://doi.org/10.1109/ISM.2005.80","url":null,"abstract":"The development of highly interactive multimedia applications is still a challenging and complex task. In addition to the application logic multimedia applications typically provide a sophisticated user interface with integrated media objects. As a consequence, the development process involves different experts for software design, user interface design, and media design. There is still a lack of concepts for a structured development process to integrate these requirements. In this paper we introduce the Multimedia Modeling Language (MML), a visual modeling language supporting the design process in multimedia application development. It is part of a model-driven development approach for multimedia applications. The language is oriented on well-established software engineering concepts, in particular UML 2.0. It integrates the results of two different research lines: application-oriented multimedia modeling and model-based user interface development. In this paper we describe the concepts of the language and present the modeling process with MML. In particular we show how MML aims to integrate the different developer roles in multimedia application design.","PeriodicalId":322363,"journal":{"name":"Seventh IEEE International Symposium on Multimedia (ISM'05)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127019150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
As technology trends yield shorter cycle times and larger, wider datapaths in architectures for multimedia systems, global broadcast networks for operand communication are becoming a major bottleneck in processor performance. New low latency operand transport techniques are needed. This paper proposes and evaluates lower cost mechanisms than traditional bypass networks, exploiting regular operand distribution patterns in multimedia applications. To reduce latency associated with operand movement within a datapath, our mechanism, called dynamic instruction clustering, groups chains of dependent instructions within a basic block at runtime, identifies intermediate value transportation, and schedules it on networked ALUs which are connected by a local dedicated network. By converting global communication into local, the transport latency can be minimized and the critical path of the application code can be executed in consecutive, shortened cycles, resulting in improved performance. We demonstrated that 28% and 30% of total dependence edges residing in the instruction window can be localized on 8 and 16-way machines, respectively. Our results show that the overall performance gains over a wide range of multimedia applications are 16% for 8-way and 35% for 16-way on average.
{"title":"Reducing operand communication overhead using instruction clustering for multimedia applications","authors":"Hongkyu Kim, D. S. Wills, L. Wills","doi":"10.1109/ISM.2005.95","DOIUrl":"https://doi.org/10.1109/ISM.2005.95","url":null,"abstract":"As technology trends yield shorter cycle times and larger, wider datapaths in architectures for multimedia systems, global broadcast networks for operand communication are becoming a major bottleneck in processor performance. New low latency operand transport techniques are needed. This paper proposes and evaluates lower cost mechanisms than traditional bypass networks, exploiting regular operand distribution patterns in multimedia applications. To reduce latency associated with operand movement within a datapath, our mechanism, called dynamic instruction clustering, groups chains of dependent instructions within a basic block at runtime, identifies intermediate value transportation, and schedules it on networked ALUs which are connected by a local dedicated network. By converting global communication into local, the transport latency can be minimized and the critical path of the application code can be executed in consecutive, shortened cycles, resulting in improved performance. We demonstrated that 28% and 30% of total dependence edges residing in the instruction window can be localized on 8 and 16-way machines, respectively. Our results show that the overall performance gains over a wide range of multimedia applications are 16% for 8-way and 35% for 16-way on average.","PeriodicalId":322363,"journal":{"name":"Seventh IEEE International Symposium on Multimedia (ISM'05)","volume":"2015 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127640624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Zhaoyu Liu, Dichao Peng, Yuliang Zheng, Jeffrey Liu
For decades, video surveillance systems have been crucial in protecting critical infrastructures. Currently the systems are adopting the emerging internetworking digital technology, changing from the traditional isolated analog technology. While there are several advantages of IP-based video surveillance systems, the systems are also exposed to additional new security threats. Security is a critical factor to make the IP-based video surveillance systems practical and usable. In this paper, we address the security threats to the data communications in IP-based video surveillance systems. We first analyze the current approaches, mainly naive and selective methods, to secure real-time video data, and identify their limitations to IP-based video surveillance systems. Then we propose a system design of secure internetworking video surveillance systems. Our proposed approach adopts a randomly adjustable encryption method to provide confidentiality to the data. Our implementation and performance study demonstrate that our approach achieves a good balance of cost-efficiency and security.
{"title":"Communication protection in IP-based video surveillance systems","authors":"Zhaoyu Liu, Dichao Peng, Yuliang Zheng, Jeffrey Liu","doi":"10.1109/ISM.2005.42","DOIUrl":"https://doi.org/10.1109/ISM.2005.42","url":null,"abstract":"For decades, video surveillance systems have been crucial in protecting critical infrastructures. Currently the systems are adopting the emerging internetworking digital technology, changing from the traditional isolated analog technology. While there are several advantages of IP-based video surveillance systems, the systems are also exposed to additional new security threats. Security is a critical factor to make the IP-based video surveillance systems practical and usable. In this paper, we address the security threats to the data communications in IP-based video surveillance systems. We first analyze the current approaches, mainly naive and selective methods, to secure real-time video data, and identify their limitations to IP-based video surveillance systems. Then we propose a system design of secure internetworking video surveillance systems. Our proposed approach adopts a randomly adjustable encryption method to provide confidentiality to the data. Our implementation and performance study demonstrate that our approach achieves a good balance of cost-efficiency and security.","PeriodicalId":322363,"journal":{"name":"Seventh IEEE International Symposium on Multimedia (ISM'05)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126722048","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
R. V. Prasad, H. N. Shankar, P. Pawełczak, H. Jamadagni
For efficient computer supported cooperative work (CSCW) audio conferencing is an essential component where video and text are add-ons. The specifications for enabling CSCW over Internet are incomplete if they are blind to actual conduct of participants. Indeed, a blind conference mimics quite closely a virtual voice-only conference. In this paper, we analyze the results of sessions of face-to-face blind conversations and gain penetrating insights. In particular, we focus on the impact of users' behavior on the design of a scalable architecture for virtual voice-only conferencing over VoIP and arrive at a meaningful number of floors for such conferences. We also present the features and the requirements for the proposed service.
{"title":"bFixing number of floors for virtual voice-only conference - an empirical study","authors":"R. V. Prasad, H. N. Shankar, P. Pawełczak, H. Jamadagni","doi":"10.1109/ISM.2005.59","DOIUrl":"https://doi.org/10.1109/ISM.2005.59","url":null,"abstract":"For efficient computer supported cooperative work (CSCW) audio conferencing is an essential component where video and text are add-ons. The specifications for enabling CSCW over Internet are incomplete if they are blind to actual conduct of participants. Indeed, a blind conference mimics quite closely a virtual voice-only conference. In this paper, we analyze the results of sessions of face-to-face blind conversations and gain penetrating insights. In particular, we focus on the impact of users' behavior on the design of a scalable architecture for virtual voice-only conferencing over VoIP and arrive at a meaningful number of floors for such conferences. We also present the features and the requirements for the proposed service.","PeriodicalId":322363,"journal":{"name":"Seventh IEEE International Symposium on Multimedia (ISM'05)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121389918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
While mobile technology precludes large electronic displays for visual immersion, sound heard over headphones - a widely accepted technology - is ideally suited for mobile applications. In this paper, we report on a newly developed immersive headphone-based approach that opens new opportunities for mobile multimedia. This new motion-tracked binaural sound technology (abbreviated as MTB) provides and exploits the strong perceptual cues that are created by the voluntary motions of the listener's head. A head-tracker is used to modify dynamically the sound heard by the listener, recreating the natural experience. After presenting an overview of the key concepts and of the challenges of the implementation of MTB, we describe examples of MTB spatial sound applications. Finally, we outline new mobile multimedia applications that would combine immersive spatial sound, head-tracking, and small visual displays.
{"title":"Immersive spatial sound for mobile multimedia","authors":"V. Algazi, R. Duda","doi":"10.1109/ISM.2005.69","DOIUrl":"https://doi.org/10.1109/ISM.2005.69","url":null,"abstract":"While mobile technology precludes large electronic displays for visual immersion, sound heard over headphones - a widely accepted technology - is ideally suited for mobile applications. In this paper, we report on a newly developed immersive headphone-based approach that opens new opportunities for mobile multimedia. This new motion-tracked binaural sound technology (abbreviated as MTB) provides and exploits the strong perceptual cues that are created by the voluntary motions of the listener's head. A head-tracker is used to modify dynamically the sound heard by the listener, recreating the natural experience. After presenting an overview of the key concepts and of the challenges of the implementation of MTB, we describe examples of MTB spatial sound applications. Finally, we outline new mobile multimedia applications that would combine immersive spatial sound, head-tracking, and small visual displays.","PeriodicalId":322363,"journal":{"name":"Seventh IEEE International Symposium on Multimedia (ISM'05)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122713091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In this paper, we present our work on a ubiquitous context-sensitive multimodal multimedia computing system that progressively acquires machine knowledge. This ubiquitous computing system supports an automatic selection of media and modalities deemed appropriate for the user's context and user's profile. The ability of the system to do so constitutes its acquired knowledge. The decision making for media/modality selection takes into account if the user has some special needs due to disability. The architecture of the system is designed to be pervasive and is conceived to resist failure. In case of one or more components being missing or found defective, the machine would resist failure by reconfiguring itself dynamically in the architectural level. It finds alternative replacement to the failed component using its acquired knowledge.
{"title":"A ubiquitous context-sensitive multimodal multimedia computing system and its machine learning-based reconfiguration at the architectural level","authors":"M. D. Hina, A. Ramdane-Cherif, C. Tadj","doi":"10.1109/ISM.2005.24","DOIUrl":"https://doi.org/10.1109/ISM.2005.24","url":null,"abstract":"In this paper, we present our work on a ubiquitous context-sensitive multimodal multimedia computing system that progressively acquires machine knowledge. This ubiquitous computing system supports an automatic selection of media and modalities deemed appropriate for the user's context and user's profile. The ability of the system to do so constitutes its acquired knowledge. The decision making for media/modality selection takes into account if the user has some special needs due to disability. The architecture of the system is designed to be pervasive and is conceived to resist failure. In case of one or more components being missing or found defective, the machine would resist failure by reconfiguring itself dynamically in the architectural level. It finds alternative replacement to the failed component using its acquired knowledge.","PeriodicalId":322363,"journal":{"name":"Seventh IEEE International Symposium on Multimedia (ISM'05)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128240054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}