Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000376
Hsing Mei, Chun-Yaung Lu, Chi-Chen Lai
Internet Cache Protocol (ICP) is useful for constructing a cooperative cache environment on the Internet. The proxy servers exchange ICP queries and replies in order to share the cached objects with others. However, the required bandwidth of the ICP procedure in the cooperative environment is fixed no matter how high or low the ICP hit rate is. We propose a querying filter mechanism to select proper ICP queries before sending them out. With the querying filter scheme, every proxy server builds a status table to store favorite Web hosts of its neighbors. When the original hosts are requesting, those objects will be delivered freely in order to keep the ICP hit rate stable. Other queries have to be filtered for decreasing additional traffic cost. By doing so the unqualified queries will be filtered and the network bandwidth can be saved; this eventually raises the utilization of the ICP mechanism.
{"title":"An automatic cache cooperative environment using ICP","authors":"Hsing Mei, Chun-Yaung Lu, Chi-Chen Lai","doi":"10.1109/ITCC.2002.1000376","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000376","url":null,"abstract":"Internet Cache Protocol (ICP) is useful for constructing a cooperative cache environment on the Internet. The proxy servers exchange ICP queries and replies in order to share the cached objects with others. However, the required bandwidth of the ICP procedure in the cooperative environment is fixed no matter how high or low the ICP hit rate is. We propose a querying filter mechanism to select proper ICP queries before sending them out. With the querying filter scheme, every proxy server builds a status table to store favorite Web hosts of its neighbors. When the original hosts are requesting, those objects will be delivered freely in order to keep the ICP hit rate stable. Other queries have to be filtered for decreasing additional traffic cost. By doing so the unqualified queries will be filtered and the network bandwidth can be saved; this eventually raises the utilization of the ICP mechanism.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116388904","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000421
Dongyang Long, W. Jia, Ming Li
Novel synchronous coding schemes are introduced and relationships between optimal synchronous codes and Huffman codes are also discussed. Although the problem of existence of optimal synchronous codes has not been resolved yet, we show that any synchronous code can consider as an optimal synchronous code for some information source and that there always exist optimal synchronous codes for the information source with a dyadic probability distribution. Comparing with Huffman coding, the synchronous coding is used not only for statistical modeling but also for dictionary methods. Moreover, it is proven that breaking a synchronous code is NP-complete.
{"title":"On synchronous coding","authors":"Dongyang Long, W. Jia, Ming Li","doi":"10.1109/ITCC.2002.1000421","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000421","url":null,"abstract":"Novel synchronous coding schemes are introduced and relationships between optimal synchronous codes and Huffman codes are also discussed. Although the problem of existence of optimal synchronous codes has not been resolved yet, we show that any synchronous code can consider as an optimal synchronous code for some information source and that there always exist optimal synchronous codes for the information source with a dyadic probability distribution. Comparing with Huffman coding, the synchronous coding is used not only for statistical modeling but also for dictionary methods. Moreover, it is proven that breaking a synchronous code is NP-complete.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125944676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000411
King-Ip Lin, Hui Chen
A large amount of online information resides on the "invisible Web" - Web pages that are generated dynamically from databases and other data sources hidden from the user. They are not indexed by a static URL but are generated when queries are made via a search interface (a specialized search engine). In this paper, we propose a system that is capable of automatically making use of these specialized engines to find information on the invisible Web. We describe our overall architecture and process: from obtaining the search engines to picking the right engines to query. Experiments show that we can find information that is not found by traditional search engines.
{"title":"Automatic information discovery from the \"invisible Web\"","authors":"King-Ip Lin, Hui Chen","doi":"10.1109/ITCC.2002.1000411","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000411","url":null,"abstract":"A large amount of online information resides on the \"invisible Web\" - Web pages that are generated dynamically from databases and other data sources hidden from the user. They are not indexed by a static URL but are generated when queries are made via a search interface (a specialized search engine). In this paper, we propose a system that is capable of automatically making use of these specialized engines to find information on the invisible Web. We describe our overall architecture and process: from obtaining the search engines to picking the right engines to query. Experiments show that we can find information that is not found by traditional search engines.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"07 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125961166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000393
Abdur Chowdhury, S. Beitzel, Eric C. Jensen
We examine the issues of combining multiple query representations in a single IR engine. Differing query representations are used to retrieve different documents. Thus, when combining their results, improvements are observed in effectiveness. We use multiple TREC query representations (title, description and narrative) as a basis for experimentation. We examine several combination approaches presented in the literature (vector addition, CombSUM and CombMNZ) and present a new combination approach using query vector length normalization. We examine two query representation combination approaches (title + description and title + narrative) for 150 queries from TREC 6, 7 and 8 topics. Our QLN (Query Length Normalization) technique outperformed vector addition and data fusion approaches by as much as 32% and was on average 24% better. Additionally, QLN always outperformed the single best query representation in terms of effectiveness.
{"title":"Analysis of combining multiple query representations with varying lengths in a single engine","authors":"Abdur Chowdhury, S. Beitzel, Eric C. Jensen","doi":"10.1109/ITCC.2002.1000393","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000393","url":null,"abstract":"We examine the issues of combining multiple query representations in a single IR engine. Differing query representations are used to retrieve different documents. Thus, when combining their results, improvements are observed in effectiveness. We use multiple TREC query representations (title, description and narrative) as a basis for experimentation. We examine several combination approaches presented in the literature (vector addition, CombSUM and CombMNZ) and present a new combination approach using query vector length normalization. We examine two query representation combination approaches (title + description and title + narrative) for 150 queries from TREC 6, 7 and 8 topics. Our QLN (Query Length Normalization) technique outperformed vector addition and data fusion approaches by as much as 32% and was on average 24% better. Additionally, QLN always outperformed the single best query representation in terms of effectiveness.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132384621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000375
S. Al-Khayatt, S. Shaikh, Babak Akhgar, J. Siddiqi
The concept of virtual private networks offers a simple and cheap alternative to dedicated secure networks in corporate networks and Internet environments. The choice, however, of authentication and encryption techniques and protocols affect issues such as data throughput and performance throughout the network. This paper looks into two virtual private network models within a corporate network. Two operating system platforms are implemented in this investigation; Novell Netware and Windows 2000. Performance of the two arrangements is, subsequently, analyzed and evaluated.
{"title":"A study of encrypted, tunneling models in virtual private networks","authors":"S. Al-Khayatt, S. Shaikh, Babak Akhgar, J. Siddiqi","doi":"10.1109/ITCC.2002.1000375","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000375","url":null,"abstract":"The concept of virtual private networks offers a simple and cheap alternative to dedicated secure networks in corporate networks and Internet environments. The choice, however, of authentication and encryption techniques and protocols affect issues such as data throughput and performance throughout the network. This paper looks into two virtual private network models within a corporate network. Two operating system platforms are implemented in this investigation; Novell Netware and Windows 2000. Performance of the two arrangements is, subsequently, analyzed and evaluated.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134018363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000379
J. Domingo-Ferrer, F. Sebé
Invertible watermarking has been introduced in the literature for the purpose of image authentication. We present a spread-spectrum invertible watermarking system which can be used to authenticate images in any lossless format, i.e. establish their integrity. A second application of invertible watermarking is multilevel access to watermarked images: depending on her clearance, the image user can "clean" the marks of more or less parts of the image, so as to gain in precision. Both applications make sense for precision-critical images (e.g. military, satellite, medical, quality control, reverse engineering images) whose copyright should still be protected to some extent.
{"title":"Invertible spread-spectrum watermarking for image authentication and multilevel access to precision-critical watermarked images","authors":"J. Domingo-Ferrer, F. Sebé","doi":"10.1109/ITCC.2002.1000379","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000379","url":null,"abstract":"Invertible watermarking has been introduced in the literature for the purpose of image authentication. We present a spread-spectrum invertible watermarking system which can be used to authenticate images in any lossless format, i.e. establish their integrity. A second application of invertible watermarking is multilevel access to watermarked images: depending on her clearance, the image user can \"clean\" the marks of more or less parts of the image, so as to gain in precision. Both applications make sense for precision-critical images (e.g. military, satellite, medical, quality control, reverse engineering images) whose copyright should still be protected to some extent.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"12 6","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132236696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000353
Ying Zhao, G. Karypis
The emergence of the World Wide Web has led to an increased interest in methods for searching for information. A key characteristic of many online document collections is that the documents have pre-defined category information, such as the variety of scientific articles accessible via digital libraries (e.g. ACM, IEEE, etc.), medical articles, news-wires and various directories (e.g. Yahoo, OpenDirectory Project, etc.). However, most previous information retrieval systems have not taken the pre-existing category information into account. In this paper, we present weight adjustment schemes based upon the category information in the vector-space model, which are able to select the most content-specific and discriminating features. Our experimental results on TREC data sets show that the pre-existing category information does provide additional beneficial information to improve retrieval. The proposed weight adjustment schemes perform better than the vector-space model with the inverse document frequency (IDF) weighting scheme when queries are less specific. The proposed weighting schemes can also benefit retrieval when clusters are used as an approximations to categories.
{"title":"Improve precategorized collection retrieval by using supervised term weighting schemes","authors":"Ying Zhao, G. Karypis","doi":"10.1109/ITCC.2002.1000353","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000353","url":null,"abstract":"The emergence of the World Wide Web has led to an increased interest in methods for searching for information. A key characteristic of many online document collections is that the documents have pre-defined category information, such as the variety of scientific articles accessible via digital libraries (e.g. ACM, IEEE, etc.), medical articles, news-wires and various directories (e.g. Yahoo, OpenDirectory Project, etc.). However, most previous information retrieval systems have not taken the pre-existing category information into account. In this paper, we present weight adjustment schemes based upon the category information in the vector-space model, which are able to select the most content-specific and discriminating features. Our experimental results on TREC data sets show that the pre-existing category information does provide additional beneficial information to improve retrieval. The proposed weight adjustment schemes perform better than the vector-space model with the inverse document frequency (IDF) weighting scheme when queries are less specific. The proposed weighting schemes can also benefit retrieval when clusters are used as an approximations to categories.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114783596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000410
M. Cannataro, C. Comito, Andrea Pugliese
XML is emerging as the "universal" language for semi-structured data description/exchange, and new issues regarding the management of XML data, both in terms of performance and usability, are becoming critical. The application of knowledge-based synthesis and compression methods (i.e. derivation of synthetic views and lossless/lossy approximation of contents) can be extremely beneficial in this scenario. In this paper, we present SqueezeX, a system for the synthesis/compression of XML data based on a multidimensional/classification-based interpretation of their schema and on the application of semantic and type-oriented compression techniques.
{"title":"SqueezeX: synthesis and compression of XML data","authors":"M. Cannataro, C. Comito, Andrea Pugliese","doi":"10.1109/ITCC.2002.1000410","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000410","url":null,"abstract":"XML is emerging as the \"universal\" language for semi-structured data description/exchange, and new issues regarding the management of XML data, both in terms of performance and usability, are becoming critical. The application of knowledge-based synthesis and compression methods (i.e. derivation of synthetic views and lossless/lossy approximation of contents) can be extremely beneficial in this scenario. In this paper, we present SqueezeX, a system for the synthesis/compression of XML data based on a multidimensional/classification-based interpretation of their schema and on the application of semantic and type-oriented compression techniques.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"117 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116725521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000443
Faïza Abbaci, M. Beigbeder, J. Savoy
Demonstrates that, in an ideal distributed information retrieval environment, it can be effective to take into account the ability of each collection server to return relevant documents when selecting collections. Based on this assumption, we suggest a new approach to resolve the collection selection problem. In order to predict a collection's ability to return relevant documents, we inspect a limited number (n) of documents retrieved from each collection and analyse the proximity of search keywords within them. In our experiments, we vary the underlying parameter n of our suggested model in order to define the most appropriate number of top documents to be inspected. Moreover, we evaluate the retrieval effectiveness of our approach and compare it with both the centralized indexing and the CORI (COllection Retrieval Inference) approaches. Preliminary results from these experiments, conducted on the WT10g test collection of Web pages, tend to demonstrate that our suggested method can achieve appreciable retrieval effectiveness.
{"title":"A methodology for collection selection in heterogeneous contexts","authors":"Faïza Abbaci, M. Beigbeder, J. Savoy","doi":"10.1109/ITCC.2002.1000443","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000443","url":null,"abstract":"Demonstrates that, in an ideal distributed information retrieval environment, it can be effective to take into account the ability of each collection server to return relevant documents when selecting collections. Based on this assumption, we suggest a new approach to resolve the collection selection problem. In order to predict a collection's ability to return relevant documents, we inspect a limited number (n) of documents retrieved from each collection and analyse the proximity of search keywords within them. In our experiments, we vary the underlying parameter n of our suggested model in order to define the most appropriate number of top documents to be inspected. Moreover, we evaluate the retrieval effectiveness of our approach and compare it with both the centralized indexing and the CORI (COllection Retrieval Inference) approaches. Preliminary results from these experiments, conducted on the WT10g test collection of Web pages, tend to demonstrate that our suggested method can achieve appreciable retrieval effectiveness.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"316 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116771116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000368
M. Steinebach, A. Lang, J. Dittmann, Chr. Neubauer
Audio watermarking has become an important technology for the recording and advertisement industry. Today's embedding and detection strategies often rely on digital/ high quality channels, like CD or mp3. However for numerous applications, watermarks surviving noisy analog environments are better suited or even required. Using such schemes allows to identify, audio signals with monitoring devices which are 'listening' to the incoming signals. In the same way, illegal recordings of concerts can be identified. In this paper we describe a test environment for noisy, acoustic, (analog) channels and present results obtained from a particular watermarking scheme. We identify changes in the affected audio material, like frequency response or the amount of noise added to the signal. In order to do automatic robustness evaluations of watermarking schemes, the Stirmark Benchmark environment can be used. We analyze if it is possible to simulate an acoustic, noisy DA/AD environment with filters, quantization and noise generators. Based on the presented test results, we identify the parameters relevant for watermarks to successfully survive noisy acoustic channels and thereby provide valuable hints for audio watermark designers. We also describe a design concept for a DA/AD simulation.
{"title":"Audio watermarking quality evaluation: robustness to DA/AD processes","authors":"M. Steinebach, A. Lang, J. Dittmann, Chr. Neubauer","doi":"10.1109/ITCC.2002.1000368","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000368","url":null,"abstract":"Audio watermarking has become an important technology for the recording and advertisement industry. Today's embedding and detection strategies often rely on digital/ high quality channels, like CD or mp3. However for numerous applications, watermarks surviving noisy analog environments are better suited or even required. Using such schemes allows to identify, audio signals with monitoring devices which are 'listening' to the incoming signals. In the same way, illegal recordings of concerts can be identified. In this paper we describe a test environment for noisy, acoustic, (analog) channels and present results obtained from a particular watermarking scheme. We identify changes in the affected audio material, like frequency response or the amount of noise added to the signal. In order to do automatic robustness evaluations of watermarking schemes, the Stirmark Benchmark environment can be used. We analyze if it is possible to simulate an acoustic, noisy DA/AD environment with filters, quantization and noise generators. Based on the presented test results, we identify the parameters relevant for watermarks to successfully survive noisy acoustic channels and thereby provide valuable hints for audio watermark designers. We also describe a design concept for a DA/AD simulation.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132850911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}