Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000421
Dongyang Long, W. Jia, Ming Li
Novel synchronous coding schemes are introduced and relationships between optimal synchronous codes and Huffman codes are also discussed. Although the problem of existence of optimal synchronous codes has not been resolved yet, we show that any synchronous code can consider as an optimal synchronous code for some information source and that there always exist optimal synchronous codes for the information source with a dyadic probability distribution. Comparing with Huffman coding, the synchronous coding is used not only for statistical modeling but also for dictionary methods. Moreover, it is proven that breaking a synchronous code is NP-complete.
{"title":"On synchronous coding","authors":"Dongyang Long, W. Jia, Ming Li","doi":"10.1109/ITCC.2002.1000421","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000421","url":null,"abstract":"Novel synchronous coding schemes are introduced and relationships between optimal synchronous codes and Huffman codes are also discussed. Although the problem of existence of optimal synchronous codes has not been resolved yet, we show that any synchronous code can consider as an optimal synchronous code for some information source and that there always exist optimal synchronous codes for the information source with a dyadic probability distribution. Comparing with Huffman coding, the synchronous coding is used not only for statistical modeling but also for dictionary methods. Moreover, it is proven that breaking a synchronous code is NP-complete.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125944676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000411
King-Ip Lin, Hui Chen
A large amount of online information resides on the "invisible Web" - Web pages that are generated dynamically from databases and other data sources hidden from the user. They are not indexed by a static URL but are generated when queries are made via a search interface (a specialized search engine). In this paper, we propose a system that is capable of automatically making use of these specialized engines to find information on the invisible Web. We describe our overall architecture and process: from obtaining the search engines to picking the right engines to query. Experiments show that we can find information that is not found by traditional search engines.
{"title":"Automatic information discovery from the \"invisible Web\"","authors":"King-Ip Lin, Hui Chen","doi":"10.1109/ITCC.2002.1000411","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000411","url":null,"abstract":"A large amount of online information resides on the \"invisible Web\" - Web pages that are generated dynamically from databases and other data sources hidden from the user. They are not indexed by a static URL but are generated when queries are made via a search interface (a specialized search engine). In this paper, we propose a system that is capable of automatically making use of these specialized engines to find information on the invisible Web. We describe our overall architecture and process: from obtaining the search engines to picking the right engines to query. Experiments show that we can find information that is not found by traditional search engines.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"07 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125961166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000414
Ling Wang, H. Selvaraj
In this paper, the problem of performance driven circuit partitioning is considered. The parameters taken into consideration to measure performance are power interconnection resource constraints. An algorithm is presented to build clusters in a bottom up manner while decomposing clusters for cutsize and delay minimization as well as power consumption and resource constraint. A partitioning method in a top down manner is applied based on the probability function.
{"title":"Performance driven circuit clustering and partitioning","authors":"Ling Wang, H. Selvaraj","doi":"10.1109/ITCC.2002.1000414","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000414","url":null,"abstract":"In this paper, the problem of performance driven circuit partitioning is considered. The parameters taken into consideration to measure performance are power interconnection resource constraints. An algorithm is presented to build clusters in a bottom up manner while decomposing clusters for cutsize and delay minimization as well as power consumption and resource constraint. A partitioning method in a top down manner is applied based on the probability function.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114491716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000375
S. Al-Khayatt, S. Shaikh, Babak Akhgar, J. Siddiqi
The concept of virtual private networks offers a simple and cheap alternative to dedicated secure networks in corporate networks and Internet environments. The choice, however, of authentication and encryption techniques and protocols affect issues such as data throughput and performance throughout the network. This paper looks into two virtual private network models within a corporate network. Two operating system platforms are implemented in this investigation; Novell Netware and Windows 2000. Performance of the two arrangements is, subsequently, analyzed and evaluated.
{"title":"A study of encrypted, tunneling models in virtual private networks","authors":"S. Al-Khayatt, S. Shaikh, Babak Akhgar, J. Siddiqi","doi":"10.1109/ITCC.2002.1000375","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000375","url":null,"abstract":"The concept of virtual private networks offers a simple and cheap alternative to dedicated secure networks in corporate networks and Internet environments. The choice, however, of authentication and encryption techniques and protocols affect issues such as data throughput and performance throughout the network. This paper looks into two virtual private network models within a corporate network. Two operating system platforms are implemented in this investigation; Novell Netware and Windows 2000. Performance of the two arrangements is, subsequently, analyzed and evaluated.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134018363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000379
J. Domingo-Ferrer, F. Sebé
Invertible watermarking has been introduced in the literature for the purpose of image authentication. We present a spread-spectrum invertible watermarking system which can be used to authenticate images in any lossless format, i.e. establish their integrity. A second application of invertible watermarking is multilevel access to watermarked images: depending on her clearance, the image user can "clean" the marks of more or less parts of the image, so as to gain in precision. Both applications make sense for precision-critical images (e.g. military, satellite, medical, quality control, reverse engineering images) whose copyright should still be protected to some extent.
{"title":"Invertible spread-spectrum watermarking for image authentication and multilevel access to precision-critical watermarked images","authors":"J. Domingo-Ferrer, F. Sebé","doi":"10.1109/ITCC.2002.1000379","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000379","url":null,"abstract":"Invertible watermarking has been introduced in the literature for the purpose of image authentication. We present a spread-spectrum invertible watermarking system which can be used to authenticate images in any lossless format, i.e. establish their integrity. A second application of invertible watermarking is multilevel access to watermarked images: depending on her clearance, the image user can \"clean\" the marks of more or less parts of the image, so as to gain in precision. Both applications make sense for precision-critical images (e.g. military, satellite, medical, quality control, reverse engineering images) whose copyright should still be protected to some extent.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"12 6","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132236696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000393
Abdur Chowdhury, S. Beitzel, Eric C. Jensen
We examine the issues of combining multiple query representations in a single IR engine. Differing query representations are used to retrieve different documents. Thus, when combining their results, improvements are observed in effectiveness. We use multiple TREC query representations (title, description and narrative) as a basis for experimentation. We examine several combination approaches presented in the literature (vector addition, CombSUM and CombMNZ) and present a new combination approach using query vector length normalization. We examine two query representation combination approaches (title + description and title + narrative) for 150 queries from TREC 6, 7 and 8 topics. Our QLN (Query Length Normalization) technique outperformed vector addition and data fusion approaches by as much as 32% and was on average 24% better. Additionally, QLN always outperformed the single best query representation in terms of effectiveness.
{"title":"Analysis of combining multiple query representations with varying lengths in a single engine","authors":"Abdur Chowdhury, S. Beitzel, Eric C. Jensen","doi":"10.1109/ITCC.2002.1000393","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000393","url":null,"abstract":"We examine the issues of combining multiple query representations in a single IR engine. Differing query representations are used to retrieve different documents. Thus, when combining their results, improvements are observed in effectiveness. We use multiple TREC query representations (title, description and narrative) as a basis for experimentation. We examine several combination approaches presented in the literature (vector addition, CombSUM and CombMNZ) and present a new combination approach using query vector length normalization. We examine two query representation combination approaches (title + description and title + narrative) for 150 queries from TREC 6, 7 and 8 topics. Our QLN (Query Length Normalization) technique outperformed vector addition and data fusion approaches by as much as 32% and was on average 24% better. Additionally, QLN always outperformed the single best query representation in terms of effectiveness.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132384621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000410
M. Cannataro, C. Comito, Andrea Pugliese
XML is emerging as the "universal" language for semi-structured data description/exchange, and new issues regarding the management of XML data, both in terms of performance and usability, are becoming critical. The application of knowledge-based synthesis and compression methods (i.e. derivation of synthetic views and lossless/lossy approximation of contents) can be extremely beneficial in this scenario. In this paper, we present SqueezeX, a system for the synthesis/compression of XML data based on a multidimensional/classification-based interpretation of their schema and on the application of semantic and type-oriented compression techniques.
{"title":"SqueezeX: synthesis and compression of XML data","authors":"M. Cannataro, C. Comito, Andrea Pugliese","doi":"10.1109/ITCC.2002.1000410","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000410","url":null,"abstract":"XML is emerging as the \"universal\" language for semi-structured data description/exchange, and new issues regarding the management of XML data, both in terms of performance and usability, are becoming critical. The application of knowledge-based synthesis and compression methods (i.e. derivation of synthetic views and lossless/lossy approximation of contents) can be extremely beneficial in this scenario. In this paper, we present SqueezeX, a system for the synthesis/compression of XML data based on a multidimensional/classification-based interpretation of their schema and on the application of semantic and type-oriented compression techniques.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"117 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116725521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000443
Faïza Abbaci, M. Beigbeder, J. Savoy
Demonstrates that, in an ideal distributed information retrieval environment, it can be effective to take into account the ability of each collection server to return relevant documents when selecting collections. Based on this assumption, we suggest a new approach to resolve the collection selection problem. In order to predict a collection's ability to return relevant documents, we inspect a limited number (n) of documents retrieved from each collection and analyse the proximity of search keywords within them. In our experiments, we vary the underlying parameter n of our suggested model in order to define the most appropriate number of top documents to be inspected. Moreover, we evaluate the retrieval effectiveness of our approach and compare it with both the centralized indexing and the CORI (COllection Retrieval Inference) approaches. Preliminary results from these experiments, conducted on the WT10g test collection of Web pages, tend to demonstrate that our suggested method can achieve appreciable retrieval effectiveness.
{"title":"A methodology for collection selection in heterogeneous contexts","authors":"Faïza Abbaci, M. Beigbeder, J. Savoy","doi":"10.1109/ITCC.2002.1000443","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000443","url":null,"abstract":"Demonstrates that, in an ideal distributed information retrieval environment, it can be effective to take into account the ability of each collection server to return relevant documents when selecting collections. Based on this assumption, we suggest a new approach to resolve the collection selection problem. In order to predict a collection's ability to return relevant documents, we inspect a limited number (n) of documents retrieved from each collection and analyse the proximity of search keywords within them. In our experiments, we vary the underlying parameter n of our suggested model in order to define the most appropriate number of top documents to be inspected. Moreover, we evaluate the retrieval effectiveness of our approach and compare it with both the centralized indexing and the CORI (COllection Retrieval Inference) approaches. Preliminary results from these experiments, conducted on the WT10g test collection of Web pages, tend to demonstrate that our suggested method can achieve appreciable retrieval effectiveness.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"316 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116771116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000353
Ying Zhao, G. Karypis
The emergence of the World Wide Web has led to an increased interest in methods for searching for information. A key characteristic of many online document collections is that the documents have pre-defined category information, such as the variety of scientific articles accessible via digital libraries (e.g. ACM, IEEE, etc.), medical articles, news-wires and various directories (e.g. Yahoo, OpenDirectory Project, etc.). However, most previous information retrieval systems have not taken the pre-existing category information into account. In this paper, we present weight adjustment schemes based upon the category information in the vector-space model, which are able to select the most content-specific and discriminating features. Our experimental results on TREC data sets show that the pre-existing category information does provide additional beneficial information to improve retrieval. The proposed weight adjustment schemes perform better than the vector-space model with the inverse document frequency (IDF) weighting scheme when queries are less specific. The proposed weighting schemes can also benefit retrieval when clusters are used as an approximations to categories.
{"title":"Improve precategorized collection retrieval by using supervised term weighting schemes","authors":"Ying Zhao, G. Karypis","doi":"10.1109/ITCC.2002.1000353","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000353","url":null,"abstract":"The emergence of the World Wide Web has led to an increased interest in methods for searching for information. A key characteristic of many online document collections is that the documents have pre-defined category information, such as the variety of scientific articles accessible via digital libraries (e.g. ACM, IEEE, etc.), medical articles, news-wires and various directories (e.g. Yahoo, OpenDirectory Project, etc.). However, most previous information retrieval systems have not taken the pre-existing category information into account. In this paper, we present weight adjustment schemes based upon the category information in the vector-space model, which are able to select the most content-specific and discriminating features. Our experimental results on TREC data sets show that the pre-existing category information does provide additional beneficial information to improve retrieval. The proposed weight adjustment schemes perform better than the vector-space model with the inverse document frequency (IDF) weighting scheme when queries are less specific. The proposed weighting schemes can also benefit retrieval when clusters are used as an approximations to categories.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114783596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2002-04-08DOI: 10.1109/ITCC.2002.1000351
Mohammed Aljlayl, O. Frieder, D. Grossman
Machine translation (MT) is an automatic process that translates from one human language to another by using context information. We empirically evaluate the use of an MT-based approach for query translation in an Arabic-English cross-language information retrieval (CLIR) system, called ALKAFI, using the TREC-7 and TREC-9 topics and collections. The effect of the query length on the MT performance is also investigated in order to explore how much context is actually required for successful MT processing.
{"title":"On Arabic-English cross-language information retrieval: a machine translation approach","authors":"Mohammed Aljlayl, O. Frieder, D. Grossman","doi":"10.1109/ITCC.2002.1000351","DOIUrl":"https://doi.org/10.1109/ITCC.2002.1000351","url":null,"abstract":"Machine translation (MT) is an automatic process that translates from one human language to another by using context information. We empirically evaluate the use of an MT-based approach for query translation in an Arabic-English cross-language information retrieval (CLIR) system, called ALKAFI, using the TREC-7 and TREC-9 topics and collections. The effect of the query length on the MT performance is also investigated in order to explore how much context is actually required for successful MT processing.","PeriodicalId":115190,"journal":{"name":"Proceedings. International Conference on Information Technology: Coding and Computing","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124380828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}