Pub Date : 2010-08-23DOI: 10.1109/ICCCE.2010.5556753
D. Abdulwahab, Sheroz Khan, J. Chebil, M. Ahmed, A. Naji, Al-Kaissi, A. Alam
Transducers have found wide applications in industries, though, they possess nonlinear responses. The nonlinearity feature does not prevail over the entire range; rather there are sections (or regions) on the response curves that are easily regarded as linear. This work focuses on analyzing transducer characteristics by finding regions exhibiting larger level of linearity. The point of inflection is a joint between two consecutive linear segments, and it is where changes in the output signal in response to external parameter variations remain at minimum compared to its immediate neighboring points. It is shown through simulations on how, assuming that a given system does not operate over the entire range of the response curve, a linear response is assured. This is done by translating the circuit operating over the linear section by employing additional external circuit elements. This paper attempts at analyzing a transducer characteristics by showing regions around the points of inflections where variations remain most linear.
{"title":"Identification of linearized regions of non-linear transducers responses","authors":"D. Abdulwahab, Sheroz Khan, J. Chebil, M. Ahmed, A. Naji, Al-Kaissi, A. Alam","doi":"10.1109/ICCCE.2010.5556753","DOIUrl":"https://doi.org/10.1109/ICCCE.2010.5556753","url":null,"abstract":"Transducers have found wide applications in industries, though, they possess nonlinear responses. The nonlinearity feature does not prevail over the entire range; rather there are sections (or regions) on the response curves that are easily regarded as linear. This work focuses on analyzing transducer characteristics by finding regions exhibiting larger level of linearity. The point of inflection is a joint between two consecutive linear segments, and it is where changes in the output signal in response to external parameter variations remain at minimum compared to its immediate neighboring points. It is shown through simulations on how, assuming that a given system does not operate over the entire range of the response curve, a linear response is assured. This is done by translating the circuit operating over the linear section by employing additional external circuit elements. This paper attempts at analyzing a transducer characteristics by showing regions around the points of inflections where variations remain most linear.","PeriodicalId":350725,"journal":{"name":"International Conference on Computer and Communication Engineering (ICCCE'10)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124880992","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-05-11DOI: 10.1109/ICCCE.2010.5556806
K. K. Htike, Othman Omran Khalifa
Rainfall forecasting is vital for making important decisions and performing strategic planning in agriculture-dependent countries. Despite its importance, statistical rainfall forecasting, especially for long-term, has been proven to be a great challenge due to the dynamic nature of climate phenomena and random fluctuations involved in the process. Artificial Neural Networks (ANNs) have recently become very popular and they are one of the most widely used forecasting models that have enjoyed fruitful applications for forecasting purposes in many domains of engineering and computer science. The main contribution of this research is in the design, implementation and comparison of rainfall forecasting models using Focused Time-Delay Neural Networks (FTDNN). The optimal parameters of the neural network architectures were obtained from experiments while networks were trained to perform one-step-ahead predictions. The daily rainfall dataset, obtained from Malaysia Meteorological Department (MMD), was converted to monthly, biannually, quarterly and monthly datasets. Training and testing were performed on each of the datasets and corresponding accuracies of the forecasts were measured using Mean Absolute Percent Error. For testing data, results indicate that yearly rainfall dataset gives the most accurate forecasts (94.25%). As future work, more parameters such as temperature, humidity and sunshine data can be incorporated into the neural network for superior forecasting performance.
{"title":"Rainfall forecasting models using focused time-delay neural networks","authors":"K. K. Htike, Othman Omran Khalifa","doi":"10.1109/ICCCE.2010.5556806","DOIUrl":"https://doi.org/10.1109/ICCCE.2010.5556806","url":null,"abstract":"Rainfall forecasting is vital for making important decisions and performing strategic planning in agriculture-dependent countries. Despite its importance, statistical rainfall forecasting, especially for long-term, has been proven to be a great challenge due to the dynamic nature of climate phenomena and random fluctuations involved in the process. Artificial Neural Networks (ANNs) have recently become very popular and they are one of the most widely used forecasting models that have enjoyed fruitful applications for forecasting purposes in many domains of engineering and computer science. The main contribution of this research is in the design, implementation and comparison of rainfall forecasting models using Focused Time-Delay Neural Networks (FTDNN). The optimal parameters of the neural network architectures were obtained from experiments while networks were trained to perform one-step-ahead predictions. The daily rainfall dataset, obtained from Malaysia Meteorological Department (MMD), was converted to monthly, biannually, quarterly and monthly datasets. Training and testing were performed on each of the datasets and corresponding accuracies of the forecasts were measured using Mean Absolute Percent Error. For testing data, results indicate that yearly rainfall dataset gives the most accurate forecasts (94.25%). As future work, more parameters such as temperature, humidity and sunshine data can be incorporated into the neural network for superior forecasting performance.","PeriodicalId":350725,"journal":{"name":"International Conference on Computer and Communication Engineering (ICCCE'10)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116697504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-05-11DOI: 10.1109/ICCCE.2010.5556830
C. Rashidi, M. Anuar, S. Aljunid
This paper investigates the performance of direct detection technique for zero cross correlation code in Optical Code Division Multiple Access (OCDMA) which is a better technique to reduce Multi Access Interference (MAI). MAI is the main reason for performance degradation in optical CDMA especially when large numbers of users are involved. The implementation of direct detection whereby only a pair of decoder and detector is required as opposed with two pairs in the AND subtraction detection technique. OptiSystem Ver.7 has been used for simulation on system's performance. It is possible since the code properties posses' one clean signal chip for each of the channels. Subsequently, phase-induced intensity noise (PIIN) is suppressed at the receiver, thus the system performance is improved in term of signal to noise ratio and bit error rate. It has been shown that by using direct detection technique is better compare with other technique such as AND subtraction technique in ZCC code system as the varying the number of user is related to the power level of the received power is lower for BER 10−9. A larger number of users have higher insertion loss and smaller output power. For ZCC code, direct detection technique can support higher number users than the conventional technique because the number of filters at the receiver is reduced, thus a smaller power loss.
{"title":"Study of direct detection technique for zero cross correlation code in OCDMA","authors":"C. Rashidi, M. Anuar, S. Aljunid","doi":"10.1109/ICCCE.2010.5556830","DOIUrl":"https://doi.org/10.1109/ICCCE.2010.5556830","url":null,"abstract":"This paper investigates the performance of direct detection technique for zero cross correlation code in Optical Code Division Multiple Access (OCDMA) which is a better technique to reduce Multi Access Interference (MAI). MAI is the main reason for performance degradation in optical CDMA especially when large numbers of users are involved. The implementation of direct detection whereby only a pair of decoder and detector is required as opposed with two pairs in the AND subtraction detection technique. OptiSystem Ver.7 has been used for simulation on system's performance. It is possible since the code properties posses' one clean signal chip for each of the channels. Subsequently, phase-induced intensity noise (PIIN) is suppressed at the receiver, thus the system performance is improved in term of signal to noise ratio and bit error rate. It has been shown that by using direct detection technique is better compare with other technique such as AND subtraction technique in ZCC code system as the varying the number of user is related to the power level of the received power is lower for BER 10−9. A larger number of users have higher insertion loss and smaller output power. For ZCC code, direct detection technique can support higher number users than the conventional technique because the number of filters at the receiver is reduced, thus a smaller power loss.","PeriodicalId":350725,"journal":{"name":"International Conference on Computer and Communication Engineering (ICCCE'10)","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121223751","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-05-11DOI: 10.1109/ICCCE.2010.5556762
I. Abba, W. Abidin, V. P. Bong, A. Othman, K. H. Ping
Mobile satellite (MS) signal performance is affected by many factors such as ionospheric effect, multipath fading and shadowing effect. Satellites transmit signals at the same strength, but things like trees and buildings can obscure a signal beyond recognition. Some experimental works have been done using the existing geo-satellite system or airborne platform in the developed countries which is quite expensive and complicated. Very little data is represents less developed countries such as Africa and Asia. Therefore, a simple and low cost data acquisition system that can be used to carry out measurement of the MS signal performance is proposed. The experiment is carried out in open area with no obstruction of the direct wave. In addition, NMEA sentences will be used to obtain the required signal propagation parameters. The measurement is then performed to determine the relationship between the signal performance with respect to the elevation and azimuth angles. Results obtained from analysis will be used to determine the characteristics of the open space environment. This method can be used to encourage more experimental works in the less develop countries. Propagation data for the open space environment can be used as a reference to determine the MS signal quality for shadowing environments.
{"title":"Open space experimental work for L-band mobile satellite (MS) using a simple and low cost data acquisition system","authors":"I. Abba, W. Abidin, V. P. Bong, A. Othman, K. H. Ping","doi":"10.1109/ICCCE.2010.5556762","DOIUrl":"https://doi.org/10.1109/ICCCE.2010.5556762","url":null,"abstract":"Mobile satellite (MS) signal performance is affected by many factors such as ionospheric effect, multipath fading and shadowing effect. Satellites transmit signals at the same strength, but things like trees and buildings can obscure a signal beyond recognition. Some experimental works have been done using the existing geo-satellite system or airborne platform in the developed countries which is quite expensive and complicated. Very little data is represents less developed countries such as Africa and Asia. Therefore, a simple and low cost data acquisition system that can be used to carry out measurement of the MS signal performance is proposed. The experiment is carried out in open area with no obstruction of the direct wave. In addition, NMEA sentences will be used to obtain the required signal propagation parameters. The measurement is then performed to determine the relationship between the signal performance with respect to the elevation and azimuth angles. Results obtained from analysis will be used to determine the characteristics of the open space environment. This method can be used to encourage more experimental works in the less develop countries. Propagation data for the open space environment can be used as a reference to determine the MS signal quality for shadowing environments.","PeriodicalId":350725,"journal":{"name":"International Conference on Computer and Communication Engineering (ICCCE'10)","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126630548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-05-11DOI: 10.1109/ICCCE.2010.5556832
M. Abushariah, R. N. Ainon, R. Zainuddin, Othman Omran Khalifa, M. Elshafei
Lack of spoken and written training data is one o f the main issues encountered by Arabic automatic speech recognition (ASR) researchers. Almost all written and spoken corpora are not readily available to the public and many of them can only be obtained by purchasing from the Linguistic Data Consortium (LDC) or the European Language Resource Association (ELRA). There is more shortage of spoken training data as compared to written training data resulting in a great need for more speech corpora in order to serve different domains of Arabic ASR. The available spoken corpora were mainly collected from broadcast news (radios and televisions), and telephone conversations having certain technical and quality shortcomings. In order to produce a robust speaker-independent continuous automatic Arabic speech recognizer, a set of speech recordings that are rich and balanced is required. The rich characteristic is in the sense that it must contain all the phonemes of Arabic language. It must be balanced in preserving the phonetics distribution of Arabic language too. This set of speech recordings must be based on a proper written set of sentences and phrases created by experts. Therefore, it is crucial to crea te a high quality written (text) set of the sentences and phrases before recording them. This work adds a new kind of possible speech data for Arabic language based text and speech applications besides other kinds such as broadcast news and telephone conversations. Therefore, this work is an invitation to all Arabic ASR developers and research groups to explore and capitalize.
{"title":"Phonetically rich and balanced arabic speech corpus: An overview","authors":"M. Abushariah, R. N. Ainon, R. Zainuddin, Othman Omran Khalifa, M. Elshafei","doi":"10.1109/ICCCE.2010.5556832","DOIUrl":"https://doi.org/10.1109/ICCCE.2010.5556832","url":null,"abstract":"Lack of spoken and written training data is one o f the main issues encountered by Arabic automatic speech recognition (ASR) researchers. Almost all written and spoken corpora are not readily available to the public and many of them can only be obtained by purchasing from the Linguistic Data Consortium (LDC) or the European Language Resource Association (ELRA). There is more shortage of spoken training data as compared to written training data resulting in a great need for more speech corpora in order to serve different domains of Arabic ASR. The available spoken corpora were mainly collected from broadcast news (radios and televisions), and telephone conversations having certain technical and quality shortcomings. In order to produce a robust speaker-independent continuous automatic Arabic speech recognizer, a set of speech recordings that are rich and balanced is required. The rich characteristic is in the sense that it must contain all the phonemes of Arabic language. It must be balanced in preserving the phonetics distribution of Arabic language too. This set of speech recordings must be based on a proper written set of sentences and phrases created by experts. Therefore, it is crucial to crea te a high quality written (text) set of the sentences and phrases before recording them. This work adds a new kind of possible speech data for Arabic language based text and speech applications besides other kinds such as broadcast news and telephone conversations. Therefore, this work is an invitation to all Arabic ASR developers and research groups to explore and capitalize.","PeriodicalId":350725,"journal":{"name":"International Conference on Computer and Communication Engineering (ICCCE'10)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122718340","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-05-11DOI: 10.1109/ICCCE.2010.5556831
E. A. Elsheikh, M. R. Islam, A. Alam, A. F. Ismail, K. Al-Khateeb, Z. Elabdin
Dust storm effect on microwave signal attenuation has attracted many researchers recently. Many prediction models have been developed in order to predict microwave signal attenuation during the dust storm. Most of the dust storm prediction models have ignored the variation of the particle size inside the dust storms. Therefore, their predicted results didn't reflect those have been measured. In this paper two of the dust prediction models [1–2] have been investigated based on particle size statistical distributions (normal, exponential and lognormal) [3–8]. The proposed modification reflects the actual variation of the dust particle size distribution in these prediction models. The predicted attenuation based on lognormal particle size distribution have been compared with the measured data and found close agreement.
{"title":"The effect of particle size distributions on dust storm attenuation prediction for microwave propagation","authors":"E. A. Elsheikh, M. R. Islam, A. Alam, A. F. Ismail, K. Al-Khateeb, Z. Elabdin","doi":"10.1109/ICCCE.2010.5556831","DOIUrl":"https://doi.org/10.1109/ICCCE.2010.5556831","url":null,"abstract":"Dust storm effect on microwave signal attenuation has attracted many researchers recently. Many prediction models have been developed in order to predict microwave signal attenuation during the dust storm. Most of the dust storm prediction models have ignored the variation of the particle size inside the dust storms. Therefore, their predicted results didn't reflect those have been measured. In this paper two of the dust prediction models [1–2] have been investigated based on particle size statistical distributions (normal, exponential and lognormal) [3–8]. The proposed modification reflects the actual variation of the dust particle size distribution in these prediction models. The predicted attenuation based on lognormal particle size distribution have been compared with the measured data and found close agreement.","PeriodicalId":350725,"journal":{"name":"International Conference on Computer and Communication Engineering (ICCCE'10)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122748918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-05-11DOI: 10.1109/ICCCE.2010.5556858
R. F. Olanrewaju, O. Khalifa, A. Abdalla, A. Aburas, A. Zeki
It has been discovered by computational experiments that Complex Back-Propagation (CBP) algorithm is well suited for learning complex pattern [1], and it has been reported that this ability can successfully be applied in image processing with complex values. In this paper, a watermarking scheme based on Complex-Valued Neural Network, CVNN trained by CBP in transform domain is proposed. Fast Fourier Transform, FFT is used to obtain the complex values (real and imaginary part) of the host image. The complex values form the input data of CVNN. Neural networks performs best on detection, mapping, classification, learning and adaption. These features are employed to simulate the Safe Region (SR) to embed the watermark, thus, watermark are appropriately mapped to the safe region of selected coefficients. The implementation results have shown that this watermarking algorithm has high level of imperceptibility.
{"title":"Watermarking in Safe Region of frequency domain using complex-valued Neural Network","authors":"R. F. Olanrewaju, O. Khalifa, A. Abdalla, A. Aburas, A. Zeki","doi":"10.1109/ICCCE.2010.5556858","DOIUrl":"https://doi.org/10.1109/ICCCE.2010.5556858","url":null,"abstract":"It has been discovered by computational experiments that Complex Back-Propagation (CBP) algorithm is well suited for learning complex pattern [1], and it has been reported that this ability can successfully be applied in image processing with complex values. In this paper, a watermarking scheme based on Complex-Valued Neural Network, CVNN trained by CBP in transform domain is proposed. Fast Fourier Transform, FFT is used to obtain the complex values (real and imaginary part) of the host image. The complex values form the input data of CVNN. Neural networks performs best on detection, mapping, classification, learning and adaption. These features are employed to simulate the Safe Region (SR) to embed the watermark, thus, watermark are appropriately mapped to the safe region of selected coefficients. The implementation results have shown that this watermarking algorithm has high level of imperceptibility.","PeriodicalId":350725,"journal":{"name":"International Conference on Computer and Communication Engineering (ICCCE'10)","volume":"104 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122569379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-05-11DOI: 10.1109/ICCCE.2010.5556776
L. Y. K. Wang, Heng-Siong Lim, A. Tan
Adopting frequency-domain equalization (FDE) in DS-CDMA can partially restore the orthogonality of users and suppress multiple-access interference (MAI). The existing FDEs are designed with the assumption of additive white Gaussian noise. Although the assumption is quite proper for a variety of applications, it has been revealed that many background noises arising in the reality are with non-Gaussian statistics. In this paper, we show that the Gaussian-based FDEs fare poorly in non-Gaussian noise environment. To achieve reliable data transmission in frequency-selective multiple-access channel with unknown non-Gaussian noise, we propose a robust adaptive FDE which consists of a nonlinear preprocessing front-end followed by a linear adaptive FDE. The nonlinear front-end adapts itself to the unknown non-Gaussian noise to suppress the impulses while the FDE effectively removes the MAI. Results from the simulation reveal that the proposed robust FDE is able to perform successfully in both Gaussian and non-Gaussian noise environments.
{"title":"Robust adaptive frequency-domain equalization for DS-CDMA in non-Gaussian noise channels","authors":"L. Y. K. Wang, Heng-Siong Lim, A. Tan","doi":"10.1109/ICCCE.2010.5556776","DOIUrl":"https://doi.org/10.1109/ICCCE.2010.5556776","url":null,"abstract":"Adopting frequency-domain equalization (FDE) in DS-CDMA can partially restore the orthogonality of users and suppress multiple-access interference (MAI). The existing FDEs are designed with the assumption of additive white Gaussian noise. Although the assumption is quite proper for a variety of applications, it has been revealed that many background noises arising in the reality are with non-Gaussian statistics. In this paper, we show that the Gaussian-based FDEs fare poorly in non-Gaussian noise environment. To achieve reliable data transmission in frequency-selective multiple-access channel with unknown non-Gaussian noise, we propose a robust adaptive FDE which consists of a nonlinear preprocessing front-end followed by a linear adaptive FDE. The nonlinear front-end adapts itself to the unknown non-Gaussian noise to suppress the impulses while the FDE effectively removes the MAI. Results from the simulation reveal that the proposed robust FDE is able to perform successfully in both Gaussian and non-Gaussian noise environments.","PeriodicalId":350725,"journal":{"name":"International Conference on Computer and Communication Engineering (ICCCE'10)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121888855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-05-11DOI: 10.1109/ICCCE.2010.5556793
S. Hameed
Hajj (pilgrimage) is the fifth pillar of Islam that must be carried out at least once in their lifetime by every able-bodied Muslim. It is the largest annual convention of Faith in the world that requires traveling to Mecca. Hajj is performed based on predetermined dates, times, and places. It is a demonstration of the solidarity of the Muslim people, and their submission to Allah the Exalted (GOD). Hajj is an important event in Muslims' life, which offers religious, educational, scientific, social, economical, political and other benefits that are rewarded by Allah SWT to Muslims. Two to three millions out of One and half billion of Muslims around the world, performed Hajj annually. Time, date and place limitations related to Hajj leads to several difficulties and problems that require suitable solutions. Literature review LR, survey, and analysis summarize Hajj difficulties as: Managerial and organizational problems; Lack of information and misunderstanding; and difficulties in dialogue between people, pilgrims and organizational committees. This research paper proposes suitable solutions to solve difficulties using the powerful capabilities of information and communication technologies ICT. It shows the main components of a Comprehensive Hajj Model that includes: comprehensive Hajj database model for managing, organizing and tracking pilgrims and other stakeholders; a Multi- lingual Hajj Educational model on Web-based or CD; and a portable Hajj emergency and guidance model. This comprehensive solution offers: religious, educational, scientific, social, economical, political and other benefits for Muslims. It can be consider as a step toward Muslim unity and offering a unified understanding and cooperation between Muslims.
{"title":"ICT to serve Hajj: Analytical study","authors":"S. Hameed","doi":"10.1109/ICCCE.2010.5556793","DOIUrl":"https://doi.org/10.1109/ICCCE.2010.5556793","url":null,"abstract":"Hajj (pilgrimage) is the fifth pillar of Islam that must be carried out at least once in their lifetime by every able-bodied Muslim. It is the largest annual convention of Faith in the world that requires traveling to Mecca. Hajj is performed based on predetermined dates, times, and places. It is a demonstration of the solidarity of the Muslim people, and their submission to Allah the Exalted (GOD). Hajj is an important event in Muslims' life, which offers religious, educational, scientific, social, economical, political and other benefits that are rewarded by Allah SWT to Muslims. Two to three millions out of One and half billion of Muslims around the world, performed Hajj annually. Time, date and place limitations related to Hajj leads to several difficulties and problems that require suitable solutions. Literature review LR, survey, and analysis summarize Hajj difficulties as: Managerial and organizational problems; Lack of information and misunderstanding; and difficulties in dialogue between people, pilgrims and organizational committees. This research paper proposes suitable solutions to solve difficulties using the powerful capabilities of information and communication technologies ICT. It shows the main components of a Comprehensive Hajj Model that includes: comprehensive Hajj database model for managing, organizing and tracking pilgrims and other stakeholders; a Multi- lingual Hajj Educational model on Web-based or CD; and a portable Hajj emergency and guidance model. This comprehensive solution offers: religious, educational, scientific, social, economical, political and other benefits for Muslims. It can be consider as a step toward Muslim unity and offering a unified understanding and cooperation between Muslims.","PeriodicalId":350725,"journal":{"name":"International Conference on Computer and Communication Engineering (ICCCE'10)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128511906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2010-05-11DOI: 10.1109/ICCCE.2010.5556814
Chen-Shu Wang, Chia-Chuan Yeh, Chun-Yi Li
Traveling-schedule (TS) arrangement is a classical ill-define problem which lacks of structure and fulfills uncertainty and dynamic complexity. In general, there are two ways to resolve TS arrangement (TSA) problem, including: package tourism provide by travel agency who arrange entire traveling program. The other is independent tourism that travelers should collect information and arrange all traveling-detail themselves. Nowadays, independent tourism is getting popular and may instead of total package one due to tourism flexibility and customization. To cater for independent tourism customer, many travel agencies have already developed recommender system to provide online traveler with particular tourism packages according to their query conditions. However, such recommendation result usually become involve in package tourism advertisements and lack of flexibility. Additionally, such recommender mechanism can not replicate important word-of-mouse effect about traveling experience. Thus, the recommender mechanism should be revised for TSA problem solving. This research proposed an intelligence traveling recommender (iTR) system based on commonsense reasoning (CR) algorithm. iTR includes two reasoning processes, the general reasoning and the exception one. Furthermore, iTR consists of three intelligent agents and three automatic mechanisms enable user constantly refine and revise suggested traveling-schedule by iTR. CR is an appropriate methodology to deal with TSA problem because of CR can replicate human decision process actually. Finally, a demonstration TSA scenario is presented to illustrate the effect and feasibility of proposed iTR recommender architecture.
{"title":"Intelligence traveling schedule recommender based on commonsense reasoning algorithm","authors":"Chen-Shu Wang, Chia-Chuan Yeh, Chun-Yi Li","doi":"10.1109/ICCCE.2010.5556814","DOIUrl":"https://doi.org/10.1109/ICCCE.2010.5556814","url":null,"abstract":"Traveling-schedule (TS) arrangement is a classical ill-define problem which lacks of structure and fulfills uncertainty and dynamic complexity. In general, there are two ways to resolve TS arrangement (TSA) problem, including: package tourism provide by travel agency who arrange entire traveling program. The other is independent tourism that travelers should collect information and arrange all traveling-detail themselves. Nowadays, independent tourism is getting popular and may instead of total package one due to tourism flexibility and customization. To cater for independent tourism customer, many travel agencies have already developed recommender system to provide online traveler with particular tourism packages according to their query conditions. However, such recommendation result usually become involve in package tourism advertisements and lack of flexibility. Additionally, such recommender mechanism can not replicate important word-of-mouse effect about traveling experience. Thus, the recommender mechanism should be revised for TSA problem solving. This research proposed an intelligence traveling recommender (iTR) system based on commonsense reasoning (CR) algorithm. iTR includes two reasoning processes, the general reasoning and the exception one. Furthermore, iTR consists of three intelligent agents and three automatic mechanisms enable user constantly refine and revise suggested traveling-schedule by iTR. CR is an appropriate methodology to deal with TSA problem because of CR can replicate human decision process actually. Finally, a demonstration TSA scenario is presented to illustrate the effect and feasibility of proposed iTR recommender architecture.","PeriodicalId":350725,"journal":{"name":"International Conference on Computer and Communication Engineering (ICCCE'10)","volume":"93 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128573912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}