Pub Date : 2024-03-01DOI: 10.1016/j.jiixd.2023.10.007
Xiaodong Wu, Ran Duan, Jianbing Ni
This paper delves into the realm of ChatGPT, an AI-powered chatbot that utilizes topic modeling and reinforcement learning to generate natural responses. Although ChatGPT holds immense promise across various industries, such as customer service, education, mental health treatment, personal productivity, and content creation, it is essential to address its security, privacy, and ethical implications. By exploring the upgrade path from GPT-1 to GPT-4, discussing the model's features, limitations, and potential applications, this study aims to shed light on the potential risks of integrating ChatGPT into our daily lives. Focusing on security, privacy, and ethics issues, we highlight the challenges these concerns pose for widespread adoption. Finally, we analyze the open problems in these areas, calling for concerted efforts to ensure the development of secure and ethically sound large language models.
{"title":"Unveiling security, privacy, and ethical concerns of ChatGPT","authors":"Xiaodong Wu, Ran Duan, Jianbing Ni","doi":"10.1016/j.jiixd.2023.10.007","DOIUrl":"10.1016/j.jiixd.2023.10.007","url":null,"abstract":"<div><p>This paper delves into the realm of ChatGPT, an AI-powered chatbot that utilizes topic modeling and reinforcement learning to generate natural responses. Although ChatGPT holds immense promise across various industries, such as customer service, education, mental health treatment, personal productivity, and content creation, it is essential to address its security, privacy, and ethical implications. By exploring the upgrade path from GPT-1 to GPT-4, discussing the model's features, limitations, and potential applications, this study aims to shed light on the potential risks of integrating ChatGPT into our daily lives. Focusing on security, privacy, and ethics issues, we highlight the challenges these concerns pose for widespread adoption. Finally, we analyze the open problems in these areas, calling for concerted efforts to ensure the development of secure and ethically sound large language models.</p></div>","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"2 2","pages":"Pages 102-115"},"PeriodicalIF":0.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949715923000707/pdfft?md5=d54b43065c82b1dd4241ba7d67e27d46&pid=1-s2.0-S2949715923000707-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136129887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jiixd.2024.02.003
Sean Lalla, Rongxing Lu, Yunguo Guan, Songnian Zhang
Outsourcing decision tree models to cloud servers can allow model providers to distribute their models at scale without purchasing dedicated hardware for model hosting. However, model providers may be forced to disclose private model details when hosting their models in the cloud. Due to the time and monetary investments associated with model training, model providers may be reluctant to host their models in the cloud due to these privacy concerns. Furthermore, clients may be reluctant to use these outsourced models because their private queries or their results may be disclosed to the cloud servers. In this paper, we propose BloomDT, a privacy-preserving scheme for decision tree inference, which uses Bloom filters to hide the original decision tree's structure, the threshold values of each node, and the order in which features are tested while maintaining reliable classification results that are secure even if the cloud servers collude. Our scheme's security and performance are verified through rigorous testing and analysis.
{"title":"BloomDT - An improved privacy-preserving decision tree inference scheme","authors":"Sean Lalla, Rongxing Lu, Yunguo Guan, Songnian Zhang","doi":"10.1016/j.jiixd.2024.02.003","DOIUrl":"10.1016/j.jiixd.2024.02.003","url":null,"abstract":"<div><p>Outsourcing decision tree models to cloud servers can allow model providers to distribute their models at scale without purchasing dedicated hardware for model hosting. However, model providers may be forced to disclose private model details when hosting their models in the cloud. Due to the time and monetary investments associated with model training, model providers may be reluctant to host their models in the cloud due to these privacy concerns. Furthermore, clients may be reluctant to use these outsourced models because their private queries or their results may be disclosed to the cloud servers. In this paper, we propose BloomDT, a privacy-preserving scheme for decision tree inference, which uses Bloom filters to hide the original decision tree's structure, the threshold values of each node, and the order in which features are tested while maintaining reliable classification results that are secure even if the cloud servers collude. Our scheme's security and performance are verified through rigorous testing and analysis.</p></div>","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"2 2","pages":"Pages 130-147"},"PeriodicalIF":0.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949715924000088/pdfft?md5=7d9b7fbb49ca778f809e1f16a75c50b6&pid=1-s2.0-S2949715924000088-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140469188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jiixd.2024.01.001
Qi Xu , Hui Zhu , Yandong Zheng , Fengwei Wang , Le Gao
With the rapid development of location-based services and online social networks, POI recommendation services considering geographic and social factors have received extensive attention. Meanwhile, the vigorous development of cloud computing has prompted service providers to outsource data to the cloud to provide POI recommendation services. However, there is a degree of distrust of the cloud by service providers. To protect digital assets, service providers encrypt data before outsourcing it. However, encryption reduces data availability, making it more challenging to provide POI recommendation services in outsourcing scenarios. Some privacy-preserving schemes for geo-social-based POI recommendation have been presented, but they have some limitations in supporting group query, considering both geographic and social factors, and query accuracy, making these schemes impractical. To solve this issue, we propose two practical and privacy-preserving geo-social-based POI recommendation schemes for single user and group users, which are named GSPR-S and GSPR-G. Specifically, we first utilize the quad tree to organize geographic data and the MinHash method to index social data. Then, we apply BGV fully homomorphic encryption to design some private algorithms, including a private max/min operation algorithm, a private rectangular set operation algorithm, and a private rectangular overlapping detection algorithm. After that, we use these algorithms as building blocks in our schemes for efficiency improvement. According to security analysis, our schemes are proven to be secure against the honest-but-curious cloud servers, and experimental results show that our schemes have good performance.
随着基于位置的服务和在线社交网络的快速发展,考虑地理和社交因素的 POI 推荐服务受到广泛关注。与此同时,云计算的蓬勃发展也促使服务提供商将数据外包给云,以提供 POI 推荐服务。然而,服务提供商对云存在一定程度的不信任。为了保护数字资产,服务提供商会在外包数据前对其进行加密。然而,加密降低了数据的可用性,使得在外包场景中提供 POI 推荐服务更具挑战性。目前已经提出了一些基于地理社交的 POI 推荐的隐私保护方案,但这些方案在支持群组查询、考虑地理和社交因素以及查询准确性方面存在一些局限性,使得这些方案不切实际。为了解决这个问题,我们提出了两种实用且能保护隐私的基于地理社交的 POI 推荐方案,分别适用于单个用户和群体用户,分别命名为 GSPR-S 和 GSPR-G。具体来说,我们首先利用四叉树来组织地理数据,并利用 MinHash 方法来索引社交数据。然后,我们应用 BGV 全同态加密技术设计了一些私有算法,包括私有最大/最小运算算法、私有矩形集运算算法和私有矩形重叠检测算法。之后,我们将这些算法作为我们方案的构建模块,以提高效率。根据安全性分析,我们的方案被证明可以安全地对抗诚实但好奇的云服务器,实验结果表明我们的方案具有良好的性能。
{"title":"Practical and privacy-preserving geo-social-based POI recommendation","authors":"Qi Xu , Hui Zhu , Yandong Zheng , Fengwei Wang , Le Gao","doi":"10.1016/j.jiixd.2024.01.001","DOIUrl":"10.1016/j.jiixd.2024.01.001","url":null,"abstract":"<div><p>With the rapid development of location-based services and online social networks, POI recommendation services considering geographic and social factors have received extensive attention. Meanwhile, the vigorous development of cloud computing has prompted service providers to outsource data to the cloud to provide POI recommendation services. However, there is a degree of distrust of the cloud by service providers. To protect digital assets, service providers encrypt data before outsourcing it. However, encryption reduces data availability, making it more challenging to provide POI recommendation services in outsourcing scenarios. Some privacy-preserving schemes for geo-social-based POI recommendation have been presented, but they have some limitations in supporting group query, considering both geographic and social factors, and query accuracy, making these schemes impractical. To solve this issue, we propose two practical and privacy-preserving geo-social-based POI recommendation schemes for single user and group users, which are named GSPR-S and GSPR-G. Specifically, we first utilize the quad tree to organize geographic data and the MinHash method to index social data. Then, we apply BGV fully homomorphic encryption to design some private algorithms, including a private max/min operation algorithm, a private rectangular set operation algorithm, and a private rectangular overlapping detection algorithm. After that, we use these algorithms as building blocks in our schemes for efficiency improvement. According to security analysis, our schemes are proven to be secure against the honest-but-curious cloud servers, and experimental results show that our schemes have good performance.</p></div>","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"2 2","pages":"Pages 148-166"},"PeriodicalIF":0.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949715924000015/pdfft?md5=a2e0865bfbb9a59bb240fc8da82554c1&pid=1-s2.0-S2949715924000015-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139395049","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Detecting ink mismatch is a significant challenge in verifying the authenticity of documents, especially when dealing with uneven ink distribution. Conventional imaging methods frequently fail to distinguish visually similar inks. Our study presents a novel hyperspectral unmixing approach to detect ink mismatches in unbalanced clusters. The proposed method identifies unique spectral characteristics of different inks employing k-means clustering and Gaussian mixture models (GMMs) to perform color segmentation on different ink types and utilizes elbow estimation and silhouette coefficient to evaluate the number of inks estimation precisely. For a more accurate estimation of quantity, which is generally not an attribute of clustering methods, we employed entropy calculations in the red, green, and blue depth channels for precise abundance estimation of ink. This unique combination of basic techniques in conjunction exhibits better efficacy in performing ink unmixing and provides a real-world document forensic solution compared to current methods that rely on assumptions like prior knowledge of the inks used in a document and deep learning-based methods that rely heavily on abundant training datasets. We evaluate our approach on the iVision handwritten hyperspectral images dataset (iVision HHID), which is a comprehensive and rich dataset that surpasses the commonly-used UWA writing inks hyperspectral images (WIHSI) database in size and diversity. This study has accomplished the unmixing task with three main challenges: unmixing of diverse ink spectral signatures (149 spectral bands instead of 33 bands in the previous dataset), without using prior knowledge and assumptions about the number of inks used in the questioned document, and not requiring large training data for performing unmixing. Furthermore, the security of the proposed document authentication methodology to address the likelihood of forgeries or manipulations in questioned documents is enhanced as compared to previous works relying on known inks and known spectrum. Randomization techniques and anomaly detection mechanisms are used in our methodology which increases the difficulty for adversaries to predict and manipulate specific aspects of the input data in questioned documents, thereby enhancing the robustness of our method. The code for conducting this research can be accessed at GitHub repository.
检测油墨不匹配是验证文件真伪的一大挑战,尤其是在油墨分布不均匀的情况下。传统的成像方法经常无法区分视觉上相似的油墨。我们的研究提出了一种新颖的高光谱非混合方法,用于检测不平衡集群中的油墨错配。所提出的方法利用 K 均值聚类和高斯混合模型(GMMs)来识别不同油墨的独特光谱特征,从而对不同类型的油墨进行颜色分割,并利用肘部估计和剪影系数来精确评估油墨估计数量。为了更精确地估算数量(这通常不是聚类方法的特性),我们在红色、绿色和蓝色深度通道中采用了熵计算,以精确估算墨水的丰度。与依赖文档中所用墨水的先验知识等假设的现有方法和严重依赖丰富训练数据集的基于深度学习的方法相比,这种将基本技术结合在一起的独特方法在进行墨水解混合时表现出更好的功效,并提供了一种真实世界的文档取证解决方案。我们在 iVision 手写高光谱图像数据集(iVision HHID)上评估了我们的方法,该数据集全面而丰富,在规模和多样性上超过了常用的 UWA 书写墨水高光谱图像(WIHSI)数据库。这项研究在完成非混合任务时面临三大挑战:非混合多种墨水光谱特征(149 个光谱带而不是之前数据集中的 33 个带),不使用关于问题文档中使用的墨水数量的先验知识和假设,以及执行非混合时不需要大量训练数据。此外,与之前依赖已知油墨和已知光谱的工作相比,所提出的文件认证方法的安全性得到了提高,可以解决受质疑文件中可能存在的伪造或篡改问题。我们的方法采用了随机化技术和异常检测机制,增加了对手预测和篡改问题文档中输入数据特定方面的难度,从而增强了我们方法的鲁棒性。本研究的代码可在 GitHub 存储库中获取。
{"title":"A hyperspectral unmixing approach for ink mismatch detection in unbalanced clusters","authors":"Faryal Aurooj Nasir , Salman Liaquat , Khurram Khurshid , Nor Muzlifah Mahyuddin","doi":"10.1016/j.jiixd.2024.01.004","DOIUrl":"10.1016/j.jiixd.2024.01.004","url":null,"abstract":"<div><p>Detecting ink mismatch is a significant challenge in verifying the authenticity of documents, especially when dealing with uneven ink distribution. Conventional imaging methods frequently fail to distinguish visually similar inks. Our study presents a novel hyperspectral unmixing approach to detect ink mismatches in unbalanced clusters. The proposed method identifies unique spectral characteristics of different inks employing k-means clustering and Gaussian mixture models (GMMs) to perform color segmentation on different ink types and utilizes elbow estimation and silhouette coefficient to evaluate the number of inks estimation precisely. For a more accurate estimation of quantity, which is generally not an attribute of clustering methods, we employed entropy calculations in the red, green, and blue depth channels for precise abundance estimation of ink. This unique combination of basic techniques in conjunction exhibits better efficacy in performing ink unmixing and provides a real-world document forensic solution compared to current methods that rely on assumptions like prior knowledge of the inks used in a document and deep learning-based methods that rely heavily on abundant training datasets. We evaluate our approach on the iVision handwritten hyperspectral images dataset (iVision HHID), which is a comprehensive and rich dataset that surpasses the commonly-used UWA writing inks hyperspectral images (WIHSI) database in size and diversity. This study has accomplished the unmixing task with three main challenges: unmixing of diverse ink spectral signatures (149 spectral bands instead of 33 bands in the previous dataset), without using prior knowledge and assumptions about the number of inks used in the questioned document, and not requiring large training data for performing unmixing. Furthermore, the security of the proposed document authentication methodology to address the likelihood of forgeries or manipulations in questioned documents is enhanced as compared to previous works relying on known inks and known spectrum. Randomization techniques and anomaly detection mechanisms are used in our methodology which increases the difficulty for adversaries to predict and manipulate specific aspects of the input data in questioned documents, thereby enhancing the robustness of our method. The code for conducting this research can be accessed at <span>GitHub repository</span><svg><path></path></svg>.</p></div>","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"2 2","pages":"Pages 177-190"},"PeriodicalIF":0.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949715924000040/pdfft?md5=3d98b093a0be134b496feff3d3fa509c&pid=1-s2.0-S2949715924000040-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139634593","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jiixd.2023.12.002
Li Luo, Yining Liu
Sensors are widely applied in the collection of voice data. Since many attributes of voice data are sensitive such as user emotions, identity, raw voice collection may lead serious privacy threat. In the past, traditional feature extraction obtains and encrypts voice features that are then transmitted to upstream servers. In order to avoid sensitive attribute disclosure, it is necessary to separate the sensitive attributes from non-sensitive attributes of voice data. Motivated by this, user-optional privacy transmission framework for voice data (called: Voice Fence Wall) is proposed. Firstly, we provide user-optional, which means users can choose the attributes (sensitive attributes) they want to be protected. Secondly, Voice Fence Wall utilizes minimum mutual information (MI) to reduce the correlation between sensitive and non-sensitive attributes, thereby separating these attributes. Finally, only the separated non-sensitive attributes are transmitted to the upstream server, the quality of voice services is satisfied without leaking sensitive attributes. To verify the reliability and practicability, three voice datasets are used to evaluate the model, the experiments demonstrate that Voice Fence Wall not only effectively separates attributes to resist attribute inference attacks, but also outperforms related work in terms of classification performance. Specifically, our framework achieves 89.84 % accuracy in sentiment recognition and 6.01 % equal error rate in voice authentication.
{"title":"Voice Fence Wall: User-optional voice privacy transmission","authors":"Li Luo, Yining Liu","doi":"10.1016/j.jiixd.2023.12.002","DOIUrl":"10.1016/j.jiixd.2023.12.002","url":null,"abstract":"<div><p>Sensors are widely applied in the collection of voice data. Since many attributes of voice data are sensitive such as user emotions, identity, raw voice collection may lead serious privacy threat. In the past, traditional feature extraction obtains and encrypts voice features that are then transmitted to upstream servers. In order to avoid sensitive attribute disclosure, it is necessary to separate the sensitive attributes from non-sensitive attributes of voice data. Motivated by this, user-optional privacy transmission framework for voice data (called: Voice Fence Wall) is proposed. Firstly, we provide user-optional, which means users can choose the attributes (sensitive attributes) they want to be protected. Secondly, Voice Fence Wall utilizes minimum mutual information (MI) to reduce the correlation between sensitive and non-sensitive attributes, thereby separating these attributes. Finally, only the separated non-sensitive attributes are transmitted to the upstream server, the quality of voice services is satisfied without leaking sensitive attributes. To verify the reliability and practicability, three voice datasets are used to evaluate the model, the experiments demonstrate that Voice Fence Wall not only effectively separates attributes to resist attribute inference attacks, but also outperforms related work in terms of classification performance. Specifically, our framework achieves 89.84 % accuracy in sentiment recognition and 6.01 % equal error rate in voice authentication.</p></div>","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"2 2","pages":"Pages 116-129"},"PeriodicalIF":0.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S294971592300080X/pdfft?md5=7d514122810a42466002016ad09b7381&pid=1-s2.0-S294971592300080X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139393204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Data security and privacy computing in artificial intelligence","authors":"Dengguo Feng, Hui Li, Rongxing Lu, Zheli Liu, Jianbing Ni, Hui Zhu","doi":"10.1016/j.jiixd.2024.02.007","DOIUrl":"https://doi.org/10.1016/j.jiixd.2024.02.007","url":null,"abstract":"","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"2 2","pages":"Pages 99-101"},"PeriodicalIF":0.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S294971592400012X/pdfft?md5=b365b0de34c8f2cd89fb4535c7790036&pid=1-s2.0-S294971592400012X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140555268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-08DOI: 10.1016/j.jiixd.2024.01.002
Alhassan Mumuni , Fuseini Mumuni
Modern approach to artificial intelligence (AI) aims to design algorithms that learn directly from data. This approach has achieved impressive results and has contributed significantly to the progress of AI, particularly in the sphere of supervised deep learning. It has also simplified the design of machine learning systems as the learning process is highly automated. However, not all data processing tasks in conventional deep learning pipelines have been automated. In most cases data has to be manually collected, preprocessed and further extended through data augmentation before they can be effective for training. Recently, special techniques for automating these tasks have emerged. The automation of data processing tasks is driven by the need to utilize large volumes of complex, heterogeneous data for machine learning and big data applications. Today, end-to-end automated data processing systems based on automated machine learning (AutoML) techniques are capable of taking raw data and transforming them into useful features for big data tasks by automating all intermediate processing stages. In this work, we present a thorough review of approaches for automating data processing tasks in deep learning pipelines, including automated data preprocessing – e.g., data cleaning, labeling, missing data imputation, and categorical data encoding – as well as data augmentation (including synthetic data generation using generative AI methods) and feature engineering – specifically, automated feature extraction, feature construction and feature selection. In addition to automating specific data processing tasks, we discuss the use of AutoML methods and tools to simultaneously optimize all stages of the machine learning pipeline.
{"title":"Automated data processing and feature engineering for deep learning and big data applications: A survey","authors":"Alhassan Mumuni , Fuseini Mumuni","doi":"10.1016/j.jiixd.2024.01.002","DOIUrl":"10.1016/j.jiixd.2024.01.002","url":null,"abstract":"<div><div>Modern approach to artificial intelligence (AI) aims to design algorithms that learn directly from data. This approach has achieved impressive results and has contributed significantly to the progress of AI, particularly in the sphere of supervised deep learning. It has also simplified the design of machine learning systems as the learning process is highly automated. However, not all data processing tasks in conventional deep learning pipelines have been automated. In most cases data has to be manually collected, preprocessed and further extended through data augmentation before they can be effective for training. Recently, special techniques for automating these tasks have emerged. The automation of data processing tasks is driven by the need to utilize large volumes of complex, heterogeneous data for machine learning and big data applications. Today, end-to-end automated data processing systems based on automated machine learning (AutoML) techniques are capable of taking raw data and transforming them into useful features for big data tasks by automating all intermediate processing stages. In this work, we present a thorough review of approaches for automating data processing tasks in deep learning pipelines, including automated data preprocessing – e.g., data cleaning, labeling, missing data imputation, and categorical data encoding – as well as data augmentation (including synthetic data generation using generative AI methods) and feature engineering – specifically, automated feature extraction, feature construction and feature selection. In addition to automating specific data processing tasks, we discuss the use of AutoML methods and tools to simultaneously optimize all stages of the machine learning pipeline.</div></div>","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"3 2","pages":"Pages 113-153"},"PeriodicalIF":0.0,"publicationDate":"2024-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139454323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-01DOI: 10.1016/j.jiixd.2023.10.002
Imrus Salehin , Md. Shamiul Islam , Pritom Saha , S.M. Noman , Azra Tuni , Md. Mehedi Hasan , Md. Abu Baten
AutoML (Automated Machine Learning) is an emerging field that aims to automate the process of building machine learning models. AutoML emerged to increase productivity and efficiency by automating as much as possible the inefficient work that occurs while repeating this process whenever machine learning is applied. In particular, research has been conducted for a long time on technologies that can effectively develop high-quality models by minimizing the intervention of model developers in the process from data preprocessing to algorithm selection and tuning. In this semantic review research, we summarize the data processing requirements for AutoML approaches and provide a detailed explanation. We place greater emphasis on neural architecture search (NAS) as it currently represents a highly popular sub-topic within the field of AutoML. NAS methods use machine learning algorithms to search through a large space of possible architectures and find the one that performs best on a given task. We provide a summary of the performance achieved by representative NAS algorithms on the CIFAR-10, CIFAR-100, ImageNet and well-known benchmark datasets. Additionally, we delve into several noteworthy research directions in NAS methods including one/two-stage NAS, one-shot NAS and joint hyperparameter with architecture optimization. We discussed how the search space size and complexity in NAS can vary depending on the specific problem being addressed. To conclude, we examine several open problems (SOTA problems) within current AutoML methods that assure further investigation in future research.
AutoML(自动化机器学习)是一个新兴领域,旨在实现机器学习模型构建过程的自动化。AutoML 的出现是为了尽可能自动化重复机器学习过程中出现的低效工作,从而提高生产率和效率。特别是,从数据预处理到算法选择和调整,模型开发人员在这一过程中的干预降到最低,从而有效开发出高质量模型的技术已经研究了很长时间。在这项语义回顾研究中,我们总结了 AutoML 方法的数据处理要求,并提供了详细的解释。我们更加重视神经架构搜索(NAS),因为它是目前 AutoML 领域非常热门的子课题。NAS 方法使用机器学习算法在大量可能的架构中进行搜索,找出在给定任务中表现最佳的架构。我们总结了具有代表性的 NAS 算法在 CIFAR-10、CIFAR-100、ImageNet 和知名基准数据集上取得的性能。此外,我们还深入探讨了 NAS 方法中几个值得关注的研究方向,包括单/两阶段 NAS、单次 NAS 和联合超参数与架构优化。我们讨论了 NAS 的搜索空间大小和复杂性如何因所解决的具体问题而异。最后,我们探讨了当前 AutoML 方法中的几个开放问题(SOTA 问题),这些问题值得在未来的研究中进一步探讨。
{"title":"AutoML: A systematic review on automated machine learning with neural architecture search","authors":"Imrus Salehin , Md. Shamiul Islam , Pritom Saha , S.M. Noman , Azra Tuni , Md. Mehedi Hasan , Md. Abu Baten","doi":"10.1016/j.jiixd.2023.10.002","DOIUrl":"10.1016/j.jiixd.2023.10.002","url":null,"abstract":"<div><p>AutoML (Automated Machine Learning) is an emerging field that aims to automate the process of building machine learning models. AutoML emerged to increase productivity and efficiency by automating as much as possible the inefficient work that occurs while repeating this process whenever machine learning is applied. In particular, research has been conducted for a long time on technologies that can effectively develop high-quality models by minimizing the intervention of model developers in the process from data preprocessing to algorithm selection and tuning. In this semantic review research, we summarize the data processing requirements for AutoML approaches and provide a detailed explanation. We place greater emphasis on neural architecture search (NAS) as it currently represents a highly popular sub-topic within the field of AutoML. NAS methods use machine learning algorithms to search through a large space of possible architectures and find the one that performs best on a given task. We provide a summary of the performance achieved by representative NAS algorithms on the CIFAR-10, CIFAR-100, ImageNet and well-known benchmark datasets. Additionally, we delve into several noteworthy research directions in NAS methods including one/two-stage NAS, one-shot NAS and joint hyperparameter with architecture optimization. We discussed how the search space size and complexity in NAS can vary depending on the specific problem being addressed. To conclude, we examine several open problems (SOTA problems) within current AutoML methods that assure further investigation in future research.</p></div>","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"2 1","pages":"Pages 52-81"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949715923000604/pdfft?md5=a79f7fb3cdab55edd3b7838063f99f50&pid=1-s2.0-S2949715923000604-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135849912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-01DOI: 10.1016/j.jiixd.2023.07.002
Chaozheng Xue , Tao Li , Yongzhao Li
With the increasing popularity of civilian unmanned aerial vehicles (UAVs), safety issues arising from unsafe operations and terrorist activities have received growing attention. To address this problem, an accurate classification and positioning system is needed. Considering that UAVs usually use radio frequency (RF) signals for video transmission, in this paper, we design a passive distributed monitoring system that can classify and locate UAVs according to their RF signals. Specifically, three passive receivers are arranged in different locations to receive RF signals. Due to the noncooperation between a UAV and receivers, it is necessary to detect whether there is a UAV signal from the received signals. Hence, convolutional neural network (CNN) is proposed to not only detect the presence of the UAV, but also classify its type. After the UAV signal is detected, the time difference of arrival (TDOA) of the UAV signal arriving at the receiver is estimated by the cross-correlation method to obtain the corresponding distance difference. Finally, the Chan algorithm is used to calculate the location of the UAV. We deploy a distributed system constructed by three software defined radio (SDR) receivers on the campus playground, and conduct extensive experiments in a real wireless environment. The experimental results have successfully validated the proposed system.
{"title":"Radio frequency based distributed system for noncooperative UAV classification and positioning","authors":"Chaozheng Xue , Tao Li , Yongzhao Li","doi":"10.1016/j.jiixd.2023.07.002","DOIUrl":"10.1016/j.jiixd.2023.07.002","url":null,"abstract":"<div><p>With the increasing popularity of civilian unmanned aerial vehicles (UAVs), safety issues arising from unsafe operations and terrorist activities have received growing attention. To address this problem, an accurate classification and positioning system is needed. Considering that UAVs usually use radio frequency (RF) signals for video transmission, in this paper, we design a passive distributed monitoring system that can classify and locate UAVs according to their RF signals. Specifically, three passive receivers are arranged in different locations to receive RF signals. Due to the noncooperation between a UAV and receivers, it is necessary to detect whether there is a UAV signal from the received signals. Hence, convolutional neural network (CNN) is proposed to not only detect the presence of the UAV, but also classify its type. After the UAV signal is detected, the time difference of arrival (TDOA) of the UAV signal arriving at the receiver is estimated by the cross-correlation method to obtain the corresponding distance difference. Finally, the Chan algorithm is used to calculate the location of the UAV. We deploy a distributed system constructed by three software defined radio (SDR) receivers on the campus playground, and conduct extensive experiments in a real wireless environment. The experimental results have successfully validated the proposed system.</p></div>","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"2 1","pages":"Pages 42-51"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949715923000446/pdfft?md5=462b514a709497f9d3e6393f3ad2f8f7&pid=1-s2.0-S2949715923000446-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84541549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-01DOI: 10.1016/j.jiixd.2023.10.006
Heng Yin , Zhaoxia Yin , Zhenzhe Gao , Hang Su , Xinpeng Zhang , Bin Luo
Deep neural networks (DNNs) are widely used in real-world applications, thanks to their exceptional performance in image recognition. However, their vulnerability to attacks, such as Trojan and data poison, can compromise the integrity and stability of DNN applications. Therefore, it is crucial to verify the integrity of DNN models to ensure their security. Previous research on model watermarking for integrity detection has encountered the issue of overexposure of model parameters during embedding and extraction of the watermark. To address this problem, we propose a novel score-based black-box DNN fragile watermarking framework called fragile trigger generation (FTG). The FTG framework only requires the prediction probability distribution of the final output of the classifier during the watermarking process. It generates different fragile samples as the trigger, based on the classification prediction probability of the target classifier and a specified prediction probability mask to watermark it. Different prediction probability masks can promote the generation of fragile samples in corresponding distribution types. The whole watermarking process does not affect the performance of the target classifier. When verifying the watermarking information, the FTG only needs to compare the prediction results of the model on the samples with the previous label. As a result, the required model parameter information is reduced, and the FTG only needs a few samples to detect slight modifications in the model. Experimental results demonstrate the effectiveness of our proposed method and show its superiority over related work. The FTG framework provides a robust solution for verifying the integrity of DNN models, and its effectiveness in detecting slight modifications makes it a valuable tool for ensuring the security and stability of DNN applications.
{"title":"FTG: Score-based black-box watermarking by fragile trigger generation for deep model integrity verification","authors":"Heng Yin , Zhaoxia Yin , Zhenzhe Gao , Hang Su , Xinpeng Zhang , Bin Luo","doi":"10.1016/j.jiixd.2023.10.006","DOIUrl":"10.1016/j.jiixd.2023.10.006","url":null,"abstract":"<div><p>Deep neural networks (DNNs) are widely used in real-world applications, thanks to their exceptional performance in image recognition. However, their vulnerability to attacks, such as Trojan and data poison, can compromise the integrity and stability of DNN applications. Therefore, it is crucial to verify the integrity of DNN models to ensure their security. Previous research on model watermarking for integrity detection has encountered the issue of overexposure of model parameters during embedding and extraction of the watermark. To address this problem, we propose a novel score-based black-box DNN fragile watermarking framework called fragile trigger generation (FTG). The FTG framework only requires the prediction probability distribution of the final output of the classifier during the watermarking process. It generates different fragile samples as the trigger, based on the classification prediction probability of the target classifier and a specified prediction probability mask to watermark it. Different prediction probability masks can promote the generation of fragile samples in corresponding distribution types. The whole watermarking process does not affect the performance of the target classifier. When verifying the watermarking information, the FTG only needs to compare the prediction results of the model on the samples with the previous label. As a result, the required model parameter information is reduced, and the FTG only needs a few samples to detect slight modifications in the model. Experimental results demonstrate the effectiveness of our proposed method and show its superiority over related work. The FTG framework provides a robust solution for verifying the integrity of DNN models, and its effectiveness in detecting slight modifications makes it a valuable tool for ensuring the security and stability of DNN applications.</p></div>","PeriodicalId":100790,"journal":{"name":"Journal of Information and Intelligence","volume":"2 1","pages":"Pages 28-41"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949715923000641/pdfft?md5=60f402130fb47c84b855a467ea72516c&pid=1-s2.0-S2949715923000641-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135412511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}