首页 > 最新文献

Proceedings of the 2021 ACM Workshop on Intelligent Cross-Data Analysis and Retrieval最新文献

英文 中文
Cross-Modal Deep Neural Networks based Smartphone Authentication for Intelligent Things System 基于跨模态深度神经网络的智能手机认证
Tran Anh Khoa, Dinh Nguyen The Truong, Duc Ngoc Minh Dang
Nowadays, identity authentication technology, including biometric identification features such as iris and fingerprints, plays an essential role in the safety of intelligent devices. However, it cannot implement real-time and continuous identification of user identity. This paper presents a framework for user authentication from motion signals such as accelerometers and gyroscope signals powered received from smartphones. The proposed innovation scheme including i) a data preprocessing, ii) a novel feature extraction and authentication scheme based on a cross-modal deep neural network by applying a time-distributed Convolutional Neural Network (CNN), and Long Short-Term Memory (LSTM) models. The experimental results of the proposed scheme show the advantage of our approach against methods.
如今,包括虹膜、指纹等生物特征识别在内的身份认证技术对智能设备的安全起着至关重要的作用。但是,它不能实现实时、连续的用户身份识别。本文提出了一个从运动信号(如从智能手机接收的加速度计和陀螺仪信号)中进行用户认证的框架。提出的创新方案包括i)数据预处理,ii)基于时间分布卷积神经网络(CNN)和长短期记忆(LSTM)模型的跨模态深度神经网络特征提取和认证方案。实验结果表明,该方法相对于其他方法具有一定的优越性。
{"title":"Cross-Modal Deep Neural Networks based Smartphone Authentication for Intelligent Things System","authors":"Tran Anh Khoa, Dinh Nguyen The Truong, Duc Ngoc Minh Dang","doi":"10.1145/3463944.3469101","DOIUrl":"https://doi.org/10.1145/3463944.3469101","url":null,"abstract":"Nowadays, identity authentication technology, including biometric identification features such as iris and fingerprints, plays an essential role in the safety of intelligent devices. However, it cannot implement real-time and continuous identification of user identity. This paper presents a framework for user authentication from motion signals such as accelerometers and gyroscope signals powered received from smartphones. The proposed innovation scheme including i) a data preprocessing, ii) a novel feature extraction and authentication scheme based on a cross-modal deep neural network by applying a time-distributed Convolutional Neural Network (CNN), and Long Short-Term Memory (LSTM) models. The experimental results of the proposed scheme show the advantage of our approach against methods.","PeriodicalId":394510,"journal":{"name":"Proceedings of the 2021 ACM Workshop on Intelligent Cross-Data Analysis and Retrieval","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115196436","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Two-Faced Humans on Twitter and Facebook: Harvesting Social Multimedia for Human Personality Profiling Twitter和Facebook上的双面人:收集社交多媒体用于人类个性分析
Qi Yang, Aleksandr Farseev, A. Filchenkov
Human personality traits are the key drivers behind our decision-making, influencing our life path on a daily basis. Inference of personality traits, such as Myers-Briggs Personality Type, as well as an understanding of dependencies between personality traits and users' behavior on various social media platforms is of crucial importance to modern research and industry applications. The emergence of diverse and cross-purpose social media avenues makes it possible to perform user personality profiling automatically and efficiently based on data represented across multiple data modalities. However, the research efforts on personality profiling from multi-source multi-modal social media data are relatively sparse, and the level of impact of different social network data on machine learning performance has yet to be comprehensively evaluated. Furthermore, there is not such dataset in the research community to benchmark. This study is one of the first attempts towards bridging such an important research gap. Specifically, in this work, we infer the Myers-Briggs Personality Type indicators, by applying a novel multi-view fusion framework, called "PERS" and comparing the performance results not just across data modalities but also with respect to different social network data sources. Our experimental results demonstrate the PERS's ability to learn from multi-view data for personality profiling by efficiently leveraging on the significantly different data arriving from diverse social multimedia sources. We have also found that the selection of a machine learning approach is of crucial importance when choosing social network data sources and that people tend to reveal multiple facets of their personality in different social media avenues. Our released social multimedia dataset facilitates future research on this direction.
人的性格特征是我们决策背后的关键驱动力,影响着我们每天的生活道路。对Myers-Briggs人格类型等人格特征的推断,以及对人格特征与用户在各种社交媒体平台上的行为之间的依赖关系的理解,对于现代研究和行业应用至关重要。多样化和跨用途的社交媒体渠道的出现使得基于跨多种数据模式表示的数据自动有效地执行用户个性分析成为可能。然而,基于多源多模态社交媒体数据的人格分析研究相对较少,不同社交网络数据对机器学习性能的影响程度尚未得到全面评估。此外,在研究界没有这样的数据集来基准。这项研究是试图弥合这一重要研究鸿沟的首次尝试之一。具体来说,在这项工作中,我们通过应用一种新的多视角融合框架(称为“PERS”)来推断迈尔斯-布里格斯人格类型指标,并比较了不同数据模式以及不同社交网络数据源的表现结果。我们的实验结果表明,通过有效地利用来自不同社交多媒体来源的显著不同的数据,PERS能够从多视图数据中学习人格分析。我们还发现,在选择社交网络数据源时,机器学习方法的选择至关重要,人们倾向于在不同的社交媒体渠道中揭示自己个性的多个方面。我们发布的社交多媒体数据集有助于这一方向的未来研究。
{"title":"Two-Faced Humans on Twitter and Facebook: Harvesting Social Multimedia for Human Personality Profiling","authors":"Qi Yang, Aleksandr Farseev, A. Filchenkov","doi":"10.1145/3463944.3469270","DOIUrl":"https://doi.org/10.1145/3463944.3469270","url":null,"abstract":"Human personality traits are the key drivers behind our decision-making, influencing our life path on a daily basis. Inference of personality traits, such as Myers-Briggs Personality Type, as well as an understanding of dependencies between personality traits and users' behavior on various social media platforms is of crucial importance to modern research and industry applications. The emergence of diverse and cross-purpose social media avenues makes it possible to perform user personality profiling automatically and efficiently based on data represented across multiple data modalities. However, the research efforts on personality profiling from multi-source multi-modal social media data are relatively sparse, and the level of impact of different social network data on machine learning performance has yet to be comprehensively evaluated. Furthermore, there is not such dataset in the research community to benchmark. This study is one of the first attempts towards bridging such an important research gap. Specifically, in this work, we infer the Myers-Briggs Personality Type indicators, by applying a novel multi-view fusion framework, called \"PERS\" and comparing the performance results not just across data modalities but also with respect to different social network data sources. Our experimental results demonstrate the PERS's ability to learn from multi-view data for personality profiling by efficiently leveraging on the significantly different data arriving from diverse social multimedia sources. We have also found that the selection of a machine learning approach is of crucial importance when choosing social network data sources and that people tend to reveal multiple facets of their personality in different social media avenues. Our released social multimedia dataset facilitates future research on this direction.","PeriodicalId":394510,"journal":{"name":"Proceedings of the 2021 ACM Workshop on Intelligent Cross-Data Analysis and Retrieval","volume":"145 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122141632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
ST-HOI: A Spatial-Temporal Baseline for Human-Object Interaction Detection in Videos ST-HOI:视频中人-物交互检测的时空基线
Meng-Jiun Chiou, Chun-Yu Liao, Li-Wei Wang, Roger Zimmermann, Jiashi Feng
Detecting human-object interactions (HOI) is an important step toward a comprehensive visual understanding of machines. While detecting non-temporal HOIs (e.g., sitting on a chair) from static images is feasible, it is unlikely even for humans to guess temporal-related HOIs (e.g., opening/closing a door) from a single video frame, where the neighboring frames play an essential role. However, conventional HOI methods operating on only static images have been used to predict temporal-related interactions, which is essentially guessing without temporal contexts and may lead to sub-optimal performance. In this paper, we bridge this gap by detecting video-based HOIs with explicit temporal information. We first show that a naive temporal-aware variant of a common action detection baseline does not work on video-based HOIs due to a feature-inconsistency issue. We then propose a simple yet effective architecture named Spatial-Temporal HOI Detection (ST-HOI) utilizing temporal information such as human and object trajectories, correctly-localized visual features, and spatial-temporal masking pose features. We construct a new video HOI benchmark dubbed VidHOI where our proposed approach serves as a solid baseline.
检测人机交互(HOI)是对机器进行全面视觉理解的重要一步。虽然从静态图像中检测非时间hoi(例如,坐在椅子上)是可行的,但人类甚至不太可能从单个视频帧中猜测与时间相关的hoi(例如,打开/关闭一扇门),其中相邻帧起着至关重要的作用。然而,传统的仅对静态图像操作的HOI方法已被用于预测与时间相关的交互,这实际上是在没有时间上下文的情况下进行猜测,可能导致性能次优。在本文中,我们通过检测具有明确时间信息的基于视频的hoi来弥补这一差距。我们首先表明,由于特征不一致问题,普通动作检测基线的朴素时间感知变体不适用于基于视频的hoi。然后,我们提出了一个简单而有效的架构,称为时空HOI检测(ST-HOI),利用时间信息,如人和物体轨迹,正确定位的视觉特征和时空掩蔽姿态特征。我们构建了一个新的视频HOI基准,称为VidHOI,其中我们提出的方法作为坚实的基线。
{"title":"ST-HOI: A Spatial-Temporal Baseline for Human-Object Interaction Detection in Videos","authors":"Meng-Jiun Chiou, Chun-Yu Liao, Li-Wei Wang, Roger Zimmermann, Jiashi Feng","doi":"10.1145/3463944.3469097","DOIUrl":"https://doi.org/10.1145/3463944.3469097","url":null,"abstract":"Detecting human-object interactions (HOI) is an important step toward a comprehensive visual understanding of machines. While detecting non-temporal HOIs (e.g., sitting on a chair) from static images is feasible, it is unlikely even for humans to guess temporal-related HOIs (e.g., opening/closing a door) from a single video frame, where the neighboring frames play an essential role. However, conventional HOI methods operating on only static images have been used to predict temporal-related interactions, which is essentially guessing without temporal contexts and may lead to sub-optimal performance. In this paper, we bridge this gap by detecting video-based HOIs with explicit temporal information. We first show that a naive temporal-aware variant of a common action detection baseline does not work on video-based HOIs due to a feature-inconsistency issue. We then propose a simple yet effective architecture named Spatial-Temporal HOI Detection (ST-HOI) utilizing temporal information such as human and object trajectories, correctly-localized visual features, and spatial-temporal masking pose features. We construct a new video HOI benchmark dubbed VidHOI where our proposed approach serves as a solid baseline.","PeriodicalId":394510,"journal":{"name":"Proceedings of the 2021 ACM Workshop on Intelligent Cross-Data Analysis and Retrieval","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129353848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Scattering Transform Based Image Clustering using Projection onto Orthogonal Complement 基于散射变换的正交补投影图像聚类
Angel Villar-Corrales, V. Morgenshtern
In the last few years, large improvements in image clustering have been driven by the recent advances in deep learning. However, due to the architectural complexity of deep neural networks, there is no mathematical theory that explains the success of deep clustering techniques. In this work we introduce Projected-Scattering Spectral Clustering (PSSC), a state-of-the-art, stable, and fast algorithm for image clustering, which is also mathematically interpretable. PSSC includes a novel method to exploit the geometric structure of the scattering transform of small images. This method is inspired by the observation that, in the scattering transform domain, the subspaces formed by the eigenvectors corresponding to the few largest eigenvalues of the data matrices of individual classes are nearly shared among different classes. Therefore, projecting out those shared subspaces reduces the intra-class variability, substantially increasing the clustering performance. We call this method 'Projection onto Orthogonal Complement' (POC). Our experiments demonstrate that PSSC obtains the best results among all shallow clustering algorithms. Moreover, it achieves comparable clustering performance to that of recent state-of-the-art clustering techniques, while reducing the execution time by more than one order of magnitude.
在过去的几年里,深度学习的最新进展推动了图像聚类的巨大改进。然而,由于深度神经网络架构的复杂性,没有数学理论可以解释深度聚类技术的成功。在这项工作中,我们介绍了投影散射光谱聚类(PSSC),这是一种最先进的、稳定的、快速的图像聚类算法,它也是数学上可解释的。PSSC包括一种利用小图像散射变换的几何结构的新方法。该方法的灵感来自于在散射变换域中,不同类别的数据矩阵的几个最大特征值所对应的特征向量所形成的子空间在不同类别之间几乎是共享的。因此,投影出这些共享子空间可以减少类内的可变性,从而大大提高聚类性能。我们称这种方法为“正交补投影”(POC)。实验结果表明,在所有浅聚类算法中,PSSC算法的聚类效果最好。此外,它实现了与最近最先进的集群技术相当的集群性能,同时将执行时间减少了一个数量级以上。
{"title":"Scattering Transform Based Image Clustering using Projection onto Orthogonal Complement","authors":"Angel Villar-Corrales, V. Morgenshtern","doi":"10.1145/3463944.3469098","DOIUrl":"https://doi.org/10.1145/3463944.3469098","url":null,"abstract":"In the last few years, large improvements in image clustering have been driven by the recent advances in deep learning. However, due to the architectural complexity of deep neural networks, there is no mathematical theory that explains the success of deep clustering techniques. In this work we introduce Projected-Scattering Spectral Clustering (PSSC), a state-of-the-art, stable, and fast algorithm for image clustering, which is also mathematically interpretable. PSSC includes a novel method to exploit the geometric structure of the scattering transform of small images. This method is inspired by the observation that, in the scattering transform domain, the subspaces formed by the eigenvectors corresponding to the few largest eigenvalues of the data matrices of individual classes are nearly shared among different classes. Therefore, projecting out those shared subspaces reduces the intra-class variability, substantially increasing the clustering performance. We call this method 'Projection onto Orthogonal Complement' (POC). Our experiments demonstrate that PSSC obtains the best results among all shallow clustering algorithms. Moreover, it achieves comparable clustering performance to that of recent state-of-the-art clustering techniques, while reducing the execution time by more than one order of magnitude.","PeriodicalId":394510,"journal":{"name":"Proceedings of the 2021 ACM Workshop on Intelligent Cross-Data Analysis and Retrieval","volume":"409 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116526808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
Proceedings of the 2021 ACM Workshop on Intelligent Cross-Data Analysis and Retrieval
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1