首页 > 最新文献

IEEE Transactions on Pattern Analysis and Machine Intelligence最新文献

英文 中文
Unsupervised Dual Deep Hashing with Semantic-Index and Content-Code for Cross-Modal Retrieval 利用语义索引和内容代码的无监督双深度散列技术实现跨模态检索
IF 23.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-24 DOI: 10.1109/tpami.2024.3467130
Bin Zhang, Yue Zhang, Junyu Li, Jiazhou Chen, Tatsuya Akutsu, Yiu-ming Cheung, Hongmin Cai
{"title":"Unsupervised Dual Deep Hashing with Semantic-Index and Content-Code for Cross-Modal Retrieval","authors":"Bin Zhang, Yue Zhang, Junyu Li, Jiazhou Chen, Tatsuya Akutsu, Yiu-ming Cheung, Hongmin Cai","doi":"10.1109/tpami.2024.3467130","DOIUrl":"https://doi.org/10.1109/tpami.2024.3467130","url":null,"abstract":"","PeriodicalId":13426,"journal":{"name":"IEEE Transactions on Pattern Analysis and Machine Intelligence","volume":null,"pages":null},"PeriodicalIF":23.6,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142317357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Weakly Supervised Monocular 3D Object Detection by Spatial-Temporal View Consistency 通过时空视图一致性进行弱监督单目三维物体检测
IF 23.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-24 DOI: 10.1109/tpami.2024.3466915
Wencheng Han, Runzhou Tao, Haibin Ling, Jianbing Shen
{"title":"Weakly Supervised Monocular 3D Object Detection by Spatial-Temporal View Consistency","authors":"Wencheng Han, Runzhou Tao, Haibin Ling, Jianbing Shen","doi":"10.1109/tpami.2024.3466915","DOIUrl":"https://doi.org/10.1109/tpami.2024.3466915","url":null,"abstract":"","PeriodicalId":13426,"journal":{"name":"IEEE Transactions on Pattern Analysis and Machine Intelligence","volume":null,"pages":null},"PeriodicalIF":23.6,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142317286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CoVR-2: Automatic Data Construction for Composed Video Retrieval. CoVR-2:用于合成视频检索的自动数据构建。
IF 23.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-20 DOI: 10.1109/tpami.2024.3463799
Lucas Ventura,Antoine Yang,Cordelia Schmid,Gul Varol
Composed Image Retrieval (CoIR) has recently gained popularity as a task that considers both text and image queries together, to search for relevant images in a database. Most CoIR approaches require manually annotated datasets, comprising image-text-image triplets, where the text describes a modification from the query image to the target image. However, manual curation of CoIR triplets is expensive and prevents scalability. In this work, we instead propose a scalable automatic dataset creation methodology that generates triplets given video-caption pairs, while also expanding the scope of the task to include composed video retrieval (CoVR). To this end, we mine paired videos with a similar caption from a large database, and leverage a large language model to generate the corresponding modification text. Applying this methodology to the extensive WebVid2M collection, we automatically construct our WebVid-CoVR dataset, resulting in 1.6 million triplets. Moreover, we introduce a new benchmark for CoVR with a manually annotated evaluation set, along with baseline results. We further validate that our methodology is equally applicable to image-caption pairs, by generating 3.3 million CoIR training triplets using the Conceptual Captions dataset. Our model builds on BLIP-2 pretraining, adapting it to composed video (or image) retrieval, and incorporates an additional caption retrieval loss to exploit extra supervision beyond the triplet, which is possible since captions are readily available for our training data by design. We provide extensive ablations to analyze the design choices on our new CoVR benchmark. Our experiments also demonstrate that training a CoVR model on our datasets effectively transfers to CoIR, leading to improved state-of-the-art performance in the zero-shot setup on the CIRR, FashionIQ, and CIRCO benchmarks. Our code, datasets, and models are publicly available at https://imagine.enpc.fr/ ventural/covr.
合成图像检索(CoIR)作为一种同时考虑文本和图像查询,在数据库中搜索相关图像的任务,最近越来越受欢迎。大多数 CoIR 方法都需要人工标注数据集,包括图像-文本-图像三元组,其中文本描述了从查询图像到目标图像的修改。然而,人工标注 CoIR 三元组不仅成本高昂,而且不具备可扩展性。在这项工作中,我们提出了一种可扩展的自动数据集创建方法,该方法可根据视频-字幕配对生成三元组,同时还将任务范围扩展到组合视频检索(CoVR)。为此,我们从大型数据库中挖掘具有相似标题的配对视频,并利用大型语言模型生成相应的修改文本。将这一方法应用于广泛的 WebVid2M 数据集,我们自动构建了 WebVid-CoVR 数据集,产生了 160 万个三元组。此外,我们还为 CoVR 引入了一个新的基准,即人工标注的评估集和基准结果。通过使用概念字幕数据集生成 330 万个 CoIR 训练三元组,我们进一步验证了我们的方法同样适用于图像字幕对。我们的模型建立在 BLIP-2 预训练的基础上,使其适用于视频(或图像)检索,并加入了额外的字幕检索损失,以利用三元组之外的额外监督。我们在新的 CoVR 基准上提供了广泛的消减来分析设计选择。我们的实验还证明,在我们的数据集上训练的 CoVR 模型可以有效地转移到 CoIR 上,从而在 CIRR、FashionIQ 和 CIRCO 基准的零镜头设置中提高最先进的性能。我们的代码、数据集和模型可在 https://imagine.enpc.fr/ ventural/covr 上公开获取。
{"title":"CoVR-2: Automatic Data Construction for Composed Video Retrieval.","authors":"Lucas Ventura,Antoine Yang,Cordelia Schmid,Gul Varol","doi":"10.1109/tpami.2024.3463799","DOIUrl":"https://doi.org/10.1109/tpami.2024.3463799","url":null,"abstract":"Composed Image Retrieval (CoIR) has recently gained popularity as a task that considers both text and image queries together, to search for relevant images in a database. Most CoIR approaches require manually annotated datasets, comprising image-text-image triplets, where the text describes a modification from the query image to the target image. However, manual curation of CoIR triplets is expensive and prevents scalability. In this work, we instead propose a scalable automatic dataset creation methodology that generates triplets given video-caption pairs, while also expanding the scope of the task to include composed video retrieval (CoVR). To this end, we mine paired videos with a similar caption from a large database, and leverage a large language model to generate the corresponding modification text. Applying this methodology to the extensive WebVid2M collection, we automatically construct our WebVid-CoVR dataset, resulting in 1.6 million triplets. Moreover, we introduce a new benchmark for CoVR with a manually annotated evaluation set, along with baseline results. We further validate that our methodology is equally applicable to image-caption pairs, by generating 3.3 million CoIR training triplets using the Conceptual Captions dataset. Our model builds on BLIP-2 pretraining, adapting it to composed video (or image) retrieval, and incorporates an additional caption retrieval loss to exploit extra supervision beyond the triplet, which is possible since captions are readily available for our training data by design. We provide extensive ablations to analyze the design choices on our new CoVR benchmark. Our experiments also demonstrate that training a CoVR model on our datasets effectively transfers to CoIR, leading to improved state-of-the-art performance in the zero-shot setup on the CIRR, FashionIQ, and CIRCO benchmarks. Our code, datasets, and models are publicly available at https://imagine.enpc.fr/ ventural/covr.","PeriodicalId":13426,"journal":{"name":"IEEE Transactions on Pattern Analysis and Machine Intelligence","volume":null,"pages":null},"PeriodicalIF":23.6,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142275195","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-sensor Learning Enables Information Transfer across Different Sensory Data and Augments Multi-modality Imaging. 多传感器学习实现了不同感官数据之间的信息传递,并增强了多模态成像能力。
IF 23.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-20 DOI: 10.1109/tpami.2024.3465649
Lingting Zhu,Yizheng Chen,Lianli Liu,Lei Xing,Lequan Yu
Multi-modality imaging is widely used in clinical practice and biomedical research to gain a comprehensive understanding of an imaging subject. Currently, multi-modality imaging is accomplished by post hoc fusion of independently reconstructed images under the guidance of mutual information or spatially registered hardware, which limits the accuracy and utility of multi-modality imaging. Here, we investigate a data-driven multi-modality imaging (DMI) strategy for synergetic imaging of CT and MRI. We reveal two distinct types of features in multi-modality imaging, namely intra- and inter-modality features, and present a multi-sensor learning (MSL) framework to utilize the crossover inter-modality features for augmented multi-modality imaging. The MSL imaging approach breaks down the boundaries of traditional imaging modalities and allows for optimal hybridization of CT and MRI, which maximizes the use of sensory data. We showcase the effectiveness of our DMI strategy through synergetic CT-MRI brain imaging. The principle of DMI is quite general and holds enormous potential for various DMI applications across disciplines.
多模态成像被广泛应用于临床实践和生物医学研究,以全面了解成像对象。目前,多模态成像是在互信息或空间注册硬件的指导下,通过对独立重建的图像进行事后融合来实现的,这限制了多模态成像的准确性和实用性。在这里,我们研究了一种数据驱动的多模态成像(DMI)策略,用于 CT 和 MRI 的协同成像。我们揭示了多模态成像中两种不同类型的特征,即模态内特征和模态间特征,并提出了一个多传感器学习(MSL)框架,利用交叉的模态间特征来增强多模态成像。MSL 成像方法打破了传统成像模式的界限,实现了 CT 和 MRI 的最佳混合,最大限度地利用了感官数据。我们通过 CT-MRI 脑成像协同技术展示了 DMI 策略的有效性。DMI 的原理非常普遍,在各学科的各种 DMI 应用中蕴藏着巨大的潜力。
{"title":"Multi-sensor Learning Enables Information Transfer across Different Sensory Data and Augments Multi-modality Imaging.","authors":"Lingting Zhu,Yizheng Chen,Lianli Liu,Lei Xing,Lequan Yu","doi":"10.1109/tpami.2024.3465649","DOIUrl":"https://doi.org/10.1109/tpami.2024.3465649","url":null,"abstract":"Multi-modality imaging is widely used in clinical practice and biomedical research to gain a comprehensive understanding of an imaging subject. Currently, multi-modality imaging is accomplished by post hoc fusion of independently reconstructed images under the guidance of mutual information or spatially registered hardware, which limits the accuracy and utility of multi-modality imaging. Here, we investigate a data-driven multi-modality imaging (DMI) strategy for synergetic imaging of CT and MRI. We reveal two distinct types of features in multi-modality imaging, namely intra- and inter-modality features, and present a multi-sensor learning (MSL) framework to utilize the crossover inter-modality features for augmented multi-modality imaging. The MSL imaging approach breaks down the boundaries of traditional imaging modalities and allows for optimal hybridization of CT and MRI, which maximizes the use of sensory data. We showcase the effectiveness of our DMI strategy through synergetic CT-MRI brain imaging. The principle of DMI is quite general and holds enormous potential for various DMI applications across disciplines.","PeriodicalId":13426,"journal":{"name":"IEEE Transactions on Pattern Analysis and Machine Intelligence","volume":null,"pages":null},"PeriodicalIF":23.6,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142275199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Continual Learning From a Stream of APIs 从应用程序接口流中不断学习
IF 23.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-20 DOI: 10.1109/tpami.2024.3460871
Enneng Yang, Zhenyi Wang, Li Shen, Nan Yin, Tongliang Liu, Guibing Guo, Xingwei Wang, Dacheng Tao
{"title":"Continual Learning From a Stream of APIs","authors":"Enneng Yang, Zhenyi Wang, Li Shen, Nan Yin, Tongliang Liu, Guibing Guo, Xingwei Wang, Dacheng Tao","doi":"10.1109/tpami.2024.3460871","DOIUrl":"https://doi.org/10.1109/tpami.2024.3460871","url":null,"abstract":"","PeriodicalId":13426,"journal":{"name":"IEEE Transactions on Pattern Analysis and Machine Intelligence","volume":null,"pages":null},"PeriodicalIF":23.6,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142275362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Event-enhanced Snapshot Mosaic Hyperspectral Frame Deblurring. 事件增强快照马赛克高光谱帧去模糊。
IF 23.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-20 DOI: 10.1109/tpami.2024.3465455
Mengyue Geng,Lizhi Wang,Lin Zhu,Wei Zhang,Ruiqin Xiong,Yonghong Tian
Snapshot Mosaic Hyperspectral Cameras (SMHCs) are popular hyperspectral imaging devices for acquiring both color and motion details of scenes. However, the narrow-band spectral filters in SMHCs may negatively impact their motion perception ability, resulting in blurry SMHC frames. In this paper, we propose a hardware-software collaborative approach to address the blurring issue of SMHCs. Our approach involves integrating SMHCs with neuromorphic event cameras for efficient event-enhanced SMHC frame deblurring. To achieve spectral information recovery guided by event signals, we formulate a spectral-aware Event-based Double Integral (sEDI) model that links SMHC frames and events from a spectral perspective, providing principled model design insights. Then, we develop a Diffusion-guided Noise Awareness (DNA) training framework that utilizes diffusion models to learn noise-aware features and promote model robustness towards camera noise. Furthermore, we design an Event-enhanced Hyperspectral frame Deblurring Network (EvHDNet) based on sEDI, which is trained with DNA and features improved spatial-spectral learning and modality interaction for reliable SMHC frame deblurring. Experiments on both synthetic data and real data show that the proposed DNA + EvHDNet outperforms stateof-the-art methods on both spatial and spectral fidelity. The code and dataset will be made publicly available.
快照马赛克高光谱相机(SMHC)是一种流行的高光谱成像设备,用于获取场景的色彩和运动细节。然而,SMHC 中的窄带光谱滤波器可能会对其运动感知能力产生负面影响,从而导致 SMHC 图像模糊。在本文中,我们提出了一种软硬件协同方法来解决 SMHC 的模糊问题。我们的方法涉及将 SMHC 与神经形态事件相机集成,以实现高效的事件增强 SMHC 帧去模糊。为了在事件信号的引导下实现光谱信息恢复,我们提出了光谱感知的基于事件的双积分(sEDI)模型,该模型从光谱角度将 SMHC 帧和事件联系起来,提供了原则性的模型设计见解。然后,我们开发了扩散引导噪声感知(DNA)训练框架,利用扩散模型学习噪声感知特征,提高模型对摄像机噪声的鲁棒性。此外,我们还设计了基于 sEDI 的事件增强型高光谱帧去模糊网络(EvHDNet),该网络使用 DNA 进行训练,具有改进的空间-光谱学习和模态交互功能,可用于可靠的 SMHC 帧去模糊。在合成数据和真实数据上的实验表明,拟议的 DNA + EvHDNet 在空间和光谱保真度上都优于最先进的方法。代码和数据集将公开发布。
{"title":"Event-enhanced Snapshot Mosaic Hyperspectral Frame Deblurring.","authors":"Mengyue Geng,Lizhi Wang,Lin Zhu,Wei Zhang,Ruiqin Xiong,Yonghong Tian","doi":"10.1109/tpami.2024.3465455","DOIUrl":"https://doi.org/10.1109/tpami.2024.3465455","url":null,"abstract":"Snapshot Mosaic Hyperspectral Cameras (SMHCs) are popular hyperspectral imaging devices for acquiring both color and motion details of scenes. However, the narrow-band spectral filters in SMHCs may negatively impact their motion perception ability, resulting in blurry SMHC frames. In this paper, we propose a hardware-software collaborative approach to address the blurring issue of SMHCs. Our approach involves integrating SMHCs with neuromorphic event cameras for efficient event-enhanced SMHC frame deblurring. To achieve spectral information recovery guided by event signals, we formulate a spectral-aware Event-based Double Integral (sEDI) model that links SMHC frames and events from a spectral perspective, providing principled model design insights. Then, we develop a Diffusion-guided Noise Awareness (DNA) training framework that utilizes diffusion models to learn noise-aware features and promote model robustness towards camera noise. Furthermore, we design an Event-enhanced Hyperspectral frame Deblurring Network (EvHDNet) based on sEDI, which is trained with DNA and features improved spatial-spectral learning and modality interaction for reliable SMHC frame deblurring. Experiments on both synthetic data and real data show that the proposed DNA + EvHDNet outperforms stateof-the-art methods on both spatial and spectral fidelity. The code and dataset will be made publicly available.","PeriodicalId":13426,"journal":{"name":"IEEE Transactions on Pattern Analysis and Machine Intelligence","volume":null,"pages":null},"PeriodicalIF":23.6,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142275196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RoBoSS: A Robust, Bounded, Sparse, and Smooth Loss Function for Supervised Learning RoBoSS:用于监督学习的鲁棒、有界、稀疏且平滑的损失函数
IF 23.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-20 DOI: 10.1109/tpami.2024.3465535
Mushir Akhtar, M. Tanveer, Mohd. Arshad
{"title":"RoBoSS: A Robust, Bounded, Sparse, and Smooth Loss Function for Supervised Learning","authors":"Mushir Akhtar, M. Tanveer, Mohd. Arshad","doi":"10.1109/tpami.2024.3465535","DOIUrl":"https://doi.org/10.1109/tpami.2024.3465535","url":null,"abstract":"","PeriodicalId":13426,"journal":{"name":"IEEE Transactions on Pattern Analysis and Machine Intelligence","volume":null,"pages":null},"PeriodicalIF":23.6,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142275575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unveiling the Power of Self-Supervision for Multi-View Multi-Human Association and Tracking 为多视角多人类关联和跟踪揭示自我监督的力量
IF 23.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-19 DOI: 10.1109/tpami.2024.3463966
Wei Feng, Feifan Wang, Ruize Han, Yiyang Gan, Zekun Qian, Junhui Hou, Song Wang
{"title":"Unveiling the Power of Self-Supervision for Multi-View Multi-Human Association and Tracking","authors":"Wei Feng, Feifan Wang, Ruize Han, Yiyang Gan, Zekun Qian, Junhui Hou, Song Wang","doi":"10.1109/tpami.2024.3463966","DOIUrl":"https://doi.org/10.1109/tpami.2024.3463966","url":null,"abstract":"","PeriodicalId":13426,"journal":{"name":"IEEE Transactions on Pattern Analysis and Machine Intelligence","volume":null,"pages":null},"PeriodicalIF":23.6,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142275448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Revisiting Nonlocal Self-Similarity from Continuous Representation 从连续表征再看非局部自相似性
IF 23.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-19 DOI: 10.1109/tpami.2024.3464875
Yisi Luo, Xile Zhao, Deyu Meng
{"title":"Revisiting Nonlocal Self-Similarity from Continuous Representation","authors":"Yisi Luo, Xile Zhao, Deyu Meng","doi":"10.1109/tpami.2024.3464875","DOIUrl":"https://doi.org/10.1109/tpami.2024.3464875","url":null,"abstract":"","PeriodicalId":13426,"journal":{"name":"IEEE Transactions on Pattern Analysis and Machine Intelligence","volume":null,"pages":null},"PeriodicalIF":23.6,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142275450","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
T2TD: Text-3D Generation Model Based on Prior Knowledge Guidance T2TD:基于先验知识指导的文本三维生成模型
IF 23.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-18 DOI: 10.1109/tpami.2024.3463753
Weizhi Nie, Ruidong Chen, Weijie Wang, Bruno Lepri, Nicu Sebe
{"title":"T2TD: Text-3D Generation Model Based on Prior Knowledge Guidance","authors":"Weizhi Nie, Ruidong Chen, Weijie Wang, Bruno Lepri, Nicu Sebe","doi":"10.1109/tpami.2024.3463753","DOIUrl":"https://doi.org/10.1109/tpami.2024.3463753","url":null,"abstract":"","PeriodicalId":13426,"journal":{"name":"IEEE Transactions on Pattern Analysis and Machine Intelligence","volume":null,"pages":null},"PeriodicalIF":23.6,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142245659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Pattern Analysis and Machine Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1