Multi-Modal Colour Extraction Using Deep Learning Techniques

Karthik Kulkarni, Prakash J. Patil, Suvarna G. Kanakaraddi
{"title":"Multi-Modal Colour Extraction Using Deep Learning Techniques","authors":"Karthik Kulkarni, Prakash J. Patil, Suvarna G. Kanakaraddi","doi":"10.1109/ICERECT56837.2022.10060086","DOIUrl":null,"url":null,"abstract":"Multimodal learning research has advanced quickly over the past ten years in a variety of fields, especially computer vision. Due to the growing opportunities of multimodal streaming data and deep learning algorithms, deep multimodal learning is increasingly common. This calls for the development of models that can reliably handle and interpret the multimodal data. Unstructured real-world data, often known as modalities, can naturally assume many different shapes, including both text and images often. Deep learning researchers are still driven by the need to extract useful patterns from this type of data. It is crucial to have well-organized product catalogues to enhance customers' experiences as they explore the plethora of possibilities provided by online marketplaces. The availability of product characteristics like colour or material is a crucial component of it. However, attribute data is frequently erroneous or absent on several of the markets we focus on. Utilizing deep models that have been trained on huge corpora to predict features from unstructured data, such as product descriptions and photographs (referred to as modalities in this study), is one potential approach to solving this issue. To receive a comprehensive rundown of the various multi-modal colour extraction techniques and their advantages, drawbacks and open challenges.","PeriodicalId":205485,"journal":{"name":"2022 Fourth International Conference on Emerging Research in Electronics, Computer Science and Technology (ICERECT)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 Fourth International Conference on Emerging Research in Electronics, Computer Science and Technology (ICERECT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICERECT56837.2022.10060086","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Multimodal learning research has advanced quickly over the past ten years in a variety of fields, especially computer vision. Due to the growing opportunities of multimodal streaming data and deep learning algorithms, deep multimodal learning is increasingly common. This calls for the development of models that can reliably handle and interpret the multimodal data. Unstructured real-world data, often known as modalities, can naturally assume many different shapes, including both text and images often. Deep learning researchers are still driven by the need to extract useful patterns from this type of data. It is crucial to have well-organized product catalogues to enhance customers' experiences as they explore the plethora of possibilities provided by online marketplaces. The availability of product characteristics like colour or material is a crucial component of it. However, attribute data is frequently erroneous or absent on several of the markets we focus on. Utilizing deep models that have been trained on huge corpora to predict features from unstructured data, such as product descriptions and photographs (referred to as modalities in this study), is one potential approach to solving this issue. To receive a comprehensive rundown of the various multi-modal colour extraction techniques and their advantages, drawbacks and open challenges.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
使用深度学习技术的多模态颜色提取
在过去的十年中,多模态学习的研究在各个领域都取得了迅速的进展,尤其是计算机视觉。由于多模态流数据和深度学习算法的机会越来越多,深度多模态学习越来越普遍。这就要求开发能够可靠地处理和解释多模态数据的模型。非结构化的现实世界数据,通常被称为模态,可以自然地呈现许多不同的形状,包括文本和图像。深度学习研究人员仍然被从这类数据中提取有用模式的需求所驱动。当客户探索在线市场提供的大量可能性时,组织良好的产品目录对于增强客户体验至关重要。颜色或材料等产品特性的可用性是其中至关重要的组成部分。然而,在我们关注的几个市场中,属性数据经常是错误的或缺失的。利用在巨大的语料库上训练的深度模型来预测非结构化数据的特征,如产品描述和照片(在本研究中称为模态),是解决这个问题的一种潜在方法。全面介绍各种多模态颜色提取技术及其优缺点和面临的挑战。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Research of Computer Simulation based on Digital Design in the Design Application A Novel Study on EVs Smart Charging Optimization Modelling of Eye Blink Monitoring Mechanism utilizing ML Techniques Performance Evaluation of a Network on Chip Based on Ghz Throughput and Low Power for Streaming Data Transmission on FPGA Way Forward to Digital Society – Digital Transformation of Msmes from Industry 4.0 to Industry 5.0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1