首页 > 最新文献

Proceedings of the 30th ACM International Conference on Multimedia最新文献

英文 中文
PDAS pda
Pub Date : 2022-10-10 DOI: 10.1201/b11663-4
Chao Zhou, Yixuan Ban, Yangchao Zhao, Liang Guo, Bing Yu
{"title":"PDAS","authors":"Chao Zhou, Yixuan Ban, Yangchao Zhao, Liang Guo, Bing Yu","doi":"10.1201/b11663-4","DOIUrl":"https://doi.org/10.1201/b11663-4","url":null,"abstract":"","PeriodicalId":412792,"journal":{"name":"Proceedings of the 30th ACM International Conference on Multimedia","volume":"26 9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114454843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CyclicShift: A Data Augmentation Method For Enriching Data Patterns CyclicShift:一种丰富数据模式的数据增强方法
Pub Date : 2022-10-10 DOI: 10.1145/3503161.3548188
Hui Lu, Xuan Cheng, Wentao Xia, Pan Deng, Minghui Liu, Tianshu Xie, Xiaomin Wang, Meilin Liu
In this paper, we propose a simple yet effective data augmentation strategy, dubbed CyclicShift, to enrich data patterns. The idea is to shift the image in a certain direction and then circularly refill the resultant out-of-frame part to the other side. Compared with previous related methods, Translation, and Shuffle, our proposed method is able to avoid losing pixels of the original image and preserve its semantic information as much as possible. Visually and emprically, we show that our method indeed brings new data patterns and thereby improves the generalization ability as well as the performance of models. Extensive experiments demonstrate our method's effectiveness in image classification and fine-grained recognition over multiple datasets and various network architectures. Furthermore, our method can also be superimposed on other data augmentation methods in a very simple way. CyclicMix, the simultaneous use of CyclicShift and CutMix, hits a new high in most cases. Our code is open-source and available at https://github.com/dejavunHui/CyclicShift.
在本文中,我们提出了一种简单而有效的数据增强策略,称为CyclicShift,以丰富数据模式。这个想法是将图像在某个方向上移动,然后循环地将生成的帧外部分重新填充到另一边。与之前的相关方法(Translation, Shuffle)相比,我们提出的方法既避免了原始图像的像素丢失,又尽可能地保留了原始图像的语义信息。从视觉和经验上表明,我们的方法确实带来了新的数据模式,从而提高了模型的泛化能力和性能。大量的实验证明了我们的方法在多数据集和各种网络架构的图像分类和细粒度识别方面的有效性。此外,我们的方法还可以以一种非常简单的方式叠加在其他数据增强方法上。CyclicMix,同时使用CyclicShift和CutMix,在大多数情况下达到新高。我们的代码是开源的,可以在https://github.com/dejavunHui/CyclicShift上找到。
{"title":"CyclicShift: A Data Augmentation Method For Enriching Data Patterns","authors":"Hui Lu, Xuan Cheng, Wentao Xia, Pan Deng, Minghui Liu, Tianshu Xie, Xiaomin Wang, Meilin Liu","doi":"10.1145/3503161.3548188","DOIUrl":"https://doi.org/10.1145/3503161.3548188","url":null,"abstract":"In this paper, we propose a simple yet effective data augmentation strategy, dubbed CyclicShift, to enrich data patterns. The idea is to shift the image in a certain direction and then circularly refill the resultant out-of-frame part to the other side. Compared with previous related methods, Translation, and Shuffle, our proposed method is able to avoid losing pixels of the original image and preserve its semantic information as much as possible. Visually and emprically, we show that our method indeed brings new data patterns and thereby improves the generalization ability as well as the performance of models. Extensive experiments demonstrate our method's effectiveness in image classification and fine-grained recognition over multiple datasets and various network architectures. Furthermore, our method can also be superimposed on other data augmentation methods in a very simple way. CyclicMix, the simultaneous use of CyclicShift and CutMix, hits a new high in most cases. Our code is open-source and available at https://github.com/dejavunHui/CyclicShift.","PeriodicalId":412792,"journal":{"name":"Proceedings of the 30th ACM International Conference on Multimedia","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114632640","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Tackling Instance-Dependent Label Noise with Dynamic Distribution Calibration 动态分布校准处理实例相关标签噪声
Pub Date : 2022-10-10 DOI: 10.1145/3503161.3547984
Manyi Zhang, Yuxin Ren, Zihao Wang, C. Yuan
Instance-dependent label noise is realistic but rather challenging, where the label-corruption process depends on instances directly. It causes a severe distribution shift between the distributions of training and test data, which impairs the generalization of trained models. Prior works put great effort into tackling the issue. Unfortunately, these works always highly rely on strong assumptions or remain heuristic without theoretical guarantees. In this paper, to address the distribution shift in learning with instance-dependent label noise, a dynamic distribution-calibration strategy is adopted. Specifically, we hypothesize that, before training data are corrupted by label noise, each class conforms to a multivariate Gaussian distribution at the feature level. Label noise produces outliers to shift the Gaussian distribution. During training, to calibrate the shifted distribution, we propose two methods based on the mean and covariance of multivariate Gaussian distribution respectively. The mean-based method works in a recursive dimension-reduction manner for robust mean estimation, which is theoretically guaranteed to train a high-quality model against label noise. The covariance-based method works in a distribution disturbance manner, which is experimentally verified to improve the model robustness. We demonstrate the utility and effectiveness of our methods on datasets with synthetic label noise and real-world unknown noise.
依赖于实例的标签噪声是现实的,但相当具有挑战性,因为标签损坏过程直接依赖于实例。它会导致训练数据和测试数据之间的严重分布偏移,从而影响训练模型的泛化。先前的作品为解决这个问题付出了很大的努力。不幸的是,这些工作总是高度依赖于强有力的假设,或者在没有理论保证的情况下仍然是启发式的。本文采用一种动态分布校正策略,解决了实例相关标签噪声下学习中的分布偏移问题。具体来说,我们假设,在训练数据被标签噪声破坏之前,每个类在特征级别上都符合多元高斯分布。标签噪声产生离群值以移位高斯分布。在训练过程中,为了校正偏移分布,我们分别提出了基于多元高斯分布均值和协方差的两种方法。基于均值的方法以递归降维的方式进行鲁棒均值估计,理论上保证训练出高质量的抗标签噪声模型。基于协方差的方法以分布扰动的方式工作,实验验证了该方法提高了模型的鲁棒性。我们展示了我们的方法在具有合成标签噪声和现实世界未知噪声的数据集上的实用性和有效性。
{"title":"Tackling Instance-Dependent Label Noise with Dynamic Distribution Calibration","authors":"Manyi Zhang, Yuxin Ren, Zihao Wang, C. Yuan","doi":"10.1145/3503161.3547984","DOIUrl":"https://doi.org/10.1145/3503161.3547984","url":null,"abstract":"Instance-dependent label noise is realistic but rather challenging, where the label-corruption process depends on instances directly. It causes a severe distribution shift between the distributions of training and test data, which impairs the generalization of trained models. Prior works put great effort into tackling the issue. Unfortunately, these works always highly rely on strong assumptions or remain heuristic without theoretical guarantees. In this paper, to address the distribution shift in learning with instance-dependent label noise, a dynamic distribution-calibration strategy is adopted. Specifically, we hypothesize that, before training data are corrupted by label noise, each class conforms to a multivariate Gaussian distribution at the feature level. Label noise produces outliers to shift the Gaussian distribution. During training, to calibrate the shifted distribution, we propose two methods based on the mean and covariance of multivariate Gaussian distribution respectively. The mean-based method works in a recursive dimension-reduction manner for robust mean estimation, which is theoretically guaranteed to train a high-quality model against label noise. The covariance-based method works in a distribution disturbance manner, which is experimentally verified to improve the model robustness. We demonstrate the utility and effectiveness of our methods on datasets with synthetic label noise and real-world unknown noise.","PeriodicalId":412792,"journal":{"name":"Proceedings of the 30th ACM International Conference on Multimedia","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117001230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Zero-shot Generalization of Multimodal Dialogue Agents 多模态对话代理的零概率泛化
Pub Date : 2022-10-10 DOI: 10.1145/3503161.3548759
Diogo Tavares
Multimodal conversational agents are an ever expanding field which benefits from the introduction of large language models. Production-ready robust conversational assistants trade breadth of scope for higher accuracy and general dialogue quality. These conversational assistants must be able to maintain the conversation focused, respond appropriately to user requests, maintain a certain level of natural response generation, be robust to out-of-scope and chitchat attempts, and, of course, be accurate in assisting the user in reaching their domain-specific goals. This work discusses data-centric observations, alongside providing research hypothesis for future, and some of my already developed work, to be expanded throughout my PhD.
多模态会话代理是一个不断发展的领域,它得益于大型语言模型的引入。生产就绪的健壮的会话助手用广度换取更高的准确性和一般的对话质量。这些会话助手必须能够保持对话的焦点,适当地响应用户请求,保持一定程度的自然响应生成,对超出范围和闲聊的尝试具有健壮性,当然,还要准确地帮助用户达到特定于领域的目标。这项工作讨论了以数据为中心的观察,同时为未来提供了研究假设,以及我已经开发的一些工作,将在我的博士学位中扩展。
{"title":"Zero-shot Generalization of Multimodal Dialogue Agents","authors":"Diogo Tavares","doi":"10.1145/3503161.3548759","DOIUrl":"https://doi.org/10.1145/3503161.3548759","url":null,"abstract":"Multimodal conversational agents are an ever expanding field which benefits from the introduction of large language models. Production-ready robust conversational assistants trade breadth of scope for higher accuracy and general dialogue quality. These conversational assistants must be able to maintain the conversation focused, respond appropriately to user requests, maintain a certain level of natural response generation, be robust to out-of-scope and chitchat attempts, and, of course, be accurate in assisting the user in reaching their domain-specific goals. This work discusses data-centric observations, alongside providing research hypothesis for future, and some of my already developed work, to be expanded throughout my PhD.","PeriodicalId":412792,"journal":{"name":"Proceedings of the 30th ACM International Conference on Multimedia","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116006558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
DrawMon: A Distributed System for Detection of Atypical Sketch Content in Concurrent Pictionary Games DrawMon:一种用于并发图像猜词游戏中非典型素描内容检测的分布式系统
Pub Date : 2022-10-10 DOI: 10.1145/3503161.3547747
Nikhil Bansal, Kartiki Gupta, Kiruthika Kannan, Sivani Pentapati, R. Sarvadevabhatla
Pictionary, the popular sketch-based guessing game, provides an opportunity to analyze shared goal cooperative game play in restricted communication settings. However, some players occasionally draw atypical sketch content. While such content is occasionally relevant in the game context, it sometimes represents a rule violation and impairs the game experience. To address such situations in a timely and scalable manner, we introduce DrawMon, a novel distributed framework for automatic detection of atypical sketch content in concurrently occurring Pictionary game sessions. We build specialized online interfaces to collect game session data and annotate atypical sketch content, resulting in AtyPict, the first ever atypical sketch content dataset. We use AtyPict to train CanvasNet, a deep neural atypical content detection network. We utilize CanvasNet as a core component of DrawMon. Our analysis of post deployment game session data indicates DrawMon's effectiveness for scalable monitoring and atypical sketch content detection. Beyond Pictionary, our contributions also serve as a design guide for customized atypical content response systems involving shared and interactive whiteboards. Code and datasets are available at https://drawm0n.github.io.
Pictionary是一款流行的基于草图的猜谜游戏,它提供了一个在有限的交流环境中分析共同目标合作游戏玩法的机会。然而,有些玩家偶尔会画出非典型的素描内容。虽然这些内容偶尔会与游戏情境相关,但有时却会违反规则并损害游戏体验。为了及时和可扩展地解决这种情况,我们引入了DrawMon,这是一个新颖的分布式框架,用于自动检测并发发生的Pictionary游戏会话中的非典型素描内容。我们建立了专门的在线界面来收集游戏会话数据并注释非典型草图内容,从而产生了AtyPict,这是有史以来第一个非典型草图内容数据集。我们使用AtyPict来训练CanvasNet,一个深度神经非典型内容检测网络。我们利用CanvasNet作为DrawMon的核心组件。我们对部署后游戏会话数据的分析表明,DrawMon在可扩展监控和非典型草图内容检测方面的有效性。除了Pictionary之外,我们的贡献还可以作为定制的非典型内容响应系统的设计指南,涉及共享和交互式白板。代码和数据集可在https://drawm0n.github.io上获得。
{"title":"DrawMon: A Distributed System for Detection of Atypical Sketch Content in Concurrent Pictionary Games","authors":"Nikhil Bansal, Kartiki Gupta, Kiruthika Kannan, Sivani Pentapati, R. Sarvadevabhatla","doi":"10.1145/3503161.3547747","DOIUrl":"https://doi.org/10.1145/3503161.3547747","url":null,"abstract":"Pictionary, the popular sketch-based guessing game, provides an opportunity to analyze shared goal cooperative game play in restricted communication settings. However, some players occasionally draw atypical sketch content. While such content is occasionally relevant in the game context, it sometimes represents a rule violation and impairs the game experience. To address such situations in a timely and scalable manner, we introduce DrawMon, a novel distributed framework for automatic detection of atypical sketch content in concurrently occurring Pictionary game sessions. We build specialized online interfaces to collect game session data and annotate atypical sketch content, resulting in AtyPict, the first ever atypical sketch content dataset. We use AtyPict to train CanvasNet, a deep neural atypical content detection network. We utilize CanvasNet as a core component of DrawMon. Our analysis of post deployment game session data indicates DrawMon's effectiveness for scalable monitoring and atypical sketch content detection. Beyond Pictionary, our contributions also serve as a design guide for customized atypical content response systems involving shared and interactive whiteboards. Code and datasets are available at https://drawm0n.github.io.","PeriodicalId":412792,"journal":{"name":"Proceedings of the 30th ACM International Conference on Multimedia","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116143681","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Webly Supervised Image Hashing with Lightweight Semantic Transfer Network 基于轻量级语义传输网络的网络监督图像哈希
Pub Date : 2022-10-10 DOI: 10.1145/3503161.3548342
Hui Cui, Lei Zhu, Jingjing Li, Zheng Zhang, Weili Guan
Recent studies have verified the success of deep hashing for efficient image retrieval. However, most existing methods require abundant human labeling data to optimize the large number of involved network parameters, which consequently restricts the scalability of deep image hashing. Alternatively, learning from freely available web images that inherently include rich semantics is a promising strategy. Nevertheless, the domain distribution gap will prevent transferring the semantics involved in the source web images to the target images. Besides, most existing deep image hashing methods suffer from excessive training time to achieve satisfactory performance without explicit supervision. How to efficiently train the deep image hashing network is another important problem that needs to be seriously considered. In this paper, we propose a Webly Supervised Image Hashing (WSIH) with a well-designed lightweight network. Our model enhances the semantics of unsupervised image hashing with the weak supervision from freely available web images, and simultaneously avoids involving over-abundant parameters in the deep network architecture. Particularly, we train a concept prototype learning network on the web images, learning well-trained network parameters and the prototype codes that hold the discriminative semantics of the potential visual concepts in target images. Further, we meticulously design a lightweight siamese network architecture and a dual-level transfer mechanism to efficiently translate the semantics learned from source web images to the target images. Experiments on two widely-tested image datasets show the superiority of the proposed method in both retrieval accuracy and training efficiency compared to state-of-the-art image hashing methods.The source codes of our method are available at: https://github.com/christinecui/WSIH.
最近的研究已经验证了深度哈希在高效图像检索方面的成功。然而,大多数现有方法需要大量的人工标记数据来优化涉及的大量网络参数,从而限制了深度图像哈希的可扩展性。另外,从包含丰富语义的免费网络图像中学习是一种很有前途的策略。然而,域分布差距会阻碍源web图像中涉及的语义向目标图像的传递。此外,现有的大多数深度图像哈希方法在没有明确监督的情况下,训练时间过长,难以达到令人满意的效果。如何有效地训练深度图像哈希网络是需要认真考虑的另一个重要问题。在本文中,我们提出了一个设计良好的轻量级网络的Webly监督图像哈希(WSIH)。我们的模型利用来自可自由获取的web图像的弱监督来增强无监督图像哈希的语义,同时避免在深度网络架构中涉及过多的参数。特别地,我们在网络图像上训练概念原型学习网络,学习训练良好的网络参数和原型代码,这些原型代码包含目标图像中潜在视觉概念的判别语义。此外,我们精心设计了轻量级的暹罗网络架构和双层传输机制,以有效地将从源web图像学习到的语义转换为目标图像。在两个广泛测试的图像数据集上的实验表明,与最先进的图像哈希方法相比,该方法在检索精度和训练效率方面都具有优势。我们的方法的源代码可在:https://github.com/christinecui/WSIH。
{"title":"Webly Supervised Image Hashing with Lightweight Semantic Transfer Network","authors":"Hui Cui, Lei Zhu, Jingjing Li, Zheng Zhang, Weili Guan","doi":"10.1145/3503161.3548342","DOIUrl":"https://doi.org/10.1145/3503161.3548342","url":null,"abstract":"Recent studies have verified the success of deep hashing for efficient image retrieval. However, most existing methods require abundant human labeling data to optimize the large number of involved network parameters, which consequently restricts the scalability of deep image hashing. Alternatively, learning from freely available web images that inherently include rich semantics is a promising strategy. Nevertheless, the domain distribution gap will prevent transferring the semantics involved in the source web images to the target images. Besides, most existing deep image hashing methods suffer from excessive training time to achieve satisfactory performance without explicit supervision. How to efficiently train the deep image hashing network is another important problem that needs to be seriously considered. In this paper, we propose a Webly Supervised Image Hashing (WSIH) with a well-designed lightweight network. Our model enhances the semantics of unsupervised image hashing with the weak supervision from freely available web images, and simultaneously avoids involving over-abundant parameters in the deep network architecture. Particularly, we train a concept prototype learning network on the web images, learning well-trained network parameters and the prototype codes that hold the discriminative semantics of the potential visual concepts in target images. Further, we meticulously design a lightweight siamese network architecture and a dual-level transfer mechanism to efficiently translate the semantics learned from source web images to the target images. Experiments on two widely-tested image datasets show the superiority of the proposed method in both retrieval accuracy and training efficiency compared to state-of-the-art image hashing methods.The source codes of our method are available at: https://github.com/christinecui/WSIH.","PeriodicalId":412792,"journal":{"name":"Proceedings of the 30th ACM International Conference on Multimedia","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123551233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
MONOPOLY 垄断
Pub Date : 2022-10-10 DOI: 10.1145/3503161.3548380
Puneet Mathur, A. Neerkaje, Malika Chhibber, Ramit Sawhney, Fuming Guo, Franck Dernoncourt, Sanghamitra Dutta, Dinesh Manocha
Big business in manufacturing (in the past) and, more recently, giants in the Internet and distribution industry have attracted the attention of both policymakers and economists because of, among other factors, their power to distort competition and to impose conditions on consumers. Market power is therefore a key concept in economics and a central, maybe unwelcome, presence in the economy. On these bases it would be logical to expect full attention to have been paid to the history of this concept and of its causes. But in fact, a gap in the literature seems to exist, and this book aims to fill it. The author’s choice is to do so by looking at the work of four major Italian economists: Vilfredo Pareto, Maffeo Pantaleoni, Antonio De Viti de Marco, and Enrico Barone. In so doing, the book takes on multiple tasks: not only to look for the roots of ideas on market power and competition but also to define these relevant figures and to explore and highlight their more general contribution to the development of economic ideas and to policymaking. The book is organized in five main chapters, completed by an introduction and a final section with some general conclusions. The introduction properly paves the way for the following analysis, putting the reader in the best position to navigate the volume and appreciate the logic behind the connections among the various parts of the book. In the first chapter the author sets the scene for an in-depth analysis of the work of the four Italian economists who are the protagonists of the book by reviewing, with a historical approach, the literature(s) engaging with the concept of monopoly power. Given the complex nature of this idea, the chapter moves along four different paths: the history of formal models of profit maximization under imperfect competition; the history of competition policy; the theory of competition; and the definition of the concept of entry barriers. The following chapter (chapter 2) deals more directly with the contribution by the Italian marginalists, focusing on what looks like one side of the debate on market power: the issue of competition and the conditions that might make it less than perfect. In this, a fundamental distinction is made between “static” lack of competition—resulting from structural barriers to the free movement of actors in the market —and “dynamic” (temporary) situations of limited competition caused by innovations and other factors, a theme central to the so-called classical school of economic thought, too. The view, stressed by Italian Book Reviews / 879
制造业的大企业(过去),以及最近的互联网和分销行业的巨头,吸引了政策制定者和经济学家的注意,原因之一是它们有能力扭曲竞争,并对消费者施加条件。因此,市场力量是经济学中的一个关键概念,也是经济中一个重要的、或许不受欢迎的存在。在这些基础上,期望对这一概念的历史及其原因给予充分注意是合乎逻辑的。但事实上,文学上似乎存在一个空白,这本书旨在填补它。作者的选择是通过考察四位意大利主要经济学家:维尔弗雷多·帕累托、马费奥·潘塔莱奥尼、安东尼奥·德·维蒂·德·马尔科和恩里科·巴罗内的著作来做到这一点。在此过程中,本书承担了多重任务:不仅要寻找市场力量和竞争思想的根源,还要定义这些相关人物,并探索和突出他们对经济思想发展和政策制定的更普遍贡献。本书分为五个主要章节,由引言和最后一节完成,最后一节给出了一些一般性的结论。引言为接下来的分析做了适当的铺垫,让读者处于最佳的位置来浏览这本书,并欣赏书中各个部分之间联系背后的逻辑。在第一章中,作者为深入分析四位意大利经济学家的工作做了铺垫,他们是本书的主角,用历史的方法回顾了与垄断权力概念相关的文献。考虑到这一观点的复杂性,本章将沿着四条不同的路径展开:不完全竞争下利润最大化的正式模型的历史;竞争政策的历史;竞争理论;以及进入壁垒概念的定义。下一章(第二章)更直接地讨论意大利边际主义者的贡献,聚焦于市场力量辩论的一面:竞争问题和可能使其不完美的条件。在这一点上,对“静态的”缺乏竞争——由市场参与者自由流动的结构性障碍造成——和“动态的”(暂时的)由创新和其他因素造成的有限竞争——这也是所谓的古典经济思想学派的核心主题——进行了根本性的区分。意大利书评/ 879强调了这一观点
{"title":"MONOPOLY","authors":"Puneet Mathur, A. Neerkaje, Malika Chhibber, Ramit Sawhney, Fuming Guo, Franck Dernoncourt, Sanghamitra Dutta, Dinesh Manocha","doi":"10.1145/3503161.3548380","DOIUrl":"https://doi.org/10.1145/3503161.3548380","url":null,"abstract":"Big business in manufacturing (in the past) and, more recently, giants in the Internet and distribution industry have attracted the attention of both policymakers and economists because of, among other factors, their power to distort competition and to impose conditions on consumers. Market power is therefore a key concept in economics and a central, maybe unwelcome, presence in the economy. On these bases it would be logical to expect full attention to have been paid to the history of this concept and of its causes. But in fact, a gap in the literature seems to exist, and this book aims to fill it. The author’s choice is to do so by looking at the work of four major Italian economists: Vilfredo Pareto, Maffeo Pantaleoni, Antonio De Viti de Marco, and Enrico Barone. In so doing, the book takes on multiple tasks: not only to look for the roots of ideas on market power and competition but also to define these relevant figures and to explore and highlight their more general contribution to the development of economic ideas and to policymaking. The book is organized in five main chapters, completed by an introduction and a final section with some general conclusions. The introduction properly paves the way for the following analysis, putting the reader in the best position to navigate the volume and appreciate the logic behind the connections among the various parts of the book. In the first chapter the author sets the scene for an in-depth analysis of the work of the four Italian economists who are the protagonists of the book by reviewing, with a historical approach, the literature(s) engaging with the concept of monopoly power. Given the complex nature of this idea, the chapter moves along four different paths: the history of formal models of profit maximization under imperfect competition; the history of competition policy; the theory of competition; and the definition of the concept of entry barriers. The following chapter (chapter 2) deals more directly with the contribution by the Italian marginalists, focusing on what looks like one side of the debate on market power: the issue of competition and the conditions that might make it less than perfect. In this, a fundamental distinction is made between “static” lack of competition—resulting from structural barriers to the free movement of actors in the market —and “dynamic” (temporary) situations of limited competition caused by innovations and other factors, a theme central to the so-called classical school of economic thought, too. The view, stressed by Italian Book Reviews / 879","PeriodicalId":412792,"journal":{"name":"Proceedings of the 30th ACM International Conference on Multimedia","volume":"30 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116834094","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
TransCNN-HAE: Transformer-CNN Hybrid AutoEncoder for Blind Image Inpainting transnn - hae:用于盲图像绘制的变压器- cnn混合自编码器
Pub Date : 2022-10-10 DOI: 10.1145/3503161.3547848
Haoru Zhao, Zhaorui Gu, Bing Zheng, Haiyong Zheng
Blind image inpainting is extremely challenging due to the unknown and multi-property complexity of contamination in different contaminated images. Current mainstream work decomposes blind image inpainting into two stages: mask estimating from the contaminated image and image inpainting based on the estimated mask, and this two-stage solution involves two CNN-based encoder-decoder architectures for estimating and inpainting separately. In this work, we propose a novel one-stage Transformer-CNN Hybrid AutoEncoder (TransCNN-HAE) for blind image inpainting, which intuitively follows the inpainting-then-reconstructing pipeline by leveraging global long-range contextual modeling of Transformer to repair contaminated regions and local short-range contextual modeling of CNN to reconstruct the repaired image. Moreover, a Cross-layer Dissimilarity Prompt (CDP) is devised to accelerate the identifying and inpainting of contaminated regions. Ablation studies validate the efficacy of both TransCNN-HAE and CDP, and extensive experiments on various datasets with multi-property contaminations show that our method achieves state-of-the-art performance with much lower computational cost on blind image inpainting. Our code is available at https://github.com/zhenglab/TransCNN-HAE.
由于不同污染图像中污染的未知性和多属性复杂性,使得图像盲涂非常具有挑战性。目前的主流工作将盲图像补漆分解为两个阶段:从污染图像中估计掩码和基于估计掩码的图像补漆,这个两阶段的解决方案涉及两个基于cnn的编码器-解码器架构,分别用于估计和补漆。在这项工作中,我们提出了一种用于盲图像修复的新型一级变压器-CNN混合自动编码器(TransCNN-HAE),它通过利用变压器的全局远程上下文建模来修复污染区域,利用CNN的局部短程上下文建模来重建修复后的图像,直观地遵循修复-重建的管道。此外,设计了一种跨层不相似提示(CDP),以加快污染区域的识别和涂漆。烧蚀研究验证了TransCNN-HAE和CDP的有效性,并且在具有多属性污染的各种数据集上进行的大量实验表明,我们的方法在盲图像喷漆上实现了最先进的性能,计算成本更低。我们的代码可在https://github.com/zhenglab/TransCNN-HAE上获得。
{"title":"TransCNN-HAE: Transformer-CNN Hybrid AutoEncoder for Blind Image Inpainting","authors":"Haoru Zhao, Zhaorui Gu, Bing Zheng, Haiyong Zheng","doi":"10.1145/3503161.3547848","DOIUrl":"https://doi.org/10.1145/3503161.3547848","url":null,"abstract":"Blind image inpainting is extremely challenging due to the unknown and multi-property complexity of contamination in different contaminated images. Current mainstream work decomposes blind image inpainting into two stages: mask estimating from the contaminated image and image inpainting based on the estimated mask, and this two-stage solution involves two CNN-based encoder-decoder architectures for estimating and inpainting separately. In this work, we propose a novel one-stage Transformer-CNN Hybrid AutoEncoder (TransCNN-HAE) for blind image inpainting, which intuitively follows the inpainting-then-reconstructing pipeline by leveraging global long-range contextual modeling of Transformer to repair contaminated regions and local short-range contextual modeling of CNN to reconstruct the repaired image. Moreover, a Cross-layer Dissimilarity Prompt (CDP) is devised to accelerate the identifying and inpainting of contaminated regions. Ablation studies validate the efficacy of both TransCNN-HAE and CDP, and extensive experiments on various datasets with multi-property contaminations show that our method achieves state-of-the-art performance with much lower computational cost on blind image inpainting. Our code is available at https://github.com/zhenglab/TransCNN-HAE.","PeriodicalId":412792,"journal":{"name":"Proceedings of the 30th ACM International Conference on Multimedia","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123892107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Augmented Dual-Contrastive Aggregation Learning for Unsupervised Visible-Infrared Person Re-Identification 基于增强双对比聚合学习的无监督可见-红外人再识别
Pub Date : 2022-10-10 DOI: 10.1145/3503161.3548198
Bin Yang, Mang Ye, Jun Chen, Zesen Wu
Visible infrared person re-identification (VI-ReID) aims at searching out the corresponding infrared (visible) images from a gallery set captured by other spectrum cameras. Recent works mainly focus on supervised VI-ReID methods that require plenty of cross-modality (visible-infrared) identity labels which are more expensive than the annotations in single-modality person ReID. For the unsupervised learning visible infrared re-identification (USL-VI-ReID), the large cross-modality discrepancies lead to difficulties in generating reliable cross-modality labels and learning modality-invariant features without any annotations. To address this problem, we propose a novel Augmented Dual-Contrastive Aggregation (ADCA) learning framework. Specifically, a dual-path contrastive learning framework with two modality-specific memories is proposed to learn the intra-modality person representation. To associate positive cross-modality identities, we design a cross-modality memory aggregation module with count priority to select highly associated positive samples, and aggregate their corresponding memory features at the cluster level, ensuring that the optimization is explicitly concentrated on the modality-irrelevant perspective. Extensive experiments demonstrate that our proposed ADCA significantly outperforms existing unsupervised methods under various settings, and even surpasses some supervised counterparts, facilitating VI-ReID to real-world deployment. Code is available at https://github.com/yangbincv/ADCA.
可见红外人物再识别(VI-ReID)旨在从其他光谱相机捕获的图库集中搜索出相应的红外(可见)图像。最近的工作主要集中在有监督的VI-ReID方法上,该方法需要大量的跨模态(可见-红外)身份标签,这比单模态人ReID中的注释更昂贵。对于无监督学习可见红外再识别(USL-VI-ReID),较大的跨模态差异导致在没有任何注释的情况下难以生成可靠的跨模态标签和学习模态不变特征。为了解决这个问题,我们提出了一种新的增强双对比聚合(ADCA)学习框架。具体来说,我们提出了一个具有两种特定模态记忆的双路径对比学习框架来学习模态内的人表征。为了关联正的跨模态身份,我们设计了一个具有计数优先级的跨模态记忆聚合模块,以选择高度关联的正样本,并在聚类级别聚合它们相应的记忆特征,确保优化明确地集中在模态无关的角度。大量实验表明,我们提出的ADCA在各种设置下显着优于现有的无监督方法,甚至超过了一些有监督的对应方法,从而促进了VI-ReID的实际部署。代码可从https://github.com/yangbincv/ADCA获得。
{"title":"Augmented Dual-Contrastive Aggregation Learning for Unsupervised Visible-Infrared Person Re-Identification","authors":"Bin Yang, Mang Ye, Jun Chen, Zesen Wu","doi":"10.1145/3503161.3548198","DOIUrl":"https://doi.org/10.1145/3503161.3548198","url":null,"abstract":"Visible infrared person re-identification (VI-ReID) aims at searching out the corresponding infrared (visible) images from a gallery set captured by other spectrum cameras. Recent works mainly focus on supervised VI-ReID methods that require plenty of cross-modality (visible-infrared) identity labels which are more expensive than the annotations in single-modality person ReID. For the unsupervised learning visible infrared re-identification (USL-VI-ReID), the large cross-modality discrepancies lead to difficulties in generating reliable cross-modality labels and learning modality-invariant features without any annotations. To address this problem, we propose a novel Augmented Dual-Contrastive Aggregation (ADCA) learning framework. Specifically, a dual-path contrastive learning framework with two modality-specific memories is proposed to learn the intra-modality person representation. To associate positive cross-modality identities, we design a cross-modality memory aggregation module with count priority to select highly associated positive samples, and aggregate their corresponding memory features at the cluster level, ensuring that the optimization is explicitly concentrated on the modality-irrelevant perspective. Extensive experiments demonstrate that our proposed ADCA significantly outperforms existing unsupervised methods under various settings, and even surpasses some supervised counterparts, facilitating VI-ReID to real-world deployment. Code is available at https://github.com/yangbincv/ADCA.","PeriodicalId":412792,"journal":{"name":"Proceedings of the 30th ACM International Conference on Multimedia","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124053590","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Cartoon-Flow: A Flow-Based Generative Adversarial Network for Arbitrary-Style Photo Cartoonization 卡通流:一个基于流的生成对抗网络,用于任意风格的照片卡通化
Pub Date : 2022-10-10 DOI: 10.1145/3503161.3548094
Jieun Lee, Hyeonwoo Kim, Jong-Chae Shim, Eenjun Hwang
Photo cartoonization aims to convert photos of real-world scenes into cartoon-style images. Recently, generative adversarial network (GAN)-based methods for photo cartoonization have been proposed to generate pleasable cartoonized images. However, as these methods can transfer only learned cartoon styles to photos, they are limited in general-purpose applications where unlearned styles are often required. To address this limitation, an arbitrary style transfer (AST) method that transfers arbitrary artistic style into content images can be used. However, conventional AST methods do not perform satisfactorily in cartoonization for two reasons. First, they cannot capture the unique characteristics of cartoons that differ from common artistic styles. Second, they suffer from content leaks in which the semantic structure of the content is distorted. In this paper, to solve these problems, we propose a novel arbitrary-style photo cartoonization method, Cartoon-Flow. More specifically, we construct a new hybrid GAN with an invertible neural flow generator to effectively preserve content information. In addition, we introduce two new losses for cartoonization: (1) edge-promoting smooth loss to learn the unique characteristics of cartoons with smooth surfaces and clear edges, and (2) line loss to mimic the line drawing of cartoons. Extensive experiments demonstrate that the proposed method outperforms previous methods both quantitatively and qualitatively.
照片卡通化旨在将现实世界场景的照片转换为卡通风格的图像。最近,基于生成对抗网络(GAN)的照片卡通化方法被提出,以生成令人愉悦的卡通化图像。然而,由于这些方法只能将学习过的卡通风格转换为照片,因此它们在通常需要非学习过的风格的通用应用中受到限制。为了解决这个限制,可以使用任意风格转换(AST)方法,将任意艺术风格转换为内容图像。然而,由于两个原因,传统的AST方法在卡通化中不能令人满意地执行。首先,他们无法捕捉到漫画不同于普通艺术风格的独特特征。其次,它们遭受内容泄漏,其中内容的语义结构被扭曲。为了解决这些问题,我们提出了一种新的任意风格的照片卡通化方法——卡通流。更具体地说,我们构建了一个具有可逆神经流生成器的新型混合GAN,以有效地保留内容信息。此外,我们还引入了两种新的卡通化损失:(1)促进边缘的平滑损失,以学习具有光滑表面和清晰边缘的卡通的独特特征;(2)线损失,以模仿卡通的线条绘制。大量的实验表明,该方法在定量和定性上都优于以往的方法。
{"title":"Cartoon-Flow: A Flow-Based Generative Adversarial Network for Arbitrary-Style Photo Cartoonization","authors":"Jieun Lee, Hyeonwoo Kim, Jong-Chae Shim, Eenjun Hwang","doi":"10.1145/3503161.3548094","DOIUrl":"https://doi.org/10.1145/3503161.3548094","url":null,"abstract":"Photo cartoonization aims to convert photos of real-world scenes into cartoon-style images. Recently, generative adversarial network (GAN)-based methods for photo cartoonization have been proposed to generate pleasable cartoonized images. However, as these methods can transfer only learned cartoon styles to photos, they are limited in general-purpose applications where unlearned styles are often required. To address this limitation, an arbitrary style transfer (AST) method that transfers arbitrary artistic style into content images can be used. However, conventional AST methods do not perform satisfactorily in cartoonization for two reasons. First, they cannot capture the unique characteristics of cartoons that differ from common artistic styles. Second, they suffer from content leaks in which the semantic structure of the content is distorted. In this paper, to solve these problems, we propose a novel arbitrary-style photo cartoonization method, Cartoon-Flow. More specifically, we construct a new hybrid GAN with an invertible neural flow generator to effectively preserve content information. In addition, we introduce two new losses for cartoonization: (1) edge-promoting smooth loss to learn the unique characteristics of cartoons with smooth surfaces and clear edges, and (2) line loss to mimic the line drawing of cartoons. Extensive experiments demonstrate that the proposed method outperforms previous methods both quantitatively and qualitatively.","PeriodicalId":412792,"journal":{"name":"Proceedings of the 30th ACM International Conference on Multimedia","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124620853","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
Proceedings of the 30th ACM International Conference on Multimedia
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1