M. Worring, P. Sajda, S. Santini, David A. Shamma, A. Smeaton, Qiang Yang
What started as a field with an emphasis on optimally serving users' interactive information needs has now become dominated by methods that focus on improving the mean average precision (MAP) of a clearly defined task disconnected from its application. With the pervasiveness of the Internet and all the sensors available to derive contextual user information, it is time to bring the data and the user back together. As a field, we must consider understanding the subjective and descriptive nature of users and understanding data as equally interesting research topics that are both worthy of publication. At the 2012 ACM Second Annual International Conference on Multimedia Retrieval (ICMR) in Hong Kong, a panel took place with Marcel Worring as the moderator and the other authors of this article as the panelists. This panel discussion explored this intriguing question: Where is the user in multimedia retrieval?
{"title":"Where Is the User in Multimedia Retrieval?","authors":"M. Worring, P. Sajda, S. Santini, David A. Shamma, A. Smeaton, Qiang Yang","doi":"10.1109/MMUL.2012.53","DOIUrl":"https://doi.org/10.1109/MMUL.2012.53","url":null,"abstract":"What started as a field with an emphasis on optimally serving users' interactive information needs has now become dominated by methods that focus on improving the mean average precision (MAP) of a clearly defined task disconnected from its application. With the pervasiveness of the Internet and all the sensors available to derive contextual user information, it is time to bring the data and the user back together. As a field, we must consider understanding the subjective and descriptive nature of users and understanding data as equally interesting research topics that are both worthy of publication. At the 2012 ACM Second Annual International Conference on Multimedia Retrieval (ICMR) in Hong Kong, a panel took place with Marcel Worring as the moderator and the other authors of this article as the panelists. This panel discussion explored this intriguing question: Where is the user in multimedia retrieval?","PeriodicalId":290893,"journal":{"name":"IEEE Multim.","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122815927","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The widespread adoption of smartphones equipped with high-quality image-capturing capabilities coupled with the prevalent use of social networks have resulted in an explosive growth of social media content. People now routinely capture the scenes around them and instantly share the multimedia content with their friends over a variety of social networks. The social network functions also ensure that much of this content comes with some form of social annotations. This environment sets the stage for advances in large-scale media research. This special issue hopes to address these challenges. The articles in this issue cover identification of use cases and task design, dataset development, and basic research over existing datasets.
{"title":"Large-Scale Multimedia Data Collections","authors":"B. Huet, Tat-Seng Chua, Alexander Hauptmann","doi":"10.1109/MMUL.2012.35","DOIUrl":"https://doi.org/10.1109/MMUL.2012.35","url":null,"abstract":"The widespread adoption of smartphones equipped with high-quality image-capturing capabilities coupled with the prevalent use of social networks have resulted in an explosive growth of social media content. People now routinely capture the scenes around them and instantly share the multimedia content with their friends over a variety of social networks. The social network functions also ensure that much of this content comes with some form of social annotations. This environment sets the stage for advances in large-scale media research. This special issue hopes to address these challenges. The articles in this issue cover identification of use cases and task design, dataset development, and basic research over existing datasets.","PeriodicalId":290893,"journal":{"name":"IEEE Multim.","volume":"159 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122004846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Authoring of rich media content is not prevalent despite efforts to develop standards, tools, and platforms. Average users prefer to keep it simple. However, growing interest in stylizing content and pinning media objects is putting average users on a new path of creativity that could lead to richer multimedia content.
{"title":"Rich Media, Poor Media","authors":"John R. Smith","doi":"10.1109/MMUL.2012.40","DOIUrl":"https://doi.org/10.1109/MMUL.2012.40","url":null,"abstract":"Authoring of rich media content is not prevalent despite efforts to develop standards, tools, and platforms. Average users prefer to keep it simple. However, growing interest in stylizing content and pinning media objects is putting average users on a new path of creativity that could lead to richer multimedia content.","PeriodicalId":290893,"journal":{"name":"IEEE Multim.","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127223040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Recent advances in 3D depth cameras such as Microsoft Kinect sensors (www.xbox.com/en-US/kinect) have created many opportunities for multimedia computing. The Kinect sensor lets the computer directly sense the third dimension (depth) of the players and the environment. It also understands when users talk, knows who they are when they walk up to it, and can interpret their movements and translate them into a format that developers can use to build new experiences. While the Kinect sensor incorporates several advanced sensing hardware, this article focuses on the vision aspect of the Kinect sensor and its impact beyond the gaming industry.
{"title":"Microsoft Kinect Sensor and Its Effect","authors":"Zhengyou Zhang","doi":"10.1109/MMUL.2012.24","DOIUrl":"https://doi.org/10.1109/MMUL.2012.24","url":null,"abstract":"Recent advances in 3D depth cameras such as Microsoft Kinect sensors (www.xbox.com/en-US/kinect) have created many opportunities for multimedia computing. The Kinect sensor lets the computer directly sense the third dimension (depth) of the players and the environment. It also understands when users talk, knows who they are when they walk up to it, and can interpret their movements and translate them into a format that developers can use to build new experiences. While the Kinect sensor incorporates several advanced sensing hardware, this article focuses on the vision aspect of the Kinect sensor and its impact beyond the gaming industry.","PeriodicalId":290893,"journal":{"name":"IEEE Multim.","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114394467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Social media provides new opportunities for sharing health-related data online. Although crowdsourcing medical diagnoses is not yet the trend, people are using social media to seek answers and better understand treatments and outcomes as doctors, experts, and patients converge online.
{"title":"A Virtual Opinion","authors":"John R. Smith","doi":"10.1109/MMUL.2012.18","DOIUrl":"https://doi.org/10.1109/MMUL.2012.18","url":null,"abstract":"Social media provides new opportunities for sharing health-related data online. Although crowdsourcing medical diagnoses is not yet the trend, people are using social media to seek answers and better understand treatments and outcomes as doctors, experts, and patients converge online.","PeriodicalId":290893,"journal":{"name":"IEEE Multim.","volume":"242 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125706128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Acknowledging the relevance of developments in the open source software community, the ACM International Conference on Multimedia (or ACM MM) has held its Open Source Software Competition every year since 2004. This competition has steadily grown since its inception, from two finalist entries in 2004 to 14 in 2011. This article reports on the Open Source Competition at ACM MM 2011 and the conference's associated panel discussion. To help communicate their intention and impact, videos of the winning project (OpenIMAJ and ImageTerrier) and the two honorable mentions (ClassX and OpenCast) are included here as Web extras. Video 1: "ClassX" by Sherif A. Halawa, Bernd Girod, and Derek Pang (Stanford University). Video 2: "OpenIMAJ/ImageTerrier" by Jonathon Hare, Sina Samangooei, and David Dupplaw (University of Southampton, UK). Video 3: "OpenCast" by Michelle Ziegmann (University of California at Berkeley) and Olaf A. Schulte (ETH Zurich).
认识到开源软件社区发展的相关性,ACM国际多媒体会议(或ACM MM)自2004年以来每年都举办开源软件竞赛。这项比赛自成立以来稳步增长,从2004年的两名决赛选手到2011年的14名。本文报道了ACM MM 2011上的开源竞赛和会议相关的小组讨论。为了帮助传达他们的意图和影响,获奖项目(OpenIMAJ和ImageTerrier)和两个荣誉提名项目(ClassX和OpenCast)的视频作为Web附加内容包含在这里。视频1:“ClassX”由Sherif A. Halawa, Bernd Girod和Derek Pang(斯坦福大学)制作。视频2:“OpenIMAJ/ImageTerrier”由jonathan Hare, Sina Samangooei和David Dupplaw(英国南安普顿大学)制作。视频3:米歇尔·齐格曼(加州大学伯克利分校)和奥拉夫·a·舒尔特(苏黎世联邦理工学院)的“OpenCast”。
{"title":"Embracing Open Source Multimedia Software","authors":"Monette Velasco, Wei Tsang Ooi","doi":"10.1109/MMUL.2012.16","DOIUrl":"https://doi.org/10.1109/MMUL.2012.16","url":null,"abstract":"Acknowledging the relevance of developments in the open source software community, the ACM International Conference on Multimedia (or ACM MM) has held its Open Source Software Competition every year since 2004. This competition has steadily grown since its inception, from two finalist entries in 2004 to 14 in 2011. This article reports on the Open Source Competition at ACM MM 2011 and the conference's associated panel discussion. To help communicate their intention and impact, videos of the winning project (OpenIMAJ and ImageTerrier) and the two honorable mentions (ClassX and OpenCast) are included here as Web extras. Video 1: \"ClassX\" by Sherif A. Halawa, Bernd Girod, and Derek Pang (Stanford University). Video 2: \"OpenIMAJ/ImageTerrier\" by Jonathon Hare, Sina Samangooei, and David Dupplaw (University of Southampton, UK). Video 3: \"OpenCast\" by Michelle Ziegmann (University of California at Berkeley) and Olaf A. Schulte (ETH Zurich).","PeriodicalId":290893,"journal":{"name":"IEEE Multim.","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121345001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The shift to consumption of news in digital form, as more users access information with mobile and portable devices, is driving a new premium for timeliness and immediacy of news reporting. Increasingly computer-based systems will be able to continuously monitor the pulse of the world through digital means, detecting topics of interest from open, multilingual, multimedia and social information and provide timely, relevant information.
{"title":"Open Channel for News","authors":"John R. Smith","doi":"10.1109/MMUL.2011.69","DOIUrl":"https://doi.org/10.1109/MMUL.2011.69","url":null,"abstract":"The shift to consumption of news in digital form, as more users access information with mobile and portable devices, is driving a new premium for timeliness and immediacy of news reporting. Increasingly computer-based systems will be able to continuously monitor the pulse of the world through digital means, detecting topics of interest from open, multilingual, multimedia and social information and provide timely, relevant information.","PeriodicalId":290893,"journal":{"name":"IEEE Multim.","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131516355","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The number and density of digital photos taken in cities is making it possible to automatically contextualize the photos and capture the continuous history across places, people, events, and objects of interest.
{"title":"History Made Every Day","authors":"John R. Smith","doi":"10.1109/MMUL.2011.47","DOIUrl":"https://doi.org/10.1109/MMUL.2011.47","url":null,"abstract":"The number and density of digital photos taken in cities is making it possible to automatically contextualize the photos and capture the continuous history across places, people, events, and objects of interest.","PeriodicalId":290893,"journal":{"name":"IEEE Multim.","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126048726","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The past few years have witnessed rapid development and commercialization of visual content identification and search technologies that are concerned with identifying and searching visual content, particularly image and video content, by visual similarities. This special issue contains several research articles covering a diverse range of topics in visual content identification and search.
{"title":"Visual Content Identification and Search","authors":"Jian Lu, Xiansheng Hua, Dong Xu","doi":"10.1109/MMUL.2011.52","DOIUrl":"https://doi.org/10.1109/MMUL.2011.52","url":null,"abstract":"The past few years have witnessed rapid development and commercialization of visual content identification and search technologies that are concerned with identifying and searching visual content, particularly image and video content, by visual similarities. This special issue contains several research articles covering a diverse range of topics in visual content identification and search.","PeriodicalId":290893,"journal":{"name":"IEEE Multim.","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134527663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}