首页 > 最新文献

Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications最新文献

英文 中文
Multimodal Dish Pairing: Predicting Side Dishes to Serve with a Main Dish 多模式菜肴搭配:预测配菜与主菜的搭配
Taichi Nishimura, Katsuhiko Ishiguro, Keita Higuchi, Masaaki Kotera
Planning a food menu is an essential task in our daily lives. We need to plan a menu by considering various perspectives. To reduce the burden when planning a menu, this study first tackles a novel problem of multimodal dish pairing (MDP), i.e., retrieving suitable side dishes given a query main dish. The key challenge of MDP is to learn human subjectivity, i.e., one-to-many relationships of the main and side dishes. However, in general, web resources only include one-to-one manually created pairs of main and side dishes. To tackle this problem, this study assumes that if side dishes are similar to a manually created side dish, they are also acceptable for the query main dish. We then imitate a one-to-many relationship by computing the similarity of side dishes as side dish scores and assigning them to unknown main and side dish pairs. Based on this score, we train a neural network to learn the suitability of the side dishes through learning-to-rank techniques by fully leveraging the multimodal representations of the dishes. During the experiments, we created a dataset by crawling recipes from an online menu site and evaluated the proposed method based on five criteria: retrieval evaluation, overlapping ingredients, overlapping cooking methods, consistency of the dish styles, and human evaluations. Our experiment results show that the proposed method is superior to the baseline in terms of these five criteria. The results of the qualitative analysis further demonstrates that the proposed method can retrieve side dishes suitable for the main dish.
计划食物菜单是我们日常生活中必不可少的任务。我们需要从不同的角度来规划菜单。为了减轻菜单规划的负担,本研究首先解决了一个新的多模式菜肴配对问题,即根据查询的主菜检索合适的配菜。MDP的主要挑战是学习人的主体性,即主菜和小菜的一对多关系。然而,一般来说,网络资源只包括一对一的手工制作的主菜和配菜。为了解决这个问题,本研究假设,如果配菜与手动创建的配菜相似,它们也可以作为查询主菜。然后,我们通过计算配菜作为配菜分数的相似度并将它们分配给未知的主菜和配菜对来模拟一对多关系。基于这个分数,我们训练了一个神经网络,通过充分利用菜肴的多模态表征,通过学习排序技术来学习配菜的适宜性。在实验过程中,我们通过从在线菜单站点抓取食谱创建了一个数据集,并基于五个标准对所提出的方法进行了评估:检索评估、配料重叠、烹饪方法重叠、菜肴风格一致性和人工评估。实验结果表明,本文提出的方法在这五个方面都优于基线。定性分析的结果进一步表明,该方法可以检索出适合主菜的配菜。
{"title":"Multimodal Dish Pairing: Predicting Side Dishes to Serve with a Main Dish","authors":"Taichi Nishimura, Katsuhiko Ishiguro, Keita Higuchi, Masaaki Kotera","doi":"10.1145/3552485.3554934","DOIUrl":"https://doi.org/10.1145/3552485.3554934","url":null,"abstract":"Planning a food menu is an essential task in our daily lives. We need to plan a menu by considering various perspectives. To reduce the burden when planning a menu, this study first tackles a novel problem of multimodal dish pairing (MDP), i.e., retrieving suitable side dishes given a query main dish. The key challenge of MDP is to learn human subjectivity, i.e., one-to-many relationships of the main and side dishes. However, in general, web resources only include one-to-one manually created pairs of main and side dishes. To tackle this problem, this study assumes that if side dishes are similar to a manually created side dish, they are also acceptable for the query main dish. We then imitate a one-to-many relationship by computing the similarity of side dishes as side dish scores and assigning them to unknown main and side dish pairs. Based on this score, we train a neural network to learn the suitability of the side dishes through learning-to-rank techniques by fully leveraging the multimodal representations of the dishes. During the experiments, we created a dataset by crawling recipes from an online menu site and evaluated the proposed method based on five criteria: retrieval evaluation, overlapping ingredients, overlapping cooking methods, consistency of the dish styles, and human evaluations. Our experiment results show that the proposed method is superior to the baseline in terms of these five criteria. The results of the qualitative analysis further demonstrates that the proposed method can retrieve side dishes suitable for the main dish.","PeriodicalId":338126,"journal":{"name":"Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115307646","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Few-shot Food Recognition with Pre-trained Model 基于预训练模型的少镜头食物识别
Yanqi Wu, Xue Song, Jingjing Chen
Food recognition is a challenging task due to the diversity of food. However, conventional training in food recognition networks demands large amounts of labeled images, which is laborious and expensive. In this work, we aim to tackle the challenging few-shot food recognition problem by leveraging the knowledge learning from pre-trained models, e.g., CLIP. Although CLIP has shown a remarkable zero-shot capability on a wide range of vision tasks, it performs poorly in the domain-specific food recognition task. To transfer CLIP's rich prior knowledge, we explore an adapter-based approach to fine-tune CLIP with only a few samples. Thus we combine CLIP's prior knowledge with the new knowledge extracted from the few-shot training set effectively for achieving good performance. Besides, we also design appropriate prompts to facilitate more accurate identification of foods from different cuisines. Experiments demonstrate that our approach achieves quite promising performance on two public food datasets, including VIREO Food-172 and UECFood-256.
由于食物的多样性,食物识别是一项具有挑战性的任务。然而,传统的食物识别网络训练需要大量的标记图像,这既费力又昂贵。在这项工作中,我们的目标是通过利用预训练模型(如CLIP)的知识学习来解决具有挑战性的少量食物识别问题。尽管CLIP在广泛的视觉任务中表现出了显著的零射击能力,但在特定领域的食物识别任务中表现不佳。为了转移CLIP丰富的先验知识,我们探索了一种基于适配器的方法,仅使用少量样本对CLIP进行微调。因此,我们将CLIP的先验知识与从少镜头训练集中提取的新知识有效地结合起来,以获得良好的性能。此外,我们还设计了适当的提示,以便更准确地识别不同菜系的食物。实验表明,我们的方法在两个公共食品数据集(包括VIREO food -172和UECFood-256)上取得了相当有希望的性能。
{"title":"Few-shot Food Recognition with Pre-trained Model","authors":"Yanqi Wu, Xue Song, Jingjing Chen","doi":"10.1145/3552485.3554939","DOIUrl":"https://doi.org/10.1145/3552485.3554939","url":null,"abstract":"Food recognition is a challenging task due to the diversity of food. However, conventional training in food recognition networks demands large amounts of labeled images, which is laborious and expensive. In this work, we aim to tackle the challenging few-shot food recognition problem by leveraging the knowledge learning from pre-trained models, e.g., CLIP. Although CLIP has shown a remarkable zero-shot capability on a wide range of vision tasks, it performs poorly in the domain-specific food recognition task. To transfer CLIP's rich prior knowledge, we explore an adapter-based approach to fine-tune CLIP with only a few samples. Thus we combine CLIP's prior knowledge with the new knowledge extracted from the few-shot training set effectively for achieving good performance. Besides, we also design appropriate prompts to facilitate more accurate identification of foods from different cuisines. Experiments demonstrate that our approach achieves quite promising performance on two public food datasets, including VIREO Food-172 and UECFood-256.","PeriodicalId":338126,"journal":{"name":"Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131397683","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Recipe Recommendation for Balancing Ingredient Preference and Daily Nutrients 平衡配料偏好和每日营养素的配方建议
Sara Ozeki, Masaaki Kotera, Katushiko Ishiguro, Taichi Nishimura, Keita Higuchi
In this work, we propose a recipe recommendation system for daily eating habits based on user preference and nutrient balance. This method prompts user input and allows for the substitution or addition of ingredients while reflecting the user's preferences. The system also considers daily nutrient balance to fill dietary reference intakes such as carbohydrates, protein, and fat. While users select a day's worth of preferred recipes, the system updates the recommendation based on user selection and excess/deficiency predefined nutritional criteria. We run a simulation study to see the performance of the proposed algorithm. With our recipe planning application, we also performed a user study that participants chose a day's worth of recipes with preferred ingredients. The results show that the proposed system helps make better nutrient balance recipes than traditional ingredient-based search. In addition, the participants liked recommendations from the proposed system that improved satisfaction with recipe selection.
在这项工作中,我们提出了一个基于用户偏好和营养平衡的日常饮食习惯食谱推荐系统。此方法提示用户输入,并允许替换或添加成分,同时反映用户的偏好。该系统还考虑每日营养平衡,以填补饮食参考摄入量,如碳水化合物,蛋白质和脂肪。当用户选择一天的首选食谱时,系统会根据用户的选择和预定义的营养标准来更新推荐。我们进行了模拟研究,以观察所提出算法的性能。在我们的食谱计划应用程序中,我们还执行了一个用户研究,让参与者选择一天中使用首选配料的食谱。结果表明,与传统的基于成分的搜索相比,该系统有助于制作更好的营养平衡食谱。此外,参与者喜欢的建议系统,提高了满意度的食谱选择。
{"title":"Recipe Recommendation for Balancing Ingredient Preference and Daily Nutrients","authors":"Sara Ozeki, Masaaki Kotera, Katushiko Ishiguro, Taichi Nishimura, Keita Higuchi","doi":"10.1145/3552485.3554941","DOIUrl":"https://doi.org/10.1145/3552485.3554941","url":null,"abstract":"In this work, we propose a recipe recommendation system for daily eating habits based on user preference and nutrient balance. This method prompts user input and allows for the substitution or addition of ingredients while reflecting the user's preferences. The system also considers daily nutrient balance to fill dietary reference intakes such as carbohydrates, protein, and fat. While users select a day's worth of preferred recipes, the system updates the recommendation based on user selection and excess/deficiency predefined nutritional criteria. We run a simulation study to see the performance of the proposed algorithm. With our recipe planning application, we also performed a user study that participants chose a day's worth of recipes with preferred ingredients. The results show that the proposed system helps make better nutrient balance recipes than traditional ingredient-based search. In addition, the participants liked recommendations from the proposed system that improved satisfaction with recipe selection.","PeriodicalId":338126,"journal":{"name":"Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123811897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CEA++2022 Panel - Toward Building a Global Food Network CEA++2022专题讨论-迈向构建全球食品网络
Yoko Yamakata, S. Mougiakakou, Ramesh C. Jain
How can we create a global food network? Attempts are being made worldwide to lead people to healthier eating habits. They are not always in the academic field but often on a small scale and privately, in hospitals, nursing homes, schools, and various organizations. They may be collecting and manually analyzing data such as recipes and food records. They are precious data, but in many cases, they are never made public. And in academia, dictionaries, corpora, and knowledge graphs are constructed manually at a high cost, but such knowledge is never shared, and another group continues to generate new knowledge. How can we reduce this wasteful work and allow data and knowledge to be shared and leveraged? There are several issues involved in sharing food data. First, food cultures differ from country to country and region to region. Food data produced in one area rarely works as in another. Food data, especially when linked to medical care, is likely to contain private information and must be anonymized when shared. Anonymizing data without losing its intrinsic value is complex, and knowledge sharing must be abandoned in many cases. In addition, food logging is burdensome. Eating takes place every day, multiple times a day. Recording every meal requires a tremendous amount of effort. However, the impact of a single meal on a person's body is minimal, and it is the long-term record that is important in guiding a person to good health. In this panel discussion, we invite Prof. Stavroula Mougiakakou, General Chair of MADiMa22, a workshop co-located with CEA++22, and Prof. Ramesh Jain, the keynote speaker of CEA++22, to discuss the issues raised above. The Moderator, Prof. Yoko Yamakata will make the panel discussion open to all, and participants from MADiMa22 and CEA++22 are also welcome to join the discussion.
我们如何创建一个全球食品网络?世界各地都在努力引导人们养成更健康的饮食习惯。他们并不总是在学术领域,但往往是小规模的和私人的,在医院、养老院、学校和各种组织。他们可能会收集和手动分析食谱和食物记录等数据。它们是宝贵的数据,但在许多情况下,它们从未公开过。在学术界,词典、语料库和知识图谱都是人工构建的,成本很高,但这些知识从不共享,而另一个群体继续产生新的知识。我们如何才能减少这种浪费的工作,并允许数据和知识被共享和利用?共享食品数据涉及几个问题。首先,饮食文化因国家和地区而异。在一个地区生产的食品数据很少能像在另一个地区一样有效。食品数据,特别是与医疗保健相关的数据,可能包含私人信息,共享时必须匿名。在不失去其内在价值的情况下匿名数据是复杂的,在许多情况下必须放弃知识共享。此外,粮食采伐是一项繁重的工作。每天都有吃的,一天好几次。记录每顿饭需要付出巨大的努力。然而,一顿饭对一个人的身体的影响是很小的,长期记录对指导一个人保持健康是很重要的。在本次小组讨论中,我们邀请了MADiMa22(与CEA++22共同举办的研讨会)的总主席Stavroula Mougiakakou教授和CEA++22的主讲人Ramesh Jain教授来讨论上述问题。本次论坛的主持人Yoko Yamakata教授将对所有人开放小组讨论,同时也欢迎来自MADiMa22和CEA++22的与会者参与讨论。
{"title":"CEA++2022 Panel - Toward Building a Global Food Network","authors":"Yoko Yamakata, S. Mougiakakou, Ramesh C. Jain","doi":"10.1145/3552485.3554972","DOIUrl":"https://doi.org/10.1145/3552485.3554972","url":null,"abstract":"How can we create a global food network? Attempts are being made worldwide to lead people to healthier eating habits. They are not always in the academic field but often on a small scale and privately, in hospitals, nursing homes, schools, and various organizations. They may be collecting and manually analyzing data such as recipes and food records. They are precious data, but in many cases, they are never made public. And in academia, dictionaries, corpora, and knowledge graphs are constructed manually at a high cost, but such knowledge is never shared, and another group continues to generate new knowledge. How can we reduce this wasteful work and allow data and knowledge to be shared and leveraged? There are several issues involved in sharing food data. First, food cultures differ from country to country and region to region. Food data produced in one area rarely works as in another. Food data, especially when linked to medical care, is likely to contain private information and must be anonymized when shared. Anonymizing data without losing its intrinsic value is complex, and knowledge sharing must be abandoned in many cases. In addition, food logging is burdensome. Eating takes place every day, multiple times a day. Recording every meal requires a tremendous amount of effort. However, the impact of a single meal on a person's body is minimal, and it is the long-term record that is important in guiding a person to good health. In this panel discussion, we invite Prof. Stavroula Mougiakakou, General Chair of MADiMa22, a workshop co-located with CEA++22, and Prof. Ramesh Jain, the keynote speaker of CEA++22, to discuss the issues raised above. The Moderator, Prof. Yoko Yamakata will make the panel discussion open to all, and participants from MADiMa22 and CEA++22 are also welcome to join the discussion.","PeriodicalId":338126,"journal":{"name":"Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126332196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
ABLE: Aesthetic Box Lunch Editing ABLE:美学盒饭编辑
Yutong Zhou, N. Shimada
This paper proposes an exploratory research that contains a pre-trained ordering recovery model to obtain correct placement sequences from box lunch images, and a generative adversarial network to composite novel box lunch presentations from single item food and generated layouts. Furthermore, we present Bento800, the first cleanly annotated, high-quality, and standardized dataset for aesthetic box lunch presentation generation and other downstream tasks. Bento800 dataset is available at urlhttps://github.com/Yutong-Zhou-cv/Bento800_Dataset.
本文提出了一项探索性研究,该研究包含一个预训练的排序恢复模型,用于从盒饭图像中获得正确的放置序列,以及一个生成对抗网络,用于从单个食物和生成布局中合成新颖的盒饭呈现。此外,我们提出了Bento800,这是第一个用于美学盒饭表示生成和其他下游任务的清晰注释、高质量和标准化数据集。Bento800数据集可在urlhttps://github.com/Yutong-Zhou-cv/Bento800_Dataset获得。
{"title":"ABLE: Aesthetic Box Lunch Editing","authors":"Yutong Zhou, N. Shimada","doi":"10.1145/3552485.3554935","DOIUrl":"https://doi.org/10.1145/3552485.3554935","url":null,"abstract":"This paper proposes an exploratory research that contains a pre-trained ordering recovery model to obtain correct placement sequences from box lunch images, and a generative adversarial network to composite novel box lunch presentations from single item food and generated layouts. Furthermore, we present Bento800, the first cleanly annotated, high-quality, and standardized dataset for aesthetic box lunch presentation generation and other downstream tasks. Bento800 dataset is available at urlhttps://github.com/Yutong-Zhou-cv/Bento800_Dataset.","PeriodicalId":338126,"journal":{"name":"Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications","volume":"97 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130691268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Recipe Recording by Duplicating and Editing Standard Recipe 通过复制和编辑标准配方进行配方记录
Akihisa Ishino, Yoko Yamakata, K. Aizawa
The best way to ascertain the exact nutritional value of a user's food intake is to have the user record the recipe for that food himself/herself. However, writing a recipe from scratch is tedious and impractical. Therefore, we proposed a method that allows users to write their own recipe in a short time by duplicating and editing a standard recipe. We developed a smartphone application and conducted an experiment in which 19 participants were asked to write their own recipes for about 10 food items each. The results showed that the duplication method took 74% of the time compared to writing a recipe from scratch. The number of editing operations was also reduced to 45%. Future work is to construct a dataset of standard recipes that can be rewritten with little editing cost for any person's recipe.
确定用户食物摄入的确切营养价值的最好方法是让用户自己记录食物的配方。然而,从头开始编写食谱既乏味又不切实际。因此,我们提出了一种方法,允许用户通过复制和编辑标准配方,在短时间内编写自己的配方。我们开发了一款智能手机应用程序,并进行了一项实验,要求19名参与者每人写下大约10种食物的食谱。结果表明,与从头开始编写菜谱相比,复制方法只需花费74%的时间。编辑操作的数量也减少到45%。未来的工作是构建一个标准食谱的数据集,可以以很少的编辑成本为任何人的食谱重写。
{"title":"Recipe Recording by Duplicating and Editing Standard Recipe","authors":"Akihisa Ishino, Yoko Yamakata, K. Aizawa","doi":"10.1145/3552485.3554942","DOIUrl":"https://doi.org/10.1145/3552485.3554942","url":null,"abstract":"The best way to ascertain the exact nutritional value of a user's food intake is to have the user record the recipe for that food himself/herself. However, writing a recipe from scratch is tedious and impractical. Therefore, we proposed a method that allows users to write their own recipe in a short time by duplicating and editing a standard recipe. We developed a smartphone application and conducted an experiment in which 19 participants were asked to write their own recipes for about 10 food items each. The results showed that the duplication method took 74% of the time compared to writing a recipe from scratch. The number of editing operations was also reduced to 45%. Future work is to construct a dataset of standard recipes that can be rewritten with little editing cost for any person's recipe.","PeriodicalId":338126,"journal":{"name":"Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123612534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Creating a World Food Atlas 创建世界粮食地图集
Ramesh C. Jain
I face a problem multiple times every day: What am I going to eat, how much, and where? Where can I get enjoyable healthy food? We live in a world where latest geo-spatial information of interest around us is available in the palm of our hand in our smart phone with navigational guidance, if needed. However, the most vital life information related to food remains inaccessible. Food is vital for health and enjoyment by people, society, and planet. However, data, information, and knowledge related to food suffer from inaccessibility, disinformation, and ignorance. A dependable, trusted, accessible, and dynamic source of geo-indexed food data providing culinary, nutritional, and environmental characteristics is essential for guiding wholistic food decisions. A good amount of data and knowledge related to food is already available in different silos. All those silos may be assimilated into a World Food Atlas (WFA) and made available to people to use it for designing food-centered applications, including food recommendation. WFA contains information about location of sources for food ingredients, dishes, recipes, and consumption patterns. All this information may become available through ubiquitous maps. WFA will help in making better decisions for personal, societal, and planetary health. We believe that there is an urgent need and technology is ready to make it happen. Since food varies significantly across even shorter distances and food preparations are dependent on local culture and socio-economic conditions, it is important that local people are involved in creating such an atlas. We have started an open-data World Food Atlas project and are inviting participation of all interested people to contribute. We need people from different area to help populate WFA and use it. The project is in its infancy. We are building a global community that will make this happen. We invite you to participate in this exciting project.
我每天都会多次面对这样的问题:我要吃什么,吃多少,在哪里吃?我在哪里可以吃到美味的健康食品?在我们生活的世界里,我们周围最新的地理空间信息可以在手掌上的智能手机上获得,如果需要的话,还可以使用导航指导。然而,与食物有关的最重要的生命信息仍然无法获得。食物对人类、社会和地球的健康和享受至关重要。然而,与食物有关的数据、信息和知识却难以获取、虚假信息和无知。一个可靠、可信、可访问和动态的地理索引食品数据来源,提供烹饪、营养和环境特征,对于指导整体食品决策至关重要。与粮食有关的大量数据和知识已经在不同的筒仓中获得。所有这些“筒仓”可能会被整合到世界粮食地图集(WFA)中,并提供给人们使用它来设计以食品为中心的应用程序,包括食品推荐。WFA包含有关食品配料、菜肴、食谱和消费模式的来源位置的信息。所有这些信息都可以通过无处不在的地图获得。WFA将有助于为个人、社会和地球健康做出更好的决定。我们相信这是一种迫切的需求,而且技术已经准备好了。由于即使在较短的距离内,食物也有很大差异,而食物的制作取决于当地的文化和社会经济条件,因此让当地人参与制作这样的地图集是很重要的。我们已经启动了一个开放数据的世界粮食地图集项目,并邀请所有感兴趣的人参与贡献。我们需要来自不同地区的人来帮助普及WFA并使用它。这个项目还处于起步阶段。我们正在建立一个能够实现这一目标的全球社区。我们邀请您参与这个令人兴奋的项目。
{"title":"Creating a World Food Atlas","authors":"Ramesh C. Jain","doi":"10.1145/3552485.3552517","DOIUrl":"https://doi.org/10.1145/3552485.3552517","url":null,"abstract":"I face a problem multiple times every day: What am I going to eat, how much, and where? Where can I get enjoyable healthy food? We live in a world where latest geo-spatial information of interest around us is available in the palm of our hand in our smart phone with navigational guidance, if needed. However, the most vital life information related to food remains inaccessible. Food is vital for health and enjoyment by people, society, and planet. However, data, information, and knowledge related to food suffer from inaccessibility, disinformation, and ignorance. A dependable, trusted, accessible, and dynamic source of geo-indexed food data providing culinary, nutritional, and environmental characteristics is essential for guiding wholistic food decisions. A good amount of data and knowledge related to food is already available in different silos. All those silos may be assimilated into a World Food Atlas (WFA) and made available to people to use it for designing food-centered applications, including food recommendation. WFA contains information about location of sources for food ingredients, dishes, recipes, and consumption patterns. All this information may become available through ubiquitous maps. WFA will help in making better decisions for personal, societal, and planetary health. We believe that there is an urgent need and technology is ready to make it happen. Since food varies significantly across even shorter distances and food preparations are dependent on local culture and socio-economic conditions, it is important that local people are involved in creating such an atlas. We have started an open-data World Food Atlas project and are inviting participation of all interested people to contribute. We need people from different area to help populate WFA and use it. The project is in its infancy. We are building a global community that will make this happen. We invite you to participate in this exciting project.","PeriodicalId":338126,"journal":{"name":"Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124416687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Sequential Transformation Information of Ingredients for Fine-Grained Cooking Activity Recognition 用于细粒度烹饪活动识别的食材序列变换信息学习
Atsushi Okamoto, Katsufumi Inoue, M. Yoshioka
The goal of our research is to recognize the fine-grained cooking activities (e.g., dicing or mincing in cutting) in the egocentric videos from the sequential transformation of ingredients that are processed by the camera-wearer; these types of activities are classified according to the state of ingredients after processing, and we often utilize the same cooking utensils and similar motions in such activities. Due to the above conditions, the recognition of such activities is a challenging task in computer vision and multimedia analysis. To tackle this problem, we need to perceive the sequential state transformation of ingredients precisely. In this research, to realize this, we propose a new GAN-based network whose characteristic points are 1) we crop images around the ingredient as a preprocessing to remove the environmental information, 2) we generate intermediate images from the past and future images to obtain the sequential information in the generator network, 3) the adversarial network is employed as a discriminator to classify whether the input image is generated one or not, and 4) we employ the temporally coherent network to check the temporal smoothness of input images and to predict cooking activities by comparing the original sequential images and the generated ones. To investigate the effectiveness of our proposed method, for the first step, we especially focus on "textitcutting activities ". From the experimental results with our originally prepared dataset, in this paper, we report the effectiveness of our proposed method.
我们的研究目标是在以自我为中心的视频中识别精细的烹饪活动(例如,切丁或切碎),这些视频来自于相机佩戴者处理的食材的顺序转换;这些类型的活动是根据原料加工后的状态来分类的,我们经常在这些活动中使用相同的烹饪器具和类似的动作。由于上述条件,这些活动的识别在计算机视觉和多媒体分析中是一项具有挑战性的任务。为了解决这个问题,我们需要精确地感知成分的顺序状态转换。在本研究中,为了实现这一点,我们提出了一种新的基于gan的网络,其特征点是:1)我们裁剪成分周围的图像作为预处理以去除环境信息;2)我们从过去和未来的图像中生成中间图像以获得生成器网络中的顺序信息;3)使用对抗网络作为判别器来分类输入图像是否为生成图像。4)使用时间相干网络检查输入图像的时间平滑性,并通过对比原始序列图像和生成的序列图像来预测烹饪活动。为了研究我们提出的方法的有效性,第一步,我们特别关注“文本切割活动”。根据我们最初准备的数据集的实验结果,在本文中,我们报告了我们提出的方法的有效性。
{"title":"Learning Sequential Transformation Information of Ingredients for Fine-Grained Cooking Activity Recognition","authors":"Atsushi Okamoto, Katsufumi Inoue, M. Yoshioka","doi":"10.1145/3552485.3554940","DOIUrl":"https://doi.org/10.1145/3552485.3554940","url":null,"abstract":"The goal of our research is to recognize the fine-grained cooking activities (e.g., dicing or mincing in cutting) in the egocentric videos from the sequential transformation of ingredients that are processed by the camera-wearer; these types of activities are classified according to the state of ingredients after processing, and we often utilize the same cooking utensils and similar motions in such activities. Due to the above conditions, the recognition of such activities is a challenging task in computer vision and multimedia analysis. To tackle this problem, we need to perceive the sequential state transformation of ingredients precisely. In this research, to realize this, we propose a new GAN-based network whose characteristic points are 1) we crop images around the ingredient as a preprocessing to remove the environmental information, 2) we generate intermediate images from the past and future images to obtain the sequential information in the generator network, 3) the adversarial network is employed as a discriminator to classify whether the input image is generated one or not, and 4) we employ the temporally coherent network to check the temporal smoothness of input images and to predict cooking activities by comparing the original sequential images and the generated ones. To investigate the effectiveness of our proposed method, for the first step, we especially focus on \"textitcutting activities \". From the experimental results with our originally prepared dataset, in this paper, we report the effectiveness of our proposed method.","PeriodicalId":338126,"journal":{"name":"Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications","volume":"272 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122118908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MIAIS: A Multimedia Recipe Dataset with Ingredient Annotation at Each Instructional Step 多媒体配方数据集,每个教学步骤都有配料注释
Yixin Zhang, Yoko Yamakata, Keishi Tajima
In this paper, we introduce a multimedia recipe dataset with annotation of ingredients at every instructional step, named MIAIS (Multimedia recipe dataset with Ingredient Annotation at every Instructional Step). One unique feature of recipe data is that it is usually presented in a sequential and multimedia form. However, few publicly available recipe datasets contain multimedia text-image paired data for every cooking step. Our goal is to construct a recipe dataset that contains sufficient multimedia data and the annotations to them for every cooking step, which is important for many research topics, such as cooking flow graph generation, recipe text generation, and cooking action recognition. MIAIS contains 12,000 recipes; each recipe has 9.13 cooking instruction steps on average, each of which is a tuple of a text description and an image. The text descriptions and images are collected from the NII Cookpad Dataset and Cookpad Image Dataset, respectively. We have already released our annotation data and related information.
在本文中,我们引入了一个在每个教学步骤中都有配料标注的多媒体配方数据集,命名为MIAIS (multimedia recipe dataset with Ingredient annotation at every teaching step)。配方数据的一个独特特性是,它通常以顺序的多媒体形式呈现。然而,很少有公开可用的食谱数据集包含每个烹饪步骤的多媒体文本-图像配对数据。我们的目标是构建一个食谱数据集,该数据集包含足够的多媒体数据以及对每个烹饪步骤的注释,这对于许多研究课题,如烹饪流图生成、食谱文本生成和烹饪动作识别都很重要。MIAIS包含12000个食谱;每个食谱平均有9.13个烹饪指导步骤,每个步骤都是一个由文字描述和图像组成的元组。文本描述和图像分别来自NII Cookpad Dataset和Cookpad Image Dataset。我们已经发布了我们的标注数据和相关信息。
{"title":"MIAIS: A Multimedia Recipe Dataset with Ingredient Annotation at Each Instructional Step","authors":"Yixin Zhang, Yoko Yamakata, Keishi Tajima","doi":"10.1145/3552485.3554938","DOIUrl":"https://doi.org/10.1145/3552485.3554938","url":null,"abstract":"In this paper, we introduce a multimedia recipe dataset with annotation of ingredients at every instructional step, named MIAIS (Multimedia recipe dataset with Ingredient Annotation at every Instructional Step). One unique feature of recipe data is that it is usually presented in a sequential and multimedia form. However, few publicly available recipe datasets contain multimedia text-image paired data for every cooking step. Our goal is to construct a recipe dataset that contains sufficient multimedia data and the annotations to them for every cooking step, which is important for many research topics, such as cooking flow graph generation, recipe text generation, and cooking action recognition. MIAIS contains 12,000 recipes; each recipe has 9.13 cooking instruction steps on average, each of which is a tuple of a text description and an image. The text descriptions and images are collected from the NII Cookpad Dataset and Cookpad Image Dataset, respectively. We have already released our annotation data and related information.","PeriodicalId":338126,"journal":{"name":"Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126297850","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
"Comparable Recipes": A Construction and Analysis of a Dataset of Recipes Described by Different People for the Same Dish “可比食谱”:同一道菜不同人食谱数据集的构建与分析
Rina Kagawa, Rei Miyata, Yoko Yamakata
Recording high-quality textual recipes is effective for documenting food culture. However, comparing the quality of various recipes is difficult because recipe quality might depend on a variety of description styles and dishes. Therefore, we constructed the following "Comparative Recipes" dataset. First, each of the 64 writers described five recipes after watching five home cooking videos. A total of 318 recipes were created. For each dish (video), there were 15.9 recipes on average, and each recipe was described by a different writer. Next, 335 recipe readers evaluated the quality (i.e., the reproducibility and completeness) of each recipe. A morphological analysis that used this dataset revealed that the amount of description per cooking step affects recipe quality. Furthermore, the effects of cooking procedures being integrated into cooking steps on recipe quality tended to be dependent on the reader's skill. The results suggest a need for description support that appropriately integrates cooking procedures into cooking steps according to the skills and preferences of the reader.
记录高质量的食谱文本是记录饮食文化的有效手段。然而,比较各种食谱的质量是困难的,因为食谱的质量可能取决于各种描述风格和菜肴。因此,我们构建了以下“比较食谱”数据集。首先,64位作者在观看了5个家庭烹饪视频后,每人描述了5个食谱。总共创造了318种食谱。对于每道菜(视频),平均有15.9个食谱,每个食谱由不同的作者描述。接下来,335名食谱阅读者评估每个食谱的质量(即再现性和完整性)。使用该数据集的形态学分析显示,每个烹饪步骤的描述量会影响食谱质量。此外,烹饪程序被整合到烹饪步骤对食谱质量的影响往往取决于读者的技能。结果表明需要描述支持,根据读者的技能和偏好,适当地将烹饪过程整合到烹饪步骤中。
{"title":"\"Comparable Recipes\": A Construction and Analysis of a Dataset of Recipes Described by Different People for the Same Dish","authors":"Rina Kagawa, Rei Miyata, Yoko Yamakata","doi":"10.1145/3552485.3554936","DOIUrl":"https://doi.org/10.1145/3552485.3554936","url":null,"abstract":"Recording high-quality textual recipes is effective for documenting food culture. However, comparing the quality of various recipes is difficult because recipe quality might depend on a variety of description styles and dishes. Therefore, we constructed the following \"Comparative Recipes\" dataset. First, each of the 64 writers described five recipes after watching five home cooking videos. A total of 318 recipes were created. For each dish (video), there were 15.9 recipes on average, and each recipe was described by a different writer. Next, 335 recipe readers evaluated the quality (i.e., the reproducibility and completeness) of each recipe. A morphological analysis that used this dataset revealed that the amount of description per cooking step affects recipe quality. Furthermore, the effects of cooking procedures being integrated into cooking steps on recipe quality tended to be dependent on the reader's skill. The results suggest a need for description support that appropriately integrates cooking procedures into cooking steps according to the skills and preferences of the reader.","PeriodicalId":338126,"journal":{"name":"Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications","volume":"33 12","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133784574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 1st International Workshop on Multimedia for Cooking, Eating, and related APPlications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1