首页 > 最新文献

Companion Proceedings of the Web Conference 2021最新文献

英文 中文
Task Offloading Strategy of Internet of Vehicles Based on Stackelberg Game 基于Stackelberg博弈的车联网任务卸载策略
Pub Date : 2021-04-19 DOI: 10.1145/3442442.3451139
Shuo Xiao, Shengzhi Wang, Zhenzhen Huang, Tianyu Wang, Wei Chen, Guopeng Zhang
Moving vehicles generate a large amount of sensor data every second. To ensure automatic driving in a complex driving environment, it needs to fulfill a large amount of data transmission, storage, and processing in a short time. Real-time perception of traffic, target characteristics, and traffic density are important to achieve safe driving and a stable driving experience. However, it is very difficult to adjust the pricing strategy according to the actual demand of the network. In order to analyze the interaction between task vehicle and service vehicle, the Stackelberg game model is introduced. Considering the communication model, calculation model, optimization objectives, and delay constraints, this paper constructs the utility function of service vehicle and task vehicle based on the Stackelberg game model. Based on the utility function, we can obtain the optimal price strategy of service vehicles and the optimal purchase strategy of task vehicles.
移动的车辆每秒都会产生大量的传感器数据。为了保证复杂驾驶环境下的自动驾驶,需要在短时间内完成大量的数据传输、存储和处理。实时感知交通、目标特征和交通密度对于实现安全驾驶和稳定的驾驶体验至关重要。然而,根据网络的实际需求来调整定价策略是非常困难的。为了分析任务车辆与服务车辆之间的相互作用,引入了Stackelberg博弈模型。考虑通信模型、计算模型、优化目标和时延约束,基于Stackelberg博弈模型构建了服务车和任务车的效用函数。基于效用函数,可以得到服务车的最优价格策略和任务车的最优购买策略。
{"title":"Task Offloading Strategy of Internet of Vehicles Based on Stackelberg Game","authors":"Shuo Xiao, Shengzhi Wang, Zhenzhen Huang, Tianyu Wang, Wei Chen, Guopeng Zhang","doi":"10.1145/3442442.3451139","DOIUrl":"https://doi.org/10.1145/3442442.3451139","url":null,"abstract":"Moving vehicles generate a large amount of sensor data every second. To ensure automatic driving in a complex driving environment, it needs to fulfill a large amount of data transmission, storage, and processing in a short time. Real-time perception of traffic, target characteristics, and traffic density are important to achieve safe driving and a stable driving experience. However, it is very difficult to adjust the pricing strategy according to the actual demand of the network. In order to analyze the interaction between task vehicle and service vehicle, the Stackelberg game model is introduced. Considering the communication model, calculation model, optimization objectives, and delay constraints, this paper constructs the utility function of service vehicle and task vehicle based on the Stackelberg game model. Based on the utility function, we can obtain the optimal price strategy of service vehicles and the optimal purchase strategy of task vehicles.","PeriodicalId":129420,"journal":{"name":"Companion Proceedings of the Web Conference 2021","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124378100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Collocating News Articles with Structured Web Tables✱ 以结构化的网络表对新闻文章进行译者排序
Pub Date : 2021-04-19 DOI: 10.1145/3442442.3452326
Alyssa Lees, Luciano Barbosa, Flip Korn, L. Silva, You Wu, Cong Yu
In today’s news deluge, it can often be overwhelming to understand the significance of a news article or verify the facts within. One approach to address this challenge is to identify relevant data so that crucial statistics or facts can be highlighted for the user to easily digest, and thus improve the user’s comprehension of the news story in a larger context. In this paper, we look toward structured tables on the Web, especially the high quality data tables from Wikipedia, to assist in news understanding. Specifically, we aim to automatically find tables related to a news article. For that, we leverage the content and entities extracted from news articles and their matching tables to fine-tune a Bidirectional Transformers (BERT) model. The resulting model is, therefore, an encoder tailored for article-to-table match. To find the matching tables for a given news article, the fine-tuned BERT model encodes each table in the corpus and the news article into their respective embedding vectors. The tables with the highest cosine similarities to the news article in this new representation space are considered the possible matches. Comprehensive experimental analyses show that the new approach significantly outperforms the baselines over a large, weakly-labeled, dataset obtained from Web click logs as well as a small, crowdsourced, evaluation set. Specifically, our approach achieves near 90% accuracy@5 as opposed to baselines varying between 30% and 64%.
在今天的新闻洪流中,理解一篇新闻文章的意义或核实其中的事实往往是压倒性的。解决这一挑战的一种方法是识别相关数据,以便重要的统计数据或事实可以突出显示,以便用户易于消化,从而提高用户在更大的上下文中对新闻故事的理解。在本文中,我们着眼于网络上的结构化表,特别是来自维基百科的高质量数据表,以帮助新闻理解。具体来说,我们的目标是自动查找与新闻文章相关的表格。为此,我们利用从新闻文章中提取的内容和实体及其匹配表来微调双向变形器(BERT)模型。因此,生成的模型是为条目到表匹配量身定制的编码器。为了找到给定新闻文章的匹配表,经过微调的BERT模型将语料库中的每个表和新闻文章编码到各自的嵌入向量中。在这个新的表示空间中,与新闻文章具有最高余弦相似性的表被认为是可能的匹配。综合实验分析表明,新方法在从Web点击日志中获得的大型弱标记数据集以及小型众包评估集上的表现明显优于基线。具体来说,我们的方法达到了接近90% accuracy@5,而不是在30%到64%之间变化的基线。
{"title":"Collocating News Articles with Structured Web Tables✱","authors":"Alyssa Lees, Luciano Barbosa, Flip Korn, L. Silva, You Wu, Cong Yu","doi":"10.1145/3442442.3452326","DOIUrl":"https://doi.org/10.1145/3442442.3452326","url":null,"abstract":"In today’s news deluge, it can often be overwhelming to understand the significance of a news article or verify the facts within. One approach to address this challenge is to identify relevant data so that crucial statistics or facts can be highlighted for the user to easily digest, and thus improve the user’s comprehension of the news story in a larger context. In this paper, we look toward structured tables on the Web, especially the high quality data tables from Wikipedia, to assist in news understanding. Specifically, we aim to automatically find tables related to a news article. For that, we leverage the content and entities extracted from news articles and their matching tables to fine-tune a Bidirectional Transformers (BERT) model. The resulting model is, therefore, an encoder tailored for article-to-table match. To find the matching tables for a given news article, the fine-tuned BERT model encodes each table in the corpus and the news article into their respective embedding vectors. The tables with the highest cosine similarities to the news article in this new representation space are considered the possible matches. Comprehensive experimental analyses show that the new approach significantly outperforms the baselines over a large, weakly-labeled, dataset obtained from Web click logs as well as a small, crowdsourced, evaluation set. Specifically, our approach achieves near 90% accuracy@5 as opposed to baselines varying between 30% and 64%.","PeriodicalId":129420,"journal":{"name":"Companion Proceedings of the Web Conference 2021","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120859542","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Timing-Driven X-architecture Steiner Minimum Tree Construction Based on Social Learning Multi-Objective Particle Swarm Optimization 基于社会学习多目标粒子群优化的定时驱动x结构Steiner最小树构造
Pub Date : 2021-04-19 DOI: 10.1145/3442442.3451143
Xiaohua Chen, R. Zhou, Genggeng Liu, Zhen Chen, Wenzhong Guo
The construction of timing-driven Steiner minimum tree is a critical issue in VLSI routing design. Meanwhile, since the interconnection model of X-architecture can make full use of routing resources compared to the traditional Manhattan architecture, constructing a Timing-Driven X-architecture Steiner Minimum Tree (TDXSMT) is of great significance to improving routing performance. In this paper, an efficient algorithm based on Social Learning Multi-Objective Particle Swarm Optimization (SLMOPSO) is proposed to construct a TDXSMT with minimizing the maximum source-to-sink pathlength. An X-architecture Prim-Dijkstra model is presented to construct an initial Steiner tree which can optimize both the wirelength and the maximum source-to-sink pathlength. In order to find a better solution, an SLMOPSO method based on the nearest and best select strategy is presented to improve the global exploration capability of the algorithm. Besides, the mutation and crossover operators are utilized to achieve the discrete particle update process, thereby better solving the discrete TDXSMT problem. The experimental results indicate that the proposed algorithm has an excellent trade-off between the wirelength and maximum source-to-sink pathlength of the routing tree and can greatly optimize the timing delay.
时序驱动Steiner最小树的构造是VLSI路由设计中的一个关键问题。同时,由于与传统的Manhattan架构相比,x架构的互连模型可以充分利用路由资源,因此构建时序驱动的x架构斯坦纳最小树(TDXSMT)对提高路由性能具有重要意义。本文提出了一种基于社会学习多目标粒子群优化(SLMOPSO)的高效算法,以最小化源到集的最大路径长度来构造TDXSMT。提出了一种x结构Prim-Dijkstra模型,构造了一棵既能优化无线长度又能优化最大源到汇路径长度的初始斯坦纳树。为了找到更好的解,提出了一种基于最近邻和最优选择策略的SLMOPSO方法,提高了算法的全局搜索能力。此外,利用突变算子和交叉算子实现离散粒子更新过程,从而更好地解决离散TDXSMT问题。实验结果表明,该算法在路由树的路由长度和最大源到汇路径长度之间有很好的权衡,可以极大地优化时延。
{"title":"Timing-Driven X-architecture Steiner Minimum Tree Construction Based on Social Learning Multi-Objective Particle Swarm Optimization","authors":"Xiaohua Chen, R. Zhou, Genggeng Liu, Zhen Chen, Wenzhong Guo","doi":"10.1145/3442442.3451143","DOIUrl":"https://doi.org/10.1145/3442442.3451143","url":null,"abstract":"The construction of timing-driven Steiner minimum tree is a critical issue in VLSI routing design. Meanwhile, since the interconnection model of X-architecture can make full use of routing resources compared to the traditional Manhattan architecture, constructing a Timing-Driven X-architecture Steiner Minimum Tree (TDXSMT) is of great significance to improving routing performance. In this paper, an efficient algorithm based on Social Learning Multi-Objective Particle Swarm Optimization (SLMOPSO) is proposed to construct a TDXSMT with minimizing the maximum source-to-sink pathlength. An X-architecture Prim-Dijkstra model is presented to construct an initial Steiner tree which can optimize both the wirelength and the maximum source-to-sink pathlength. In order to find a better solution, an SLMOPSO method based on the nearest and best select strategy is presented to improve the global exploration capability of the algorithm. Besides, the mutation and crossover operators are utilized to achieve the discrete particle update process, thereby better solving the discrete TDXSMT problem. The experimental results indicate that the proposed algorithm has an excellent trade-off between the wirelength and maximum source-to-sink pathlength of the routing tree and can greatly optimize the timing delay.","PeriodicalId":129420,"journal":{"name":"Companion Proceedings of the Web Conference 2021","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123603378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
aiai at the FinSBD-3 task: Structure Boundary Detection of Noisy Financial Texts in English and French Using Data Augmentation and Hybrid Deep Learning Model FinSBD-3任务:使用数据增强和混合深度学习模型对英语和法语中有噪声的金融文本进行结构边界检测
Pub Date : 2021-04-19 DOI: 10.1145/3442442.3451380
Ke Tian, Hua Chen
Both authors contributed equally to this research. This paper presents the method that we tackled the FinSBD-3 shared task (structure boundary detection) to extract the boundaries of sentences, lists, and items, including structure elements like footer, header, tables from noisy unstructured English and French financial texts. The deep attention model based on word embedding using data augmentation and BERT model named as hybrid deep learning model to detect the sentence, list-item, footer, header, tables boundaries in noisy English and French texts and classify the list-item sentences into list & different item types using deep attention model. The experiment is shown that the proposed method could be an effective solution to deal with the FinSBD-3 shared task. The submitted result ranks first based on the task metrics in the final leader board.
两位作者对这项研究的贡献相同。本文介绍了我们处理FinSBD-3共享任务(结构边界检测)的方法,以从嘈杂的非结构化英语和法语金融文本中提取句子、列表和项目的边界,包括结构元素,如页脚、页眉、表格。采用数据增强的基于词嵌入的深度注意模型和BERT模型作为混合深度学习模型,检测噪声英语和法语文本中的句子、list-item、页脚、页眉、表边界,并使用深度注意模型将list-item句子分类为list &不同的item类型。实验结果表明,该方法是处理FinSBD-3共享任务的有效解决方案。根据任务指标,提交的结果在最终排行榜中排名第一。
{"title":"aiai at the FinSBD-3 task: Structure Boundary Detection of Noisy Financial Texts in English and French Using Data Augmentation and Hybrid Deep Learning Model","authors":"Ke Tian, Hua Chen","doi":"10.1145/3442442.3451380","DOIUrl":"https://doi.org/10.1145/3442442.3451380","url":null,"abstract":"Both authors contributed equally to this research. This paper presents the method that we tackled the FinSBD-3 shared task (structure boundary detection) to extract the boundaries of sentences, lists, and items, including structure elements like footer, header, tables from noisy unstructured English and French financial texts. The deep attention model based on word embedding using data augmentation and BERT model named as hybrid deep learning model to detect the sentence, list-item, footer, header, tables boundaries in noisy English and French texts and classify the list-item sentences into list & different item types using deep attention model. The experiment is shown that the proposed method could be an effective solution to deal with the FinSBD-3 shared task. The submitted result ranks first based on the task metrics in the final leader board.","PeriodicalId":129420,"journal":{"name":"Companion Proceedings of the Web Conference 2021","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124608180","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TAKCO: A Platform for Extracting Novel Facts from Tables TAKCO:从表格中提取新事实的平台
Pub Date : 2021-04-19 DOI: 10.1145/3442442.3458611
B. Kruit, P. Boncz, J. Urbani
Web tables contain a large amount of useful knowledge. Takco is a new large-scale platform designed for extracting facts from tables that can be added to Knowledge Graphs (KGs) like Wikidata. Focusing on achieving high precision, current techniques are biased towards extracting redundant facts, i.e., facts already in the KG. Takco aims to find more novel facts, still at high precision. Our demonstration has two goals. The first one is to illustrate the main features of Takco’s novel interpretation algorithm. The second goal is to show to what extent other state-of-the-art systems are biased towards the extraction of redundant facts using our platform, thus raising awareness on this important problem.
Web表包含了大量有用的知识。Takco是一个新的大型平台,用于从表中提取事实,这些事实可以添加到知识图(KGs)中,如维基数据。专注于实现高精度,目前的技术倾向于提取冗余的事实,即已经在KG中的事实。Takco的目标是找到更多新颖的事实,仍然是高精度的。我们的演示有两个目标。第一部分是说明Takco的新解释算法的主要特点。第二个目标是展示其他最先进的系统在多大程度上倾向于使用我们的平台提取冗余事实,从而提高对这一重要问题的认识。
{"title":"TAKCO: A Platform for Extracting Novel Facts from Tables","authors":"B. Kruit, P. Boncz, J. Urbani","doi":"10.1145/3442442.3458611","DOIUrl":"https://doi.org/10.1145/3442442.3458611","url":null,"abstract":"Web tables contain a large amount of useful knowledge. Takco is a new large-scale platform designed for extracting facts from tables that can be added to Knowledge Graphs (KGs) like Wikidata. Focusing on achieving high precision, current techniques are biased towards extracting redundant facts, i.e., facts already in the KG. Takco aims to find more novel facts, still at high precision. Our demonstration has two goals. The first one is to illustrate the main features of Takco’s novel interpretation algorithm. The second goal is to show to what extent other state-of-the-art systems are biased towards the extraction of redundant facts using our platform, thus raising awareness on this important problem.","PeriodicalId":129420,"journal":{"name":"Companion Proceedings of the Web Conference 2021","volume":"240 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125932354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Fairness beyond “equal”: The Diversity Searcher as a Tool to Detect and Enhance the Representation of Socio-political Actors in News Media 超越“平等”的公平:多样性搜索器作为检测和增强新闻媒体中社会政治行动者代表性的工具
Pub Date : 2021-04-19 DOI: 10.1145/3442442.3452303
Bettina Berendt, Özgür Karadeniz, Stefan Mertens, L. d’Haenens
“Fairness” is a multi-faceted concept that is contested within and across disciplines. In machine learning, it usually denotes some form of equality of measurable outcomes of algorithmic decision making. In this paper, we start from a viewpoint of sociology and media studies, which highlights that to even claim fair treatment, individuals and groups first have to be visible. We draw on a notion and a quantitative measure of diversity that expresses this wider requirement. We used the measure to design and build the Diversity Searcher, a Web-based tool to detect and enhance the representation of socio-political actors in news media. We show how the tool's combination of natural language processing and a rich user interface can help news producers and consumers detect and understand diversity-relevant aspects of representation, which can ultimately contribute to enhancing diversity and fairness in media. We comment on our observation that, through interactions with target users during the construction of the tool, NLP results and interface questions became increasingly important, such that the formal measure of diversity has become a catalyst for functionality, but in itself less important.
“公平”是一个多方面的概念,在学科内部和学科之间都存在争议。在机器学习中,它通常表示算法决策的可测量结果的某种形式的平等。在本文中,我们从社会学和媒体研究的角度出发,强调即使要求公平待遇,个人和群体首先必须是可见的。我们利用多样性的概念和数量来表达这一更广泛的要求。我们使用该指标来设计和构建多样性搜索器,这是一个基于网络的工具,用于检测和增强新闻媒体中社会政治行为者的代表性。我们展示了该工具如何结合自然语言处理和丰富的用户界面,帮助新闻生产者和消费者检测和理解与代表性相关的多样性方面,这最终有助于增强媒体的多样性和公平性。我们评论了我们的观察,通过在工具构建过程中与目标用户的互动,NLP结果和界面问题变得越来越重要,因此多样性的正式衡量已成为功能的催化剂,但本身不那么重要。
{"title":"Fairness beyond “equal”: The Diversity Searcher as a Tool to Detect and Enhance the Representation of Socio-political Actors in News Media","authors":"Bettina Berendt, Özgür Karadeniz, Stefan Mertens, L. d’Haenens","doi":"10.1145/3442442.3452303","DOIUrl":"https://doi.org/10.1145/3442442.3452303","url":null,"abstract":"“Fairness” is a multi-faceted concept that is contested within and across disciplines. In machine learning, it usually denotes some form of equality of measurable outcomes of algorithmic decision making. In this paper, we start from a viewpoint of sociology and media studies, which highlights that to even claim fair treatment, individuals and groups first have to be visible. We draw on a notion and a quantitative measure of diversity that expresses this wider requirement. We used the measure to design and build the Diversity Searcher, a Web-based tool to detect and enhance the representation of socio-political actors in news media. We show how the tool's combination of natural language processing and a rich user interface can help news producers and consumers detect and understand diversity-relevant aspects of representation, which can ultimately contribute to enhancing diversity and fairness in media. We comment on our observation that, through interactions with target users during the construction of the tool, NLP results and interface questions became increasingly important, such that the formal measure of diversity has become a catalyst for functionality, but in itself less important.","PeriodicalId":129420,"journal":{"name":"Companion Proceedings of the Web Conference 2021","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114562815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Tracing the Factoids: the Anatomy of Information Re-organization in Wikipedia Articles 追溯事实:维基百科文章中信息重组的剖析
Pub Date : 2021-04-19 DOI: 10.1145/3442442.3452342
Amit Verma, S. Iyengar
Wikipedia articles are known for their exhaustive knowledge and extensive collaboration. Users perform various tasks that include editing in terms of adding new facts or rectifying some mistakes, looking up new topics, or simply browsing. In this paper, we investigate the impact of gradual edits on the re-positioning and organization of the factual information in Wikipedia articles. Literature shows that in a collaborative system, a set of contributors are responsible for seeking, perceiving, and organizing the information. However, very little is known about the evolution of information organization on Wikipedia articles. Based on our analysis, we show that in a Wikipedia article, the crowd is capable of placing the factual information to its correct position, eventually reducing the knowledge gaps. We also show that the majority of information re-arrangement occurs in the initial stages of the article development and gradually decreases in the later stages. Our findings advance our understanding of the fundamentals of information organization on Wikipedia articles and can have implications for developers aiming to improve the content quality and completeness of Wikipedia articles.
维基百科的文章以其详尽的知识和广泛的合作而闻名。用户执行各种任务,包括添加新事实或纠正一些错误的编辑、查找新主题或只是浏览。在本文中,我们研究了渐进式编辑对维基百科条目中事实信息重新定位和组织的影响。文献表明,在协作系统中,一组贡献者负责寻找、感知和组织信息。然而,人们对维基百科文章中信息组织的演变知之甚少。根据我们的分析,我们发现在维基百科的文章中,大众有能力将事实信息放在正确的位置,最终减少知识差距。我们还发现,大多数信息重排发生在文章开发的初始阶段,并在后期逐渐减少。我们的发现促进了我们对维基百科文章信息组织基本原理的理解,并且可以对旨在提高维基百科文章内容质量和完整性的开发人员产生影响。
{"title":"Tracing the Factoids: the Anatomy of Information Re-organization in Wikipedia Articles","authors":"Amit Verma, S. Iyengar","doi":"10.1145/3442442.3452342","DOIUrl":"https://doi.org/10.1145/3442442.3452342","url":null,"abstract":"Wikipedia articles are known for their exhaustive knowledge and extensive collaboration. Users perform various tasks that include editing in terms of adding new facts or rectifying some mistakes, looking up new topics, or simply browsing. In this paper, we investigate the impact of gradual edits on the re-positioning and organization of the factual information in Wikipedia articles. Literature shows that in a collaborative system, a set of contributors are responsible for seeking, perceiving, and organizing the information. However, very little is known about the evolution of information organization on Wikipedia articles. Based on our analysis, we show that in a Wikipedia article, the crowd is capable of placing the factual information to its correct position, eventually reducing the knowledge gaps. We also show that the majority of information re-arrangement occurs in the initial stages of the article development and gradually decreases in the later stages. Our findings advance our understanding of the fundamentals of information organization on Wikipedia articles and can have implications for developers aiming to improve the content quality and completeness of Wikipedia articles.","PeriodicalId":129420,"journal":{"name":"Companion Proceedings of the Web Conference 2021","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114365933","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Distributed Fog Computing Based on Improved LT codes for Deep Learning in Web of Things 基于改进LT代码的物联网深度学习分布式雾计算
Pub Date : 2021-04-19 DOI: 10.1145/3442442.3451140
Lei Zhang, Jie Liu, Fuquan Zhang, Yunlong Mao
With the rapid development of the Web of Things, there have been a lot of sensors deployed. Advanced knowledge can be achieved by deep learning method and easier integration with open Web standards. A large number of the data generated by sensors required extra processing resources due to the limited resources of the sensors. Due to the limitation of bandwidth or requirement of low latency, it is impossible to transfer such large amounts of data to cloud servers for processing. Thus, the concept of distributed fog computing has been proposed to process such big data into knowledge in real-time. Large scale fog computing system is built using cheap devices, denotes as fog nodes. Therefore, the resiliency to fog node failures should be considered in design of distributed fog computing. LT codes (LTC) have important applications in the design of modern distributed computing, which can reduce the latency of the computing tasks, such as matrix multiplication in deep learning methods. In this paper, we consider that fog nodes may be failure, and an improved LT codes are applied to matrix multiplication of distributed fog computing process to reduce latency. Numerical results show that the improved LTC based scheme can reduce average overhead and degree simultaneously, which reduce the latency and computation complexity of distributed fog computing.
随着物联网的快速发展,已经部署了大量的传感器。高级知识可以通过深度学习方法和更容易地与开放的Web标准集成来实现。由于传感器资源有限,传感器产生的大量数据需要额外的处理资源。由于带宽的限制或低延迟的要求,不可能将如此大量的数据传输到云服务器进行处理。因此,人们提出了分布式雾计算的概念,将这些大数据实时处理为知识。大规模的雾计算系统是用廉价的设备构建的,用雾节点表示。因此,在设计分布式雾计算时应考虑雾节点故障的弹性。LTC代码在现代分布式计算设计中有着重要的应用,它可以减少计算任务的延迟,如深度学习方法中的矩阵乘法。本文考虑雾节点可能失效,将改进的LT码应用于分布式雾计算过程的矩阵乘法中,以减少延迟。数值结果表明,改进的基于LTC的方案可以同时降低平均开销和度,从而降低分布式雾计算的延迟和计算复杂度。
{"title":"Distributed Fog Computing Based on Improved LT codes for Deep Learning in Web of Things","authors":"Lei Zhang, Jie Liu, Fuquan Zhang, Yunlong Mao","doi":"10.1145/3442442.3451140","DOIUrl":"https://doi.org/10.1145/3442442.3451140","url":null,"abstract":"With the rapid development of the Web of Things, there have been a lot of sensors deployed. Advanced knowledge can be achieved by deep learning method and easier integration with open Web standards. A large number of the data generated by sensors required extra processing resources due to the limited resources of the sensors. Due to the limitation of bandwidth or requirement of low latency, it is impossible to transfer such large amounts of data to cloud servers for processing. Thus, the concept of distributed fog computing has been proposed to process such big data into knowledge in real-time. Large scale fog computing system is built using cheap devices, denotes as fog nodes. Therefore, the resiliency to fog node failures should be considered in design of distributed fog computing. LT codes (LTC) have important applications in the design of modern distributed computing, which can reduce the latency of the computing tasks, such as matrix multiplication in deep learning methods. In this paper, we consider that fog nodes may be failure, and an improved LT codes are applied to matrix multiplication of distributed fog computing process to reduce latency. Numerical results show that the improved LTC based scheme can reduce average overhead and degree simultaneously, which reduce the latency and computation complexity of distributed fog computing.","PeriodicalId":129420,"journal":{"name":"Companion Proceedings of the Web Conference 2021","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124455306","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analyzing a Knowledge Graph of Industry 4.0 Standards
Pub Date : 2021-04-19 DOI: 10.1145/3442442.3453542
V. Janev, Maria-Esther Vidal, Kemele M. Endris, Dea Pujić
Realizing smart factories according to the Industry 4.0 vision requires intelligent human-to-machine and machine-to-machine communication. To achieve this goal, components such as actuators, sensors, and cyber-physical systems along with their data, need to be described; moreover, interoperability conflicts arisen from various semantic representations of these components demand also solutions. To empowering communication in smart factories, a variety of standards and standardization frameworks have been proposed. These standards enable the description of the main properties of components, systems, and processes, as well as interactions between them. Standardization frameworks classify, align, and integrate industrial standards according to their purposes and features. Various standardization frameworks have been proposed all over the world by industrial communities, e.g., RAMI4.0 or IICF. While being expressive to categorize existing standards, standardization frameworks may present divergent classifications of the same standard. Mismatches between standard classifications generate semantic interoperability conflicts that negatively impact the effectiveness of communication in smart factories. In this article, we tackle the problem of standard interoperability across different standardization frameworks, and devise a knowledge-driven approach that allows for the description of standards and standardization frameworks into an Industry 4.0 knowledge graph (I40KG). The STO ontology represents properties of standards and standardization frameworks, as well as relationships among them. The I40KG integrates more than 200 standards and four standardization frameworks. To populate the I40KG, the landscape of standards has been analyzed from a semantic perspective and the resulting I40KG represents knowledge expressed in more than 200 industrial related documents including technical reports, research articles, and white papers. Additionally, the I40KG has been linked to existing knowledge graphs and an automated reasoning has been implemented to reveal implicit relations between standards as well as mappings across standardization frameworks. We analyze both the number of discovered relations between standards and the accuracy of these relations. Observed results indicate that both reasoning and linking processes enable for increasing the connectivity in the knowledge graph by up to 80%, whilst up to 96% of the relations can be validated. These outcomes suggest that integrating standards and standardization frameworks into the I40KG enables the resolution of semantic interoperability conflicts, empowering the communication in smart factories.
根据工业4.0的愿景实现智能工厂,需要智能的人机和机器对机器通信。为了实现这一目标,需要对执行器、传感器和网络物理系统等组件及其数据进行描述;此外,这些组件的各种语义表示引起的互操作性冲突也需要解决方案。为了增强智能工厂的通信能力,已经提出了各种标准和标准化框架。这些标准能够描述组件、系统和过程的主要属性,以及它们之间的交互。标准化框架根据其目的和特性对工业标准进行分类、对齐和集成。工业团体在世界各地提出了各种标准化框架,例如RAMI4.0或IICF。标准化框架在表达对现有标准的分类时,可能会对同一标准提出不同的分类。标准分类之间的不匹配会产生语义互操作性冲突,从而对智能工厂中通信的有效性产生负面影响。在本文中,我们解决了跨不同标准化框架的标准互操作性问题,并设计了一种知识驱动的方法,该方法允许将标准和标准化框架描述为工业4.0知识图(I40KG)。STO本体表示标准和标准化框架的属性,以及它们之间的关系。I40KG集成了200多个标准和4个标准化框架。为了填充I40KG,从语义的角度分析了标准的格局,得到的I40KG代表了200多个工业相关文档(包括技术报告、研究文章和白皮书)中表达的知识。此外,I40KG已与现有的知识图谱相关联,并实现了自动推理,以揭示标准之间的隐含关系以及标准化框架之间的映射。我们分析了发现的标准间关系的数量和这些关系的准确性。观察结果表明,推理和链接过程使知识图中的连通性提高了80%,而高达96%的关系可以被验证。这些结果表明,将标准和标准化框架集成到I40KG中可以解决语义互操作性冲突,从而增强智能工厂的通信能力。
{"title":"Analyzing a Knowledge Graph of Industry 4.0 Standards","authors":"V. Janev, Maria-Esther Vidal, Kemele M. Endris, Dea Pujić","doi":"10.1145/3442442.3453542","DOIUrl":"https://doi.org/10.1145/3442442.3453542","url":null,"abstract":"Realizing smart factories according to the Industry 4.0 vision requires intelligent human-to-machine and machine-to-machine communication. To achieve this goal, components such as actuators, sensors, and cyber-physical systems along with their data, need to be described; moreover, interoperability conflicts arisen from various semantic representations of these components demand also solutions. To empowering communication in smart factories, a variety of standards and standardization frameworks have been proposed. These standards enable the description of the main properties of components, systems, and processes, as well as interactions between them. Standardization frameworks classify, align, and integrate industrial standards according to their purposes and features. Various standardization frameworks have been proposed all over the world by industrial communities, e.g., RAMI4.0 or IICF. While being expressive to categorize existing standards, standardization frameworks may present divergent classifications of the same standard. Mismatches between standard classifications generate semantic interoperability conflicts that negatively impact the effectiveness of communication in smart factories. In this article, we tackle the problem of standard interoperability across different standardization frameworks, and devise a knowledge-driven approach that allows for the description of standards and standardization frameworks into an Industry 4.0 knowledge graph (I40KG). The STO ontology represents properties of standards and standardization frameworks, as well as relationships among them. The I40KG integrates more than 200 standards and four standardization frameworks. To populate the I40KG, the landscape of standards has been analyzed from a semantic perspective and the resulting I40KG represents knowledge expressed in more than 200 industrial related documents including technical reports, research articles, and white papers. Additionally, the I40KG has been linked to existing knowledge graphs and an automated reasoning has been implemented to reveal implicit relations between standards as well as mappings across standardization frameworks. We analyze both the number of discovered relations between standards and the accuracy of these relations. Observed results indicate that both reasoning and linking processes enable for increasing the connectivity in the knowledge graph by up to 80%, whilst up to 96% of the relations can be validated. These outcomes suggest that integrating standards and standardization frameworks into the I40KG enables the resolution of semantic interoperability conflicts, empowering the communication in smart factories.","PeriodicalId":129420,"journal":{"name":"Companion Proceedings of the Web Conference 2021","volume":"2015 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127602726","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
AI Principles in Identifying Toxicity in Online Conversation: Keynote at the Third Workshop on Fairness, Accountability, Transparency, Ethics and Society on the Web 识别在线对话中毒性的人工智能原则:在第三届网络公平、问责制、透明度、道德和社会研讨会上的主旨演讲
Pub Date : 2021-04-19 DOI: 10.1145/3442442.3452307
Lucy Vasserman
Jigsaw’s Perspective API aims to protect voices in online conversation by developing and serving machine learning models that identify toxicity text. This talk will share how the team behind Perspective thinks about the issues of Fairness, Accountability, Transparency, Ethics and Society through the lens of Google’s AI Principles. For the Perspective team, building technology that is fair and ethical is a continuous, ongoing effort. The talk will cover concrete strategies the Perspective team has already used to mitigate bias in ML models as well as new strategies currently being explored. Finally, with examples of how Perspective is being used in the real world, the talk will show how machine learning, combined with thoughtful human moderation and participation, can help improve online conversations.
Jigsaw的Perspective API旨在通过开发和提供识别有害文本的机器学习模型来保护在线对话中的声音。本次演讲将分享Perspective背后的团队如何通过谷歌的人工智能原则来思考公平,问责制,透明度,道德和社会问题。对于Perspective团队来说,构建公平和合乎道德的技术需要持续不断的努力。演讲将涵盖Perspective团队已经用于减轻ML模型中的偏见的具体策略,以及目前正在探索的新策略。最后,通过在现实世界中如何使用Perspective的例子,该演讲将展示机器学习如何与深思熟虑的人类调节和参与相结合,有助于改善在线对话。
{"title":"AI Principles in Identifying Toxicity in Online Conversation: Keynote at the Third Workshop on Fairness, Accountability, Transparency, Ethics and Society on the Web","authors":"Lucy Vasserman","doi":"10.1145/3442442.3452307","DOIUrl":"https://doi.org/10.1145/3442442.3452307","url":null,"abstract":"Jigsaw’s Perspective API aims to protect voices in online conversation by developing and serving machine learning models that identify toxicity text. This talk will share how the team behind Perspective thinks about the issues of Fairness, Accountability, Transparency, Ethics and Society through the lens of Google’s AI Principles. For the Perspective team, building technology that is fair and ethical is a continuous, ongoing effort. The talk will cover concrete strategies the Perspective team has already used to mitigate bias in ML models as well as new strategies currently being explored. Finally, with examples of how Perspective is being used in the real world, the talk will show how machine learning, combined with thoughtful human moderation and participation, can help improve online conversations.","PeriodicalId":129420,"journal":{"name":"Companion Proceedings of the Web Conference 2021","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114378780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Companion Proceedings of the Web Conference 2021
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1