首页 > 最新文献

HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media最新文献

英文 中文
Bridging link and query intent to enhance web search 桥接链接和查询意图,以增强网络搜索
Na Dai, Xiaoguang Qi, Brian D. Davison
Understanding query intent is essential to generating appropriate rankings for users. Existing methods have provided customized rankings to answer queries with different intent. While previous methods have shown improvement over their non-discriminating counterparts, the web authors' intent when creating a hyperlink is seldom taken into consideration. To mitigate this gap, we categorize hyperlinks into two types that are reasonably comparable to query intent, i.e., links describing the target page's identity and links describing the target page's content. We argue that emphasis on one type of link when ranking documents can benefit the retrieval for that type of query. We start by presenting a link intent classification approach based on the link context representations that captures evidence from anchors, target pages, and their associated links, and then introduce our enhanced retrieval model that incorporates link intent into the estimation of anchor text importance. Comparative experiments on two large scale web corpora demonstrate the efficacy of our approaches.
理解查询意图对于为用户生成适当的排名至关重要。现有的方法提供了定制的排名来回答不同意图的查询。虽然以前的方法已经比非歧视性的方法有所改进,但web作者在创建超链接时的意图很少被考虑在内。为了减轻这种差距,我们将超链接分为两种类型,这两种类型与查询意图相当,即描述目标页面标识的链接和描述目标页面内容的链接。我们认为,在对文档进行排序时,强调一种类型的链接有利于对该类型查询的检索。我们首先提出了一种基于链接上下文表示的链接意图分类方法,该方法从锚点、目标页面及其相关链接中获取证据,然后介绍了我们的增强检索模型,该模型将链接意图纳入锚文本重要性的估计中。在两个大型网络语料库上的对比实验证明了我们的方法的有效性。
{"title":"Bridging link and query intent to enhance web search","authors":"Na Dai, Xiaoguang Qi, Brian D. Davison","doi":"10.1145/1995966.1995973","DOIUrl":"https://doi.org/10.1145/1995966.1995973","url":null,"abstract":"Understanding query intent is essential to generating appropriate rankings for users. Existing methods have provided customized rankings to answer queries with different intent. While previous methods have shown improvement over their non-discriminating counterparts, the web authors' intent when creating a hyperlink is seldom taken into consideration. To mitigate this gap, we categorize hyperlinks into two types that are reasonably comparable to query intent, i.e., links describing the target page's identity and links describing the target page's content. We argue that emphasis on one type of link when ranking documents can benefit the retrieval for that type of query. We start by presenting a link intent classification approach based on the link context representations that captures evidence from anchors, target pages, and their associated links, and then introduce our enhanced retrieval model that incorporates link intent into the estimation of anchor text importance. Comparative experiments on two large scale web corpora demonstrate the efficacy of our approaches.","PeriodicalId":91270,"journal":{"name":"HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media","volume":"70 1","pages":"17-26"},"PeriodicalIF":0.0,"publicationDate":"2011-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86747374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Extracting the mesoscopic structure from heterogeneous systems 从非均相体系中提取介观结构
Xin Liu, T. Murata
Heterogeneous systems in nature are often characterized by the mesoscopic structure known as communities. In this paper, we propose a framework to address the problem of community detection in bipartite networks and tripartite hypernetworks, which are appropriate models for many heterogeneous systems. The most important advantage of our method is that it is competent for detecting both communities of one-to-one correspondence and communities of many-to-many correspondence, while state of the art techniques can only handle the former. We demonstrate this advantage and show other desired properties of our method through extensive experiments in both synthetic and real-world datasets.
自然界中的异质系统通常以称为群落的介观结构为特征。在本文中,我们提出了一个框架来解决二部网络和三部超网络中的社区检测问题,这是许多异构系统的合适模型。我们的方法最重要的优点是它能够检测一对一通信和多对多通信的社区,而目前的技术只能处理前者。我们展示了这一优势,并通过在合成和现实世界数据集上的广泛实验展示了我们方法的其他期望属性。
{"title":"Extracting the mesoscopic structure from heterogeneous systems","authors":"Xin Liu, T. Murata","doi":"10.1145/1995966.1995995","DOIUrl":"https://doi.org/10.1145/1995966.1995995","url":null,"abstract":"Heterogeneous systems in nature are often characterized by the mesoscopic structure known as communities. In this paper, we propose a framework to address the problem of community detection in bipartite networks and tripartite hypernetworks, which are appropriate models for many heterogeneous systems. The most important advantage of our method is that it is competent for detecting both communities of one-to-one correspondence and communities of many-to-many correspondence, while state of the art techniques can only handle the former. We demonstrate this advantage and show other desired properties of our method through extensive experiments in both synthetic and real-world datasets.","PeriodicalId":91270,"journal":{"name":"HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media","volume":"29 1","pages":"211-220"},"PeriodicalIF":0.0,"publicationDate":"2011-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89280659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Automatic mining of cognitive metadata using fuzzy inference 基于模糊推理的认知元数据自动挖掘
M. Şah, V. Wade
Personalized search and browsing is increasingly vital especially for enterprises to able to reach their customers. Key challenge in supporting personalization is the need for rich metadata such as cognitive metadata about documents. As we consider size of large knowledge bases, manual annotation is not scalable and feasible. On the other hand, automatic mining of cognitive metadata is challenging since it is very difficult to understand underlying intellectual knowledge about documents automatically. To alleviate this problem, we introduce a novel metadata extraction framework, which is based on fuzzy information granulation and fuzzy inference system for automatic cognitive metadata mining. The user evaluation study shows that our approach provides reasonable precision rates for difficulty, interactivity type, and interactivity level on the examined 100 documents. In addition, proposed fuzzy inference system achieves improved results compared to a rule-based reasoner for document difficulty metadata extraction (11% improvement).
个性化搜索和浏览变得越来越重要,特别是对于企业能够接触到他们的客户。支持个性化的关键挑战是需要丰富的元数据,比如关于文档的认知元数据。当我们考虑大型知识库的规模时,手动标注是不可伸缩的和不可行的。另一方面,认知元数据的自动挖掘具有挑战性,因为很难自动理解关于文档的底层智能知识。为了解决这一问题,我们引入了一种新的元数据提取框架,该框架基于模糊信息粒化和模糊推理系统进行自动认知元数据挖掘。用户评估研究表明,我们的方法对所检查的100个文档的难度、交互性类型和交互性级别提供了合理的准确率。此外,与基于规则的推理器相比,本文提出的模糊推理系统在文档难度元数据提取方面取得了更好的结果(提高了11%)。
{"title":"Automatic mining of cognitive metadata using fuzzy inference","authors":"M. Şah, V. Wade","doi":"10.1145/1995966.1995975","DOIUrl":"https://doi.org/10.1145/1995966.1995975","url":null,"abstract":"Personalized search and browsing is increasingly vital especially for enterprises to able to reach their customers. Key challenge in supporting personalization is the need for rich metadata such as cognitive metadata about documents. As we consider size of large knowledge bases, manual annotation is not scalable and feasible. On the other hand, automatic mining of cognitive metadata is challenging since it is very difficult to understand underlying intellectual knowledge about documents automatically. To alleviate this problem, we introduce a novel metadata extraction framework, which is based on fuzzy information granulation and fuzzy inference system for automatic cognitive metadata mining. The user evaluation study shows that our approach provides reasonable precision rates for difficulty, interactivity type, and interactivity level on the examined 100 documents. In addition, proposed fuzzy inference system achieves improved results compared to a rule-based reasoner for document difficulty metadata extraction (11% improvement).","PeriodicalId":91270,"journal":{"name":"HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media","volume":"68 1","pages":"37-46"},"PeriodicalIF":0.0,"publicationDate":"2011-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84421710","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
A3P: adaptive policy prediction for shared images over popular content sharing sites A3P:对流行内容共享站点上的共享图像进行自适应策略预测
A. Squicciarini, Smitha Sundareswaran, D. Lin, Joshua Wede
More and more people go online today and share their personal images using popular web services like Picasa. While enjoying the convenience brought by advanced technology, people also become aware of the privacy issues of data being shared. Recent studies have highlighted that people expect more tools to allow them to regain control over their privacy. In this work, we propose an Adaptive Privacy Policy Prediction (A3P) system to help users compose privacy settings for their images. In particular, we examine the role of image content and metadata as possible indicators of users' privacy preferences. We propose a two-level image classification framework to obtain image categories which may be associated with similar policies. Then, we develop a policy prediction algorithm to automatically generate a policy for each newly uploaded image. Most importantly, the generated policy will follow the trend of the user's privacy concerns evolved with time. We have conducted an extensive user study and the results demonstrate effectiveness of our system with the prediction accuracy around 90%.
如今,越来越多的人上网,使用Picasa等流行的网络服务分享他们的个人照片。在享受先进技术带来的便利的同时,人们也意识到数据共享带来的隐私问题。最近的研究强调,人们希望有更多的工具可以让他们重新控制自己的隐私。在这项工作中,我们提出了一个自适应隐私策略预测(A3P)系统来帮助用户为他们的图像设置隐私。特别是,我们研究了图像内容和元数据作为用户隐私偏好的可能指标的作用。我们提出了一个两级图像分类框架,以获得可能与相似策略相关联的图像类别。然后,我们开发了一种策略预测算法,为每个新上传的图像自动生成策略。最重要的是,生成的策略将遵循用户隐私关注随时间演变的趋势。我们进行了广泛的用户研究,结果证明了我们系统的有效性,预测准确率约为90%。
{"title":"A3P: adaptive policy prediction for shared images over popular content sharing sites","authors":"A. Squicciarini, Smitha Sundareswaran, D. Lin, Joshua Wede","doi":"10.1145/1995966.1996000","DOIUrl":"https://doi.org/10.1145/1995966.1996000","url":null,"abstract":"More and more people go online today and share their personal images using popular web services like Picasa. While enjoying the convenience brought by advanced technology, people also become aware of the privacy issues of data being shared. Recent studies have highlighted that people expect more tools to allow them to regain control over their privacy. In this work, we propose an Adaptive Privacy Policy Prediction (A3P) system to help users compose privacy settings for their images. In particular, we examine the role of image content and metadata as possible indicators of users' privacy preferences. We propose a two-level image classification framework to obtain image categories which may be associated with similar policies. Then, we develop a policy prediction algorithm to automatically generate a policy for each newly uploaded image. Most importantly, the generated policy will follow the trend of the user's privacy concerns evolved with time. We have conducted an extensive user study and the results demonstrate effectiveness of our system with the prediction accuracy around 90%.","PeriodicalId":91270,"journal":{"name":"HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media","volume":"8 1","pages":"261-270"},"PeriodicalIF":0.0,"publicationDate":"2011-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86425673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 101
Little search game: term network acquisition via a human computation game 小搜索游戏:通过人工计算获取网络词汇的游戏
Jakub Simko, M. Tvarozek, M. Bieliková
Semantic structures, ranging from ontologies to flat folksonomies, are widely used on the Web despite the fact that their creation in sufficient quality is often a costly task. We propose a new approach for acquiring a lightweight network of related terms via the Little Search Game - a competitive browser game in search query formulation. The format of game queries forces players to express their perception of term relatedness. The term network is aggregated using "votes" from multiple players playing the same problem instance. We show that nearly 91% of the relationships produced by Little Search Game are correct and also elaborate on the game's unique ability to discover term relations, that are otherwise hidden to typical corpora mining methods.
语义结构(从本体到扁平的大众分类法)在Web上被广泛使用,尽管以足够的质量创建它们通常是一项代价高昂的任务。我们提出了一种新的方法,通过小搜索游戏获得一个轻量级的相关术语网络-一个竞争性的浏览器游戏在搜索查询公式。游戏查询的格式迫使玩家表达他们对术语相关性的看法。术语网络使用来自多个玩家的“投票”来聚合相同的问题实例。我们发现Little Search Game产生的近91%的关系是正确的,并且还详细说明了游戏发现术语关系的独特能力,否则典型的语料库挖掘方法是隐藏的。
{"title":"Little search game: term network acquisition via a human computation game","authors":"Jakub Simko, M. Tvarozek, M. Bieliková","doi":"10.1145/1995966.1995977","DOIUrl":"https://doi.org/10.1145/1995966.1995977","url":null,"abstract":"Semantic structures, ranging from ontologies to flat folksonomies, are widely used on the Web despite the fact that their creation in sufficient quality is often a costly task. We propose a new approach for acquiring a lightweight network of related terms via the Little Search Game - a competitive browser game in search query formulation. The format of game queries forces players to express their perception of term relatedness. The term network is aggregated using \"votes\" from multiple players playing the same problem instance. We show that nearly 91% of the relationships produced by Little Search Game are correct and also elaborate on the game's unique ability to discover term relations, that are otherwise hidden to typical corpora mining methods.","PeriodicalId":91270,"journal":{"name":"HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media","volume":"12 1","pages":"57-62"},"PeriodicalIF":0.0,"publicationDate":"2011-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88012843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Many views, many modes, many tools ... one structure: Towards a Non-disruptive Integration of Personal Information 许多视图,许多模式,许多工具……一个结构:迈向个人信息的非破坏性整合
W. Jones, K. Anderson
People yearn for more integration of their information. But tools meant to help often do the opposite-pulling people and their information in different directions. Fragmentation is potentially worsened as personal information moves onto the Web and into a myriad of special-purpose, mobile-enabled applications. How can tool developers innovate "non-disruptively" in ways that do not force people to re-organize or re-locate their information? This paper makes two arguments: 1. An integration of personal information is not likely to happen through some new release of a desktop operating system or via a Web-based "super tool." 2. Instead, integration is best supported through the development of a standards-based infrastructure that makes provision for the shared manipulation of common structure by any number of tools, each in its own way. To illustrate this approach, the paper describes an XML-based schema, considerations in its design and its current use in three separate tools. The schema in its design and use builds on the lessons learned by the open hypermedia and structural computing communities while moving forward with new techniques that address some of the changes introduced by the evolution of the term "application" to move beyond desktop apps to mobile apps, cloud-based apps and various hybrid architectures.
人们渴望更多的信息整合。但是,旨在提供帮助的工具往往适得其反——把人们和他们的信息拉向不同的方向。随着个人信息转移到网络上,进入无数特殊用途的移动应用程序,碎片化可能会加剧。工具开发人员如何以不强迫人们重新组织或重新定位信息的方式进行“非破坏性”创新?本文提出了两个论点:1.;个人信息的整合不太可能通过新发布的桌面操作系统或基于web的“超级工具”来实现。2. 相反,通过开发基于标准的基础设施来最好地支持集成,该基础设施可以通过任意数量的工具(每种工具都以自己的方式)提供对公共结构的共享操作。为了说明这种方法,本文描述了一种基于xml的模式、其设计中的注意事项以及它在三个独立工具中的当前使用情况。该模式的设计和使用基于开放超媒体和结构计算社区的经验教训,同时采用新技术,解决了“应用程序”一词的演变所带来的一些变化,从桌面应用程序转向移动应用程序、基于云的应用程序和各种混合架构。
{"title":"Many views, many modes, many tools ... one structure: Towards a Non-disruptive Integration of Personal Information","authors":"W. Jones, K. Anderson","doi":"10.1145/1995966.1995984","DOIUrl":"https://doi.org/10.1145/1995966.1995984","url":null,"abstract":"People yearn for more integration of their information. But tools meant to help often do the opposite-pulling people and their information in different directions. Fragmentation is potentially worsened as personal information moves onto the Web and into a myriad of special-purpose, mobile-enabled applications. How can tool developers innovate \"non-disruptively\" in ways that do not force people to re-organize or re-locate their information? This paper makes two arguments: 1. An integration of personal information is not likely to happen through some new release of a desktop operating system or via a Web-based \"super tool.\" 2. Instead, integration is best supported through the development of a standards-based infrastructure that makes provision for the shared manipulation of common structure by any number of tools, each in its own way. To illustrate this approach, the paper describes an XML-based schema, considerations in its design and its current use in three separate tools. The schema in its design and use builds on the lessons learned by the open hypermedia and structural computing communities while moving forward with new techniques that address some of the changes introduced by the evolution of the term \"application\" to move beyond desktop apps to mobile apps, cloud-based apps and various hybrid architectures.","PeriodicalId":91270,"journal":{"name":"HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media","volume":"1 1","pages":"113-122"},"PeriodicalIF":0.0,"publicationDate":"2011-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88244404","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
New plots for hypertext?: towards poetics of a hypertext node 超文本的新情节?:超文本节点的诗学研究
Mariusz Pisarski
While the significance of hypertext links for the new ways of telling stories has been widely discussed, there has been not many debates about the very elements that are being connected: hypertext nodes. Apart from few exceptions, poetics of the link overshadows poetics of the node. My goal is to re-focus on a single node, or lexia, by introducing the concept of contextual regulation as the major force that shapes hypertext narrative units. Because many lexias must be capable of occurring in different contexts and at different stages of the unfolding story, several compromises have to be made on the level of language, style, plot and discourse. Each node, depending on its position and importance, has a varying level of connectivity and autonomy, which affects the global coherence of text. After focusing on relations between the notion of lexia (as a coherent and flexible unit) and the notion of kernel in narrative theory, an explanation of rules behind contextual regulation is presented, along with the basic typology of nodes. Then an attempt to enhance existing plot pools for hypertext fiction is undertaken. Several suggestions for the new plots, offered by the node-centered approach, are introduced.
虽然超文本链接对于讲述故事的新方式的重要性已经被广泛讨论,但对于正在被连接的元素:超文本节点,却没有多少争论。除了少数例外,链接的诗学遮蔽了节点的诗学。我的目标是通过引入上下文规则作为塑造超文本叙事单元的主要力量的概念,重新关注单个节点或词汇。因为许多词汇必须能够出现在不同的语境和故事展开的不同阶段,所以必须在语言、风格、情节和话语的层面上做出一些妥协。每个节点根据其位置和重要性的不同,具有不同程度的连通性和自主性,从而影响文本的整体连贯性。在关注词汇概念(作为一个连贯和灵活的单位)与叙事理论中的核心概念之间的关系之后,本文提出了对语境调节背后规则的解释,以及节点的基本类型。然后尝试增强现有的超文本小说的情节池。介绍了以节点为中心的方法对新地块提出的几点建议。
{"title":"New plots for hypertext?: towards poetics of a hypertext node","authors":"Mariusz Pisarski","doi":"10.1145/1995966.1996007","DOIUrl":"https://doi.org/10.1145/1995966.1996007","url":null,"abstract":"While the significance of hypertext links for the new ways of telling stories has been widely discussed, there has been not many debates about the very elements that are being connected: hypertext nodes. Apart from few exceptions, poetics of the link overshadows poetics of the node. My goal is to re-focus on a single node, or lexia, by introducing the concept of contextual regulation as the major force that shapes hypertext narrative units. Because many lexias must be capable of occurring in different contexts and at different stages of the unfolding story, several compromises have to be made on the level of language, style, plot and discourse. Each node, depending on its position and importance, has a varying level of connectivity and autonomy, which affects the global coherence of text.\u0000 After focusing on relations between the notion of lexia (as a coherent and flexible unit) and the notion of kernel in narrative theory, an explanation of rules behind contextual regulation is presented, along with the basic typology of nodes. Then an attempt to enhance existing plot pools for hypertext fiction is undertaken. Several suggestions for the new plots, offered by the node-centered approach, are introduced.","PeriodicalId":91270,"journal":{"name":"HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media","volume":"10 1","pages":"313-318"},"PeriodicalIF":0.0,"publicationDate":"2011-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86839940","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Evaluating significance of historical entities based on tempo-spatial impacts analysis using Wikipedia link structure 基于维基百科链接结构时空影响分析的历史实体意义评价
Y. Takahashi, H. Ohshima, Mitsuo Yamamoto, H. Iwasaki, S. Oyama, Katsumi Tanaka
We propose a method to evaluate the significance of historical entities (people, events, and so on.). Here, the significance of a historical entity means how it affected other historical entities. Our proposed method first calculates the tempo-spacial impact of historical entities. The impact of a historical entity varies according to time and location. Historical entities are collected from Wikipedia. We assume that a Wikipedia link between historical entities represents an impact propagation. That is, when an entity has a link to another entity, we regard the former is influenced by the latter. Historical entities in Wikipedia usually have the date and location of their occurrence. Our proposed iteration algorithm propagates such initial tempo-spacial information through links in the similar manner as PageRank, so the tempo-spacial impact scores of all the historical entities can be calculated. We assume that a historical entity is significant if it influences many other entities that are far from it temporally or geographically. We demonstrate a prototype system and show the results of experiments that prove the effectiveness of our method.
我们提出了一种评估历史实体(人物、事件等)意义的方法。在这里,一个历史实体的意义是指它如何影响其他历史实体。我们提出的方法首先计算历史实体的时空影响。历史实体的影响因时间和地点的不同而不同。历史实体是从维基百科收集的。我们假设历史实体之间的维基百科链接代表影响传播。也就是说,当一个实体与另一个实体有联系时,我们认为前者受到后者的影响。维基百科中的历史实体通常有其出现的日期和地点。我们提出的迭代算法以类似于PageRank的方式通过链接传播这些初始的时间空间信息,因此可以计算出所有历史实体的时间空间影响得分。如果一个历史实体影响了在时间上或地理上与它相距甚远的许多其他实体,我们就认为它是重要的。我们演示了一个原型系统,并给出了实验结果,证明了我们方法的有效性。
{"title":"Evaluating significance of historical entities based on tempo-spatial impacts analysis using Wikipedia link structure","authors":"Y. Takahashi, H. Ohshima, Mitsuo Yamamoto, H. Iwasaki, S. Oyama, Katsumi Tanaka","doi":"10.1145/1995966.1995980","DOIUrl":"https://doi.org/10.1145/1995966.1995980","url":null,"abstract":"We propose a method to evaluate the significance of historical entities (people, events, and so on.). Here, the significance of a historical entity means how it affected other historical entities. Our proposed method first calculates the tempo-spacial impact of historical entities. The impact of a historical entity varies according to time and location. Historical entities are collected from Wikipedia. We assume that a Wikipedia link between historical entities represents an impact propagation. That is, when an entity has a link to another entity, we regard the former is influenced by the latter. Historical entities in Wikipedia usually have the date and location of their occurrence. Our proposed iteration algorithm propagates such initial tempo-spacial information through links in the similar manner as PageRank, so the tempo-spacial impact scores of all the historical entities can be calculated. We assume that a historical entity is significant if it influences many other entities that are far from it temporally or geographically. We demonstrate a prototype system and show the results of experiments that prove the effectiveness of our method.","PeriodicalId":91270,"journal":{"name":"HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media","volume":"96 1","pages":"83-92"},"PeriodicalIF":0.0,"publicationDate":"2011-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80514761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
GALE: a highly extensible adaptive hypermedia engine GALE:一个高度可扩展的自适应超媒体引擎
D. Smits, P. D. Bra
This paper presents GALE, the GRAPPLE Adaptive Learning Environment, which (contrary to what the word suggests) is a truly generic and general purpose adaptive hypermedia engine. Five years have passed since "The Design of AHA!" was published at ACM Hypertext (2006). GALE takes the notion of general-purpose a whole lot further. We solve shortcomings of existing adaptive systems in terms of genericity, extensibility and usability and show how GALE improves on the state of the art in all these aspects. We illustrate different authoring styles for GALE, including the use of template pages, and show how adaptation can be defined in a completely decentralized way by using the open corpus adaptation facility of GALE. GALE has been used in a number of adaptive hypermedia workshops and assignments to test whether authors can actually make use of the extensive functionality that GALE offers. Adaptation has been added to wiki sites, existing material e.g. from w3schools, and of course also to locally authored hypertext. Soon GALE will be used in cross-course adaptation at the TU/e in a pilot project to improve the success rate of university students.
本文介绍了GALE,即GRAPPLE自适应学习环境,它(与这个词所暗示的相反)是一个真正通用的、通用的自适应超媒体引擎。自“AHA的设计!”在ACM超文本(2006)上发表以来,已经过去了五年。GALE进一步发展了通用的概念。我们解决了现有自适应系统在通用性、可扩展性和可用性方面的缺点,并展示了GALE如何在所有这些方面改进了目前的技术水平。我们举例说明了GALE的不同创作风格,包括模板页面的使用,并展示了如何使用GALE的开放语料库自适应功能以完全分散的方式定义自适应。GALE已在许多自适应超媒体研讨会和作业中使用,以测试作者是否能够实际利用GALE提供的广泛功能。改编已经被添加到wiki站点,现有的材料,例如来自w3schools,当然也包括本地创作的超文本。不久,GALE将在理工大学的一个试点项目中用于跨课程适应,以提高大学生的成功率。
{"title":"GALE: a highly extensible adaptive hypermedia engine","authors":"D. Smits, P. D. Bra","doi":"10.1145/1995966.1995978","DOIUrl":"https://doi.org/10.1145/1995966.1995978","url":null,"abstract":"This paper presents GALE, the GRAPPLE Adaptive Learning Environment, which (contrary to what the word suggests) is a truly generic and general purpose adaptive hypermedia engine. Five years have passed since \"The Design of AHA!\" was published at ACM Hypertext (2006). GALE takes the notion of general-purpose a whole lot further. We solve shortcomings of existing adaptive systems in terms of genericity, extensibility and usability and show how GALE improves on the state of the art in all these aspects. We illustrate different authoring styles for GALE, including the use of template pages, and show how adaptation can be defined in a completely decentralized way by using the open corpus adaptation facility of GALE. GALE has been used in a number of adaptive hypermedia workshops and assignments to test whether authors can actually make use of the extensive functionality that GALE offers. Adaptation has been added to wiki sites, existing material e.g. from w3schools, and of course also to locally authored hypertext. Soon GALE will be used in cross-course adaptation at the TU/e in a pilot project to improve the success rate of university students.","PeriodicalId":91270,"journal":{"name":"HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media","volume":"82 1","pages":"63-72"},"PeriodicalIF":0.0,"publicationDate":"2011-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83632476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
A community question-answering refinement system 社区问答精细化系统
M. S. Pera, Yiu-Kai Ng
Community Question Answering (CQA) websites, which archive millions of questions and answers created by CQA users to provide a rich resource of information that is missing at web search engines and QA websites, have become increasingly popular. Web users who search for answers to their questions at CQA websites, however, are often required to either (i) wait for days until other CQA users post answers to their questions which might even be incorrect, offensive, or spam, or (ii) deal with restricted answer sets created by CQA websites due to the exact-match constraint that is employed and imposed between archived questions and user-formulated questions. To automate and enhance the process of locating high-quality answers to a user's question Q at a CQA website, we introduce a CQA refinement system, called QAR. Given Q, QAR first retrieves a set of CQA questions QS that are the same as, or similar to, Q in terms of its specified information need. Thereafter, QAR selects as answers to Q the top-ranked answers (among the ones to the questions in QS) based on various similarity scores and the length of the answers. Empirical studies, which were conducted using questions provided by the Text Retrieval Conference (TREC) and Text Analysis Conference (TAC), in addition to more than four millions questions (and their corresponding answers) extracted from Yahoo! Answers, show that QAR is effective in locating archived answers, if they exist, that satisfy the information need specified in Q. We have further assessed the performance of QAR by comparing its question-matching and answer-ranking strategies with their Yahoo! Answers' counterparts and verified that QAR outperforms Yahoo! Answers in (i) locating the set of questions QS that have the highest degrees of similarity with Q and (ii) ranking archived answers to QS as answers to Q.
社区问答(CQA)网站存档了数以百万计的由CQA用户创建的问题和答案,提供了网络搜索引擎和QA网站所缺少的丰富的信息资源,已经变得越来越流行。然而,在CQA网站上搜索问题答案的网络用户通常需要(i)等待数天,直到其他CQA用户发布他们的问题的答案,这些答案甚至可能是不正确的、冒犯性的或垃圾邮件的,或者(ii)处理由CQA网站创建的受限答案集,这是由于在存档问题和用户制定的问题之间采用和强加的精确匹配约束。为了自动化和增强在CQA网站上查找用户问题Q的高质量答案的过程,我们引入了一个CQA优化系统,称为QAR。给定Q, QAR首先检索一组CQA问题QS,这些问题在其指定的信息需求方面与Q相同或相似。然后,QAR根据各种相似度得分和答案长度,选择排名靠前的答案(在QS中问题的答案中)作为Q的答案。实证研究使用了文本检索会议(TREC)和文本分析会议(TAC)提供的问题,以及从Yahoo!答案,表明QAR在查找存档答案方面是有效的,如果它们存在,满足q中指定的信息需求。我们通过比较QAR的问题匹配和答案排序策略与Yahoo!回答同行并验证QAR优于Yahoo!(i)定位与Q相似度最高的一组问题,(ii)将已存档的QS答案按Q的答案排序。
{"title":"A community question-answering refinement system","authors":"M. S. Pera, Yiu-Kai Ng","doi":"10.1145/1995966.1995999","DOIUrl":"https://doi.org/10.1145/1995966.1995999","url":null,"abstract":"Community Question Answering (CQA) websites, which archive millions of questions and answers created by CQA users to provide a rich resource of information that is missing at web search engines and QA websites, have become increasingly popular. Web users who search for answers to their questions at CQA websites, however, are often required to either (i) wait for days until other CQA users post answers to their questions which might even be incorrect, offensive, or spam, or (ii) deal with restricted answer sets created by CQA websites due to the exact-match constraint that is employed and imposed between archived questions and user-formulated questions. To automate and enhance the process of locating high-quality answers to a user's question Q at a CQA website, we introduce a CQA refinement system, called QAR. Given Q, QAR first retrieves a set of CQA questions QS that are the same as, or similar to, Q in terms of its specified information need. Thereafter, QAR selects as answers to Q the top-ranked answers (among the ones to the questions in QS) based on various similarity scores and the length of the answers. Empirical studies, which were conducted using questions provided by the Text Retrieval Conference (TREC) and Text Analysis Conference (TAC), in addition to more than four millions questions (and their corresponding answers) extracted from Yahoo! Answers, show that QAR is effective in locating archived answers, if they exist, that satisfy the information need specified in Q. We have further assessed the performance of QAR by comparing its question-matching and answer-ranking strategies with their Yahoo! Answers' counterparts and verified that QAR outperforms Yahoo! Answers in (i) locating the set of questions QS that have the highest degrees of similarity with Q and (ii) ranking archived answers to QS as answers to Q.","PeriodicalId":91270,"journal":{"name":"HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media","volume":"8 1","pages":"251-260"},"PeriodicalIF":0.0,"publicationDate":"2011-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78869575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
期刊
HT ... : the proceedings of the ... ACM Conference on Hypertext and Social Media. ACM Conference on Hypertext and Social Media
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1