首页 > 最新文献

2013 International Conference on Social Computing最新文献

英文 中文
Three Fiscal Policy Experiments in an Agent-Based Macroeconomic Model 基于主体的宏观经济模型中的三个财政政策实验
Pub Date : 2013-09-08 DOI: 10.1109/SocialCom.2013.115
Carl M. Gustafson
In this paper, I build from scratch a basic agent-based macroeconomic model, featuring fifty representative agents whose decisions to consume and save depend on the current relative performance of the economy at-large. I run three different experiments in the framework: the first, on the effects of tax and spending "flexibility" on stabilizing output, the second, on the ability of spending stimulus to stabilize output, and the third, on redistributive measures across income groups and their effects on aggregate economic performance. I find that tax and spending flexibility accelerates the path back to stability after an initial imposed downturn, that spending stimulus does much the same, though with a greater initial "kick", and that redistribution in this model may take place and increase the welfare of lower-income agents without imposing a significant burden on overall performance.
在本文中,我从零开始建立了一个基本的基于主体的宏观经济模型,其中有50个代表主体,他们的消费和储蓄决策取决于当前整体经济的相对表现。我在这个框架中进行了三个不同的实验:第一个是关于税收和支出“灵活性”对稳定产出的影响,第二个是关于支出刺激稳定产出的能力,第三个是关于跨收入群体的再分配措施及其对总体经济表现的影响。我发现,税收和支出的灵活性加速了在最初强加的低迷之后回归稳定的道路,支出刺激也起到了同样的作用,尽管最初的“推力”更大,而且该模型中的再分配可能会发生,并增加低收入主体的福利,而不会对整体表现造成重大负担。
{"title":"Three Fiscal Policy Experiments in an Agent-Based Macroeconomic Model","authors":"Carl M. Gustafson","doi":"10.1109/SocialCom.2013.115","DOIUrl":"https://doi.org/10.1109/SocialCom.2013.115","url":null,"abstract":"In this paper, I build from scratch a basic agent-based macroeconomic model, featuring fifty representative agents whose decisions to consume and save depend on the current relative performance of the economy at-large. I run three different experiments in the framework: the first, on the effects of tax and spending \"flexibility\" on stabilizing output, the second, on the ability of spending stimulus to stabilize output, and the third, on redistributive measures across income groups and their effects on aggregate economic performance. I find that tax and spending flexibility accelerates the path back to stability after an initial imposed downturn, that spending stimulus does much the same, though with a greater initial \"kick\", and that redistribution in this model may take place and increase the welfare of lower-income agents without imposing a significant burden on overall performance.","PeriodicalId":129308,"journal":{"name":"2013 International Conference on Social Computing","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132814718","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Partitioning and Scaling Signed Bipartite Graphs for Polarized Political Blogosphere 两极分化政治博客圈的签名二部图的分划与标度
Pub Date : 2013-09-08 DOI: 10.1109/SocialCom.2013.32
Sedat Gokalp, M. Temkit, H. Davulcu, I. H. Toroslu
Blogosphere plays an increasingly important role as a forum for public debate. In this paper, given a mixed set of blogs debating a set of political issues from opposing camps, we use signed bipartite graphs for modeling debates, and we propose an algorithm for partitioning both the blogs, and the issues (i.e. topics, leaders, etc.) comprising the debate into binary opposing camps. Simultaneously, our algorithm scales both the blogs and the underlying issues on a univariate scale. Using this scale, a researcher can identify moderate and extreme blogs within each camp, and polarizing vs. unifying issues. Through performance evaluations we show that our proposed algorithm provides an effective solution to the problem, and performs much better than existing baseline algorithms adapted to solve this new problem. In our experiments, we used both real data from political blogosphere and US Congress records, as well as synthetic data which were obtained by varying polarization and degree distribution of the vertices of the graph to show the robustness of our algorithm.
博客圈作为公众辩论的论坛,扮演着越来越重要的角色。在本文中,给定一组混合的博客,讨论来自对立阵营的一系列政治问题,我们使用签名二部图对辩论进行建模,并且我们提出了一种算法,用于将博客和组成辩论的问题(即主题,领导人等)划分为二元对立阵营。同时,我们的算法在单变量尺度上扩展博客和潜在问题。使用这个量表,研究人员可以在每个阵营中识别出温和和极端的博客,以及两极分化和统一的问题。通过性能评估,我们表明我们提出的算法提供了一个有效的解决方案,并且比现有的用于解决这个新问题的基线算法要好得多。在我们的实验中,我们既使用了来自政治博客圈和美国国会记录的真实数据,也使用了通过图中顶点的不同极化和度分布获得的合成数据,以显示我们算法的鲁棒性。
{"title":"Partitioning and Scaling Signed Bipartite Graphs for Polarized Political Blogosphere","authors":"Sedat Gokalp, M. Temkit, H. Davulcu, I. H. Toroslu","doi":"10.1109/SocialCom.2013.32","DOIUrl":"https://doi.org/10.1109/SocialCom.2013.32","url":null,"abstract":"Blogosphere plays an increasingly important role as a forum for public debate. In this paper, given a mixed set of blogs debating a set of political issues from opposing camps, we use signed bipartite graphs for modeling debates, and we propose an algorithm for partitioning both the blogs, and the issues (i.e. topics, leaders, etc.) comprising the debate into binary opposing camps. Simultaneously, our algorithm scales both the blogs and the underlying issues on a univariate scale. Using this scale, a researcher can identify moderate and extreme blogs within each camp, and polarizing vs. unifying issues. Through performance evaluations we show that our proposed algorithm provides an effective solution to the problem, and performs much better than existing baseline algorithms adapted to solve this new problem. In our experiments, we used both real data from political blogosphere and US Congress records, as well as synthetic data which were obtained by varying polarization and degree distribution of the vertices of the graph to show the robustness of our algorithm.","PeriodicalId":129308,"journal":{"name":"2013 International Conference on Social Computing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130968686","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Entity Matching in Online Social Networks 在线社交网络中的实体匹配
Pub Date : 2013-09-08 DOI: 10.1109/SocialCom.2013.53
Olga Peled, Michael Fire, L. Rokach, Y. Elovici
In recent years, Online Social Networks (OSNs) have essentially become an integral part of our daily lives. There are hundreds of OSNs, each with its own focus and offers for particular services and functionalities. To take advantage of the full range of services and functionalities that OSNs offer, users often create several accounts on various OSNs using the same or different personal information. Retrieving all available data about an individual from several OSNs and merging it into one profile can be useful for many purposes. In this paper, we present a method for solving the Entity Resolution (ER), problem for matching user profiles across multiple OSNs. Our algorithm is able to match two user profiles from two different OSNs based on machine learning techniques, which uses features extracted from each one of the user profiles. Using supervised learning techniques and extracted features, we constructed different classifiers, which were then trained and used to rank the probability that two user profiles from two different OSNs belong to the same individual. These classifiers utilized 27 features of mainly three types: name based features (i.e., the Soundex value of two names), general user info based features (i.e., the cosine similarity between two user profiles), and social network topological based features (i.e., the number of mutual friends between two users' friends list). This experimental study uses real-life data collected from two popular OSNs, Facebook and Xing. The proposed algorithm was evaluated and its classification performance measured by AUC was 0.982 in identifying user profiles across two OSNs.
近年来,在线社交网络(OSNs)已经成为我们日常生活中不可或缺的一部分。有数百个osn,每个都有自己的重点,并提供特定的服务和功能。为了充分利用osn提供的各种服务和功能,用户通常使用相同或不同的个人信息在不同的osn上创建多个帐户。从几个osn中检索关于个人的所有可用数据,并将其合并到一个概要文件中,这对于许多用途都很有用。在本文中,我们提出了一种解决跨多个osn匹配用户配置文件的实体解析(ER)问题的方法。我们的算法能够基于机器学习技术匹配来自两个不同osn的两个用户配置文件,该技术使用从每个用户配置文件中提取的特征。使用监督学习技术和提取的特征,我们构建了不同的分类器,然后对这些分类器进行训练并用于对来自两个不同osn的两个用户配置文件属于同一个体的概率进行排序。这些分类器利用了27个特征,主要有三种类型:基于姓名的特征(即两个姓名的Soundex值)、基于一般用户信息的特征(即两个用户档案之间的余弦相似度)和基于社交网络拓扑的特征(即两个用户的好友列表之间的共同好友数量)。这项实验研究使用了从两个流行的osn, Facebook和Xing收集的真实数据。对该算法进行了评价,AUC值为0.982。
{"title":"Entity Matching in Online Social Networks","authors":"Olga Peled, Michael Fire, L. Rokach, Y. Elovici","doi":"10.1109/SocialCom.2013.53","DOIUrl":"https://doi.org/10.1109/SocialCom.2013.53","url":null,"abstract":"In recent years, Online Social Networks (OSNs) have essentially become an integral part of our daily lives. There are hundreds of OSNs, each with its own focus and offers for particular services and functionalities. To take advantage of the full range of services and functionalities that OSNs offer, users often create several accounts on various OSNs using the same or different personal information. Retrieving all available data about an individual from several OSNs and merging it into one profile can be useful for many purposes. In this paper, we present a method for solving the Entity Resolution (ER), problem for matching user profiles across multiple OSNs. Our algorithm is able to match two user profiles from two different OSNs based on machine learning techniques, which uses features extracted from each one of the user profiles. Using supervised learning techniques and extracted features, we constructed different classifiers, which were then trained and used to rank the probability that two user profiles from two different OSNs belong to the same individual. These classifiers utilized 27 features of mainly three types: name based features (i.e., the Soundex value of two names), general user info based features (i.e., the cosine similarity between two user profiles), and social network topological based features (i.e., the number of mutual friends between two users' friends list). This experimental study uses real-life data collected from two popular OSNs, Facebook and Xing. The proposed algorithm was evaluated and its classification performance measured by AUC was 0.982 in identifying user profiles across two OSNs.","PeriodicalId":129308,"journal":{"name":"2013 International Conference on Social Computing","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128784623","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 88
Big Data and Policy Design for Data Sovereignty: A Case Study on Copyright and CCL in South Korea 大数据与数据主权的政策设计:以韩国版权和CCL为例
Pub Date : 2013-09-08 DOI: 10.1109/SocialCom.2013.165
Hyejung Moon, H. Cho
The purpose of this paper is as follows. First, I am trying to conceptualize big data as a social problem. Second, I would like to explain the difference between big data and conventional mega information. Third, I would like to recommend the role of the government for utilization of big data as a policy tools. Fourth, while referring to copyright and CCL(Creative Commons License) cases, I would like to explain the regulation for big data on data sovereignty. Finally, I would like to suggest a direction of policy design for big data. As for the result of this study, policy design for big data should be distinguished from policy design for mega information to solve data sovereignty issues. From a law system perspective, big data is generated autonomously. It has been accessed openly and shared without any intention. In market perspective, big data is created without any intention. Big data can be changed automatically in case of openness with reference feature such as Linked of Data. Some policy issues such as responsibility and authenticity should be raised. Big data is generated in a distributed and diverse way without any concrete form in technology perspective. So, we need a different approach.
本文的研究目的如下。首先,我试图将大数据概念化为一个社会问题。第二,我想解释一下大数据和传统海量信息的区别。第三,我想推荐政府在利用大数据作为政策工具方面的作用。第四,在参考版权和知识共享许可(CCL)案例的同时,我想从数据主权的角度来解释大数据的监管。最后,我想提出一个大数据政策设计的方向。从本研究的结果来看,应该将大数据政策设计与大信息政策设计区分开来,以解决数据主权问题。从法律体系的角度看,大数据是自主生成的。它在没有任何意图的情况下被公开访问和分享。从市场的角度来看,大数据是无意中产生的。大数据可以在开放的情况下自动更改,具有数据关联等参考功能。应该提出责任和真实性等政策问题。从技术角度看,大数据的产生方式是分布式的、多样化的,没有具体的形式。所以,我们需要一种不同的方法。
{"title":"Big Data and Policy Design for Data Sovereignty: A Case Study on Copyright and CCL in South Korea","authors":"Hyejung Moon, H. Cho","doi":"10.1109/SocialCom.2013.165","DOIUrl":"https://doi.org/10.1109/SocialCom.2013.165","url":null,"abstract":"The purpose of this paper is as follows. First, I am trying to conceptualize big data as a social problem. Second, I would like to explain the difference between big data and conventional mega information. Third, I would like to recommend the role of the government for utilization of big data as a policy tools. Fourth, while referring to copyright and CCL(Creative Commons License) cases, I would like to explain the regulation for big data on data sovereignty. Finally, I would like to suggest a direction of policy design for big data. As for the result of this study, policy design for big data should be distinguished from policy design for mega information to solve data sovereignty issues. From a law system perspective, big data is generated autonomously. It has been accessed openly and shared without any intention. In market perspective, big data is created without any intention. Big data can be changed automatically in case of openness with reference feature such as Linked of Data. Some policy issues such as responsibility and authenticity should be raised. Big data is generated in a distributed and diverse way without any concrete form in technology perspective. So, we need a different approach.","PeriodicalId":129308,"journal":{"name":"2013 International Conference on Social Computing","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128755362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Real-Time Access Control Rule Fault Detection Using a Simulated Logic Circuit 基于仿真逻辑电路的实时访问控制规则故障检测
Pub Date : 2013-09-08 DOI: 10.1109/SocialCom.2013.76
Vincent C. Hu, K. Scarfone
Access control (AC) policies can be implemented based on different AC models, which are fundamentally composed by semantically independent AC rules in expressions of privilege assignments described by attributes of subjects/attributes, actions, objects/attributes, and environment variables of the protected systems. Incorrect implementations of AC policies result in faults that not only leak but also disable access of information, and faults in AC policies are difficult to detect without support of verification or automatic fault detection mechanisms. This research proposes an automatic method through the construction of a simulated logic circuit that simulates AC rules in AC policies or models. The simulated logic circuit allows real-time detection of policy faults including conflicts of privilege assignments, leaks of information, and conflicts of interest assignments. Such detection is traditionally done by tools that perform verification or testing after all the rules of the policy/model are completed, and it provides no information about the source of verification errors. The real-time fault detecting capability proposed by this research allows a rule fault to be detected and fixed immediately before the next rule is added to the policy/model, thus requiring no later verification and saving a significant amount of fault fixing time.
访问控制(Access control, AC)策略可以基于不同的AC模型来实现,这些AC模型从根本上是由语义独立的AC规则组成的,这些规则通过被保护系统的主体/属性、动作、对象/属性和环境变量的属性来描述权限分配的表达式。如果AC策略执行错误,不仅会导致信息泄漏,还会导致信息无法访问。如果不支持验证或故障自动检测机制,AC策略中的故障很难被发现。本研究提出一种自动化方法,透过建构模拟逻辑电路来模拟交流策略或模型中的交流规则。仿真逻辑电路允许实时检测策略错误,包括特权分配冲突、信息泄漏和利益分配冲突。这种检测传统上是由在策略/模型的所有规则完成后执行验证或测试的工具完成的,并且它不提供关于验证错误来源的信息。本研究提出的实时故障检测能力允许在将下一个规则添加到策略/模型之前立即检测和修复规则故障,从而不需要后续验证,节省了大量的故障修复时间。
{"title":"Real-Time Access Control Rule Fault Detection Using a Simulated Logic Circuit","authors":"Vincent C. Hu, K. Scarfone","doi":"10.1109/SocialCom.2013.76","DOIUrl":"https://doi.org/10.1109/SocialCom.2013.76","url":null,"abstract":"Access control (AC) policies can be implemented based on different AC models, which are fundamentally composed by semantically independent AC rules in expressions of privilege assignments described by attributes of subjects/attributes, actions, objects/attributes, and environment variables of the protected systems. Incorrect implementations of AC policies result in faults that not only leak but also disable access of information, and faults in AC policies are difficult to detect without support of verification or automatic fault detection mechanisms. This research proposes an automatic method through the construction of a simulated logic circuit that simulates AC rules in AC policies or models. The simulated logic circuit allows real-time detection of policy faults including conflicts of privilege assignments, leaks of information, and conflicts of interest assignments. Such detection is traditionally done by tools that perform verification or testing after all the rules of the policy/model are completed, and it provides no information about the source of verification errors. The real-time fault detecting capability proposed by this research allows a rule fault to be detected and fixed immediately before the next rule is added to the policy/model, thus requiring no later verification and saving a significant amount of fault fixing time.","PeriodicalId":129308,"journal":{"name":"2013 International Conference on Social Computing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129069529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Natural Language Processing and Big Data - An Ontology-Based Approach for Cross-Lingual Information Retrieval 自然语言处理与大数据——基于本体的跨语言信息检索方法
Pub Date : 2013-09-08 DOI: 10.1109/SocialCom.2013.108
J. Monti, Mario Monteleone, Maria Pia di Buono, Federica Marano
Extracting relevant information in multilingual context from massive amounts of unstructured, structured and semi-structured data is a challenging task. Various theories have been developed and applied to ease the access to multicultural and multilingual resources. This papers describes a methodology for the development of an ontology-based Cross-Language Information Retrieval (CLIR) application and shows how it is possible to achieve the translation of Natural Language (NL) queries in any language by means of a knowledge-driven approach which allows to semi-automatically map natural language to formal language, simplifying and improving in this way the human-computer interaction and communication. The outlined research activities are based on Lexicon-Grammar (LG), a method devised for natural language formalization, automatic textual analysis and parsing. Thanks to its main characteristics, LG is independent from factors which are critical for other approaches, i.e. interaction type (voice or keyboard-based), length of sentences and propositions, type of vocabulary used and restrictions due to users' idiolects. The feasibility of our knowledge-based methodological framework, which allows mapping both data and metadata, will be tested for CLIR by implementing a domain-specific early prototype system.
从大量的非结构化、结构化和半结构化数据中提取多语言环境下的相关信息是一项具有挑战性的任务。各种理论已经发展并应用于简化多元文化和多语言资源的获取。本文描述了一种基于本体的跨语言信息检索(CLIR)应用程序的开发方法,并展示了如何通过知识驱动的方法实现任何语言的自然语言(NL)查询的翻译,该方法允许半自动地将自然语言映射到形式语言,以这种方式简化和改进人机交互和通信。本文概述的研究活动是基于词典语法(Lexicon-Grammar, LG),这是一种用于自然语言形式化、自动文本分析和解析的方法。由于它的主要特点,LG独立于其他方法的关键因素,即交互类型(语音或基于键盘),句子和命题的长度,使用的词汇类型和用户的习惯限制。我们基于知识的方法框架的可行性,它允许映射数据和元数据,将通过实现一个领域特定的早期原型系统来测试CLIR。
{"title":"Natural Language Processing and Big Data - An Ontology-Based Approach for Cross-Lingual Information Retrieval","authors":"J. Monti, Mario Monteleone, Maria Pia di Buono, Federica Marano","doi":"10.1109/SocialCom.2013.108","DOIUrl":"https://doi.org/10.1109/SocialCom.2013.108","url":null,"abstract":"Extracting relevant information in multilingual context from massive amounts of unstructured, structured and semi-structured data is a challenging task. Various theories have been developed and applied to ease the access to multicultural and multilingual resources. This papers describes a methodology for the development of an ontology-based Cross-Language Information Retrieval (CLIR) application and shows how it is possible to achieve the translation of Natural Language (NL) queries in any language by means of a knowledge-driven approach which allows to semi-automatically map natural language to formal language, simplifying and improving in this way the human-computer interaction and communication. The outlined research activities are based on Lexicon-Grammar (LG), a method devised for natural language formalization, automatic textual analysis and parsing. Thanks to its main characteristics, LG is independent from factors which are critical for other approaches, i.e. interaction type (voice or keyboard-based), length of sentences and propositions, type of vocabulary used and restrictions due to users' idiolects. The feasibility of our knowledge-based methodological framework, which allows mapping both data and metadata, will be tested for CLIR by implementing a domain-specific early prototype system.","PeriodicalId":129308,"journal":{"name":"2013 International Conference on Social Computing","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132610423","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
A Conjunction for Private Stream Searching 用于私有流搜索的连接
Pub Date : 2013-09-08 DOI: 10.1109/SocialCom.2013.69
Michael J. Oehler, D. Phatak
Our contribution defines a conjunction operator for private stream searching. Private stream searching is a system of cryptographic methods that preserves the confidentiality of the search criteria and the result. The system uses an encrypted filter to conceal the search terms, processes a search without decrypting these terms, and saves the result to an encrypted buffer. Fundamentally, the system provides a private search capability based on a logical disjunction of search terms. Our conjunction operator broadens the search capability, and achieves this without significantly increasing the complexity of the private search system. The conjunction is processed as a bit wise summation of hashed keyword values to reference an encrypted entry in the filter. The method is best suited for a conjunction of fields from a record, does not impute a calculation of bilinear map, as required in prior research, and offers a practical utility that integrates into private stream searching. We demonstrate the practicality by including the conjunction operator into our domain specific language for private packet filtering.
我们的贡献定义了一个用于私有流搜索的连接运算符。私有流搜索是一种保护搜索标准和结果机密性的加密方法系统。该系统使用加密过滤器来隐藏搜索项,在不解密这些搜索项的情况下处理搜索,并将结果保存到加密缓冲区中。从根本上说,该系统提供了基于搜索词的逻辑分离的私有搜索功能。我们的联合运算符拓宽了搜索能力,并且在不显著增加私有搜索系统复杂性的情况下实现了这一目标。连接被处理为哈希关键字值的逐位求和,以引用过滤器中的加密条目。该方法最适合于记录中字段的连接,不需要像以前的研究那样计算双线性映射,并且提供了集成到私有流搜索中的实用工具。我们通过将连接运算符包含到我们的领域特定语言中用于私有包过滤来证明其实用性。
{"title":"A Conjunction for Private Stream Searching","authors":"Michael J. Oehler, D. Phatak","doi":"10.1109/SocialCom.2013.69","DOIUrl":"https://doi.org/10.1109/SocialCom.2013.69","url":null,"abstract":"Our contribution defines a conjunction operator for private stream searching. Private stream searching is a system of cryptographic methods that preserves the confidentiality of the search criteria and the result. The system uses an encrypted filter to conceal the search terms, processes a search without decrypting these terms, and saves the result to an encrypted buffer. Fundamentally, the system provides a private search capability based on a logical disjunction of search terms. Our conjunction operator broadens the search capability, and achieves this without significantly increasing the complexity of the private search system. The conjunction is processed as a bit wise summation of hashed keyword values to reference an encrypted entry in the filter. The method is best suited for a conjunction of fields from a record, does not impute a calculation of bilinear map, as required in prior research, and offers a practical utility that integrates into private stream searching. We demonstrate the practicality by including the conjunction operator into our domain specific language for private packet filtering.","PeriodicalId":129308,"journal":{"name":"2013 International Conference on Social Computing","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131728012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Automatic Labeling of Training Data for Collecting Tweets for Ambiguous TV Program Titles 模糊电视节目标题推文采集训练数据的自动标注
Pub Date : 2013-09-08 DOI: 10.1109/SocialCom.2013.119
M. Erdmann, Erik Ward, K. Ikeda, Gen Hattori, C. Ono, Y. Takishima
Twitter is a popular medium for sharing opinions on TV programs, and the analysis of TV related tweets is attracting a lot of interest. However, when collecting all tweets containing a given TV program title, we obtain a large number of unrelated tweets, due to the fact that many of the TV program titles are ambiguous. Using supervised learning, TV related tweets can be collected with high accuracy. The goal of our proposed method is to automate the labeling process, in order to eliminate the cost required for data labeling without sacrificing classification accuracy. When creating the training data, we use only tweets of unambiguous TV program titles. In order to decide whether a TV program title is ambiguous, we automatically determine whether it can be used as a common expression or named entity. In two experiments, in which we collected tweets for 32 ambiguous TV program titles, we achieved the same (78.2%) or even higher classification accuracy (79.1%) with automatically labeled training data as with manually labeled data, while effectively eliminating labeling costs.
Twitter是一种流行的分享电视节目意见的媒体,对电视相关推文的分析引起了很多人的兴趣。然而,当收集包含给定电视节目标题的所有tweet时,我们会得到大量不相关的tweet,因为许多电视节目标题是模糊的。利用监督学习,可以以较高的准确率收集电视相关的推文。我们提出的方法的目标是自动化标注过程,以便在不牺牲分类精度的情况下消除数据标注所需的成本。在创建训练数据时,我们只使用具有明确电视节目标题的tweet。为了确定电视节目标题是否有歧义,我们自动确定它是否可以用作公共表达或命名实体。在两个实验中,我们收集了32个有歧义的电视节目标题的推文,我们使用自动标记的训练数据获得了与手动标记数据相同(78.2%)甚至更高的分类准确率(79.1%),同时有效地消除了标记成本。
{"title":"Automatic Labeling of Training Data for Collecting Tweets for Ambiguous TV Program Titles","authors":"M. Erdmann, Erik Ward, K. Ikeda, Gen Hattori, C. Ono, Y. Takishima","doi":"10.1109/SocialCom.2013.119","DOIUrl":"https://doi.org/10.1109/SocialCom.2013.119","url":null,"abstract":"Twitter is a popular medium for sharing opinions on TV programs, and the analysis of TV related tweets is attracting a lot of interest. However, when collecting all tweets containing a given TV program title, we obtain a large number of unrelated tweets, due to the fact that many of the TV program titles are ambiguous. Using supervised learning, TV related tweets can be collected with high accuracy. The goal of our proposed method is to automate the labeling process, in order to eliminate the cost required for data labeling without sacrificing classification accuracy. When creating the training data, we use only tweets of unambiguous TV program titles. In order to decide whether a TV program title is ambiguous, we automatically determine whether it can be used as a common expression or named entity. In two experiments, in which we collected tweets for 32 ambiguous TV program titles, we achieved the same (78.2%) or even higher classification accuracy (79.1%) with automatically labeled training data as with manually labeled data, while effectively eliminating labeling costs.","PeriodicalId":129308,"journal":{"name":"2013 International Conference on Social Computing","volume":"111 3S 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131968723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Robust, Scalable Anomaly Detection for Large Collections of Images 鲁棒的、可扩展的大型图像异常检测
Pub Date : 2013-09-08 DOI: 10.1109/SocialCom.2013.170
Michael S. Kim
A novel robust anomaly detection algorithm is applied to an image dataset using Apache Pig, Jython and GNU Octave. Each image in the set is transformed into a feature vector that represents color, edges, and texture numerically. Data is streamed using Pig through standard and user defined GNU Octave functions for feature transformation. Once the image set is transformed into the feature space, the dataset matrix (where the rows are distinct images, and the columns are features) is input into an original anomaly detection algorithm written by the author. This unsupervised outlier detection method scores outliers in linear time. The method is linear in the number of outliers but still suffers from the curse of dimensionality (in the feature space). The top scoring images are considered anomalies. Two experiments are conducted. The first experiment tests if top scoring images coincide with images which are marked as outliers in a prior image selection step. The second examines the scalability of the implementation in Pig using a larger data set. The results are analyzed quantitatively and qualitatively.
基于Apache Pig、Jython和GNU Octave,提出了一种新的鲁棒异常检测算法。集合中的每个图像都被转换成一个特征向量,该特征向量以数字形式表示颜色、边缘和纹理。数据流使用Pig通过标准和用户定义的GNU Octave函数进行特征转换。将图像集转换为特征空间后,将数据集矩阵(行为不同图像,列为特征)输入到作者编写的原始异常检测算法中。这种无监督异常点检测方法在线性时间内对异常点进行评分。该方法在异常值的数量上是线性的,但仍然受到维度的诅咒(在特征空间中)。得分最高的图像被认为是异常的。进行了两个实验。第一个实验测试得分最高的图像是否与先前图像选择步骤中标记为异常值的图像一致。第二部分使用更大的数据集检查Pig中实现的可伸缩性。对结果进行了定量和定性分析。
{"title":"Robust, Scalable Anomaly Detection for Large Collections of Images","authors":"Michael S. Kim","doi":"10.1109/SocialCom.2013.170","DOIUrl":"https://doi.org/10.1109/SocialCom.2013.170","url":null,"abstract":"A novel robust anomaly detection algorithm is applied to an image dataset using Apache Pig, Jython and GNU Octave. Each image in the set is transformed into a feature vector that represents color, edges, and texture numerically. Data is streamed using Pig through standard and user defined GNU Octave functions for feature transformation. Once the image set is transformed into the feature space, the dataset matrix (where the rows are distinct images, and the columns are features) is input into an original anomaly detection algorithm written by the author. This unsupervised outlier detection method scores outliers in linear time. The method is linear in the number of outliers but still suffers from the curse of dimensionality (in the feature space). The top scoring images are considered anomalies. Two experiments are conducted. The first experiment tests if top scoring images coincide with images which are marked as outliers in a prior image selection step. The second examines the scalability of the implementation in Pig using a larger data set. The results are analyzed quantitatively and qualitatively.","PeriodicalId":129308,"journal":{"name":"2013 International Conference on Social Computing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130271455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A Framework for Secure Service Composition 用于安全服务组合的框架
Pub Date : 2013-09-08 DOI: 10.1109/SocialCom.2013.97
Achim D. Brucker, Francesco Malmignati, M. Merabti, Q. Shi, Bo Zhou
Modern applications are inherently heterogeneous: they are built by composing loosely coupled services that are, usually, offered and operated by different service providers. While this approach increases the flexibility of the composed applications, it makes the implementation of security and trustworthiness requirements difficult. As the number of security requirements is increasing dramatically, there is a need for new approaches that integrate security requirements right from the beginning while composing service-based applications. In this paper, we present a framework for secure service composition using a model-based approach for specifying, building, and executing composed services. As a unique feature, this framework integrates security requirements as a first class citizen and, thus, avoids the ``security as an afterthought'' paradigm.
现代应用程序本质上是异构的:它们是通过组合松散耦合的服务构建的,这些服务通常由不同的服务提供者提供和操作。虽然这种方法增加了组合应用程序的灵活性,但它使安全性和可信性需求的实现变得困难。由于安全需求的数量急剧增加,因此需要一种新的方法,在组合基于服务的应用程序时从一开始就集成安全需求。在本文中,我们提出了一个安全服务组合的框架,该框架使用基于模型的方法来指定、构建和执行组合的服务。作为一个独特的特性,该框架将安全性需求作为头等公民集成,从而避免了“事后考虑安全性”的范例。
{"title":"A Framework for Secure Service Composition","authors":"Achim D. Brucker, Francesco Malmignati, M. Merabti, Q. Shi, Bo Zhou","doi":"10.1109/SocialCom.2013.97","DOIUrl":"https://doi.org/10.1109/SocialCom.2013.97","url":null,"abstract":"Modern applications are inherently heterogeneous: they are built by composing loosely coupled services that are, usually, offered and operated by different service providers. While this approach increases the flexibility of the composed applications, it makes the implementation of security and trustworthiness requirements difficult. As the number of security requirements is increasing dramatically, there is a need for new approaches that integrate security requirements right from the beginning while composing service-based applications. In this paper, we present a framework for secure service composition using a model-based approach for specifying, building, and executing composed services. As a unique feature, this framework integrates security requirements as a first class citizen and, thus, avoids the ``security as an afterthought'' paradigm.","PeriodicalId":129308,"journal":{"name":"2013 International Conference on Social Computing","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131644532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
期刊
2013 International Conference on Social Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1