首页 > 最新文献

Computational Communication Research最新文献

英文 中文
The Accuracy and Precision of Measurement 测量的准确度和精密度
Pub Date : 2021-01-01 DOI: 10.5117/ccr2021.2.001.calc
Leandro A. Calcagnotto, Richard Huskey, Gerald M. Kosicki
Measurement noise differs by instrument and limits the validity and reliability of findings. Researchers collecting reaction time data introduce noise in the form of response time latency from hardware and software, even when collecting data on standardized computer-based experimental equipment. Reaction time is a measure with broad application for studying cognitive processing in communication research that is vulnerable to response latency noise. In this study, we utilized an Arduino microcontroller to generate a ground truth value of average response time latency in Asteroid Impact, an open source, naturalistic, experimental video game stimulus. We tested if response time latency differed across computer operating system, software, and trial modality. Here we show that reaction time measurements collected using Asteroid Impact were susceptible to response latency variability on par with other response-latency measuring software tests. These results demonstrate that Asteroid Impact is a valid and reliable stimulus for measuring reaction time data. Moreover, we provide researchers with a low-cost and open-source tool for evaluating response time latency in their own labs. Our results highlight the importance of validating measurement tools and support the philosophy of contributing methodological improvements in communication science.
测量噪声因仪器而异,并限制了结果的有效性和可靠性。收集反应时间数据的研究人员引入了以硬件和软件的反应时间延迟形式出现的噪声,即使是在标准化的基于计算机的实验设备上收集数据。在交际研究中,反应时间是一种具有广泛应用价值的研究认知加工的测量方法,极易受到反应滞后噪声的影响。在本研究中,我们利用Arduino微控制器在Asteroid Impact(一个开源的、自然的、实验性的视频游戏刺激)中生成平均响应时间延迟的ground真值。我们测试了不同计算机操作系统、软件和试验模式的响应时间延迟是否不同。在这里,我们展示了使用小行星撞击收集的反应时间测量值与其他响应延迟测量软件测试一样容易受到响应延迟变化的影响。这些结果表明,小行星撞击是测量反应时间数据的有效和可靠的刺激。此外,我们为研究人员提供了一种低成本和开源的工具,用于在他们自己的实验室中评估响应时间延迟。我们的结果强调了验证测量工具的重要性,并支持了在通信科学中贡献方法改进的哲学。
{"title":"The Accuracy and Precision of Measurement","authors":"Leandro A. Calcagnotto, Richard Huskey, Gerald M. Kosicki","doi":"10.5117/ccr2021.2.001.calc","DOIUrl":"https://doi.org/10.5117/ccr2021.2.001.calc","url":null,"abstract":"\u0000 Measurement noise differs by instrument and limits the validity and reliability of findings. Researchers collecting reaction time data introduce noise in the form of response time latency from hardware and software, even when collecting data on standardized computer-based experimental equipment. Reaction time is a measure with broad application for studying cognitive processing in communication research that is vulnerable to response latency noise. In this study, we utilized an Arduino microcontroller to generate a ground truth value of average response time latency in Asteroid Impact, an open source, naturalistic, experimental video game stimulus. We tested if response time latency differed across computer operating system, software, and trial modality. Here we show that reaction time measurements collected using Asteroid Impact were susceptible to response latency variability on par with other response-latency measuring software tests. These results demonstrate that Asteroid Impact is a valid and reliable stimulus for measuring reaction time data. Moreover, we provide researchers with a low-cost and open-source tool for evaluating response time latency in their own labs. Our results highlight the importance of validating measurement tools and support the philosophy of contributing methodological improvements in communication science.","PeriodicalId":275035,"journal":{"name":"Computational Communication Research","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117070022","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Extracting semantic relations using syntax 使用语法提取语义关系
Pub Date : 2021-01-01 DOI: 10.5117/ccr2021.2.003.welb
Kasper Welbers, W. Atteveldt, J. Kleinnijenhuis
Most common methods for automatic text analysis in communication science ignore syntactic information, focusing on the occurrence and co-occurrence of individual words, and sometimes n-grams. This is remarkably effective for some purposes, but poses a limitation for fine-grained analyses into semantic relations such as who does what to whom and according to what source. One tested, effective method for moving beyond this bag-of-words assumption is to use a rule-based approach for labeling and extracting syntactic patterns in dependency trees. Although this method can be used for a variety of purposes, its application is hindered by the lack of dedicated and accessible tools. In this paper we introduce the rsyntax R package, which is designed to make working with dependency trees easier and more intuitive for R users, and provides a framework for combining multiple rules for reliably extracting useful semantic relations.
在通信科学中,大多数常用的自动文本分析方法都忽略了句法信息,只关注单个单词的出现和共现,有时也关注n-gram。这对于某些目的来说非常有效,但是对语义关系的细粒度分析(比如谁对谁做什么,根据什么来源做什么)造成了限制。一种经过测试的有效方法可以超越这种词袋假设,即使用基于规则的方法来标记和提取依赖树中的语法模式。虽然这种方法可以用于各种目的,但由于缺乏专用的和可访问的工具,它的应用受到阻碍。在本文中,我们介绍了rsyntax R包,它旨在使R用户更容易和更直观地使用依赖树,并提供了一个框架来组合多个规则,以可靠地提取有用的语义关系。
{"title":"Extracting semantic relations using syntax","authors":"Kasper Welbers, W. Atteveldt, J. Kleinnijenhuis","doi":"10.5117/ccr2021.2.003.welb","DOIUrl":"https://doi.org/10.5117/ccr2021.2.003.welb","url":null,"abstract":"\u0000 Most common methods for automatic text analysis in communication science ignore syntactic information, focusing on the occurrence and co-occurrence of individual words, and sometimes n-grams. This is remarkably effective for some purposes, but poses a limitation for fine-grained analyses into semantic relations such as who does what to whom and according to what source. One tested, effective method for moving beyond this bag-of-words assumption is to use a rule-based approach for labeling and extracting syntactic patterns in dependency trees. Although this method can be used for a variety of purposes, its application is hindered by the lack of dedicated and accessible tools. In this paper we introduce the rsyntax R package, which is designed to make working with dependency trees easier and more intuitive for R users, and provides a framework for combining multiple rules for reliably extracting useful semantic relations.","PeriodicalId":275035,"journal":{"name":"Computational Communication Research","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121357866","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Four best practices for measuring news sentiment using ‘off-the-shelf’ dictionaries: a large-scale p-hacking experiment 使用“现成”词典衡量新闻情绪的四个最佳实践:大规模p-hacking实验
Pub Date : 2020-10-07 DOI: 10.31235/osf.io/np5wa
Chung-hong Chan, Joseph W. Bajjalieh, L. Auvil, Hartmut Wessler, Scott L. Althaus, Kasper Welbers, Wouter van Atteveldt, Marc Jungblut
We examined the validity of 37 sentiment scores based on dictionary-based methods using a large news corpus and demonstrated the risk of generating a spectrum of results with different levels of statistical significance by presenting an analysis of relationships between news sentiment and U.S. presidential approval. We summarize our findings into four best practices: 1) use a suitable sentiment dictionary; 2) do not assume that the validity and reliability of the dictionary is ‘built-in’; 3) check for the influence of content length and 4) do not use multiple dictionaries to test the same statistical hypothesis.
我们使用一个大型新闻语料库,基于基于词典的方法检验了37种情绪得分的有效性,并通过分析新闻情绪与美国总统支持率之间的关系,展示了产生具有不同统计显著性水平的结果谱的风险。我们将研究结果总结为四个最佳实践:1)使用合适的情感词典;2)不要认为字典的有效性和可靠性是“内置的”;3)检查内容长度的影响,4)不要使用多个字典来检验相同的统计假设。
{"title":"Four best practices for measuring news sentiment using ‘off-the-shelf’ dictionaries: a large-scale p-hacking experiment","authors":"Chung-hong Chan, Joseph W. Bajjalieh, L. Auvil, Hartmut Wessler, Scott L. Althaus, Kasper Welbers, Wouter van Atteveldt, Marc Jungblut","doi":"10.31235/osf.io/np5wa","DOIUrl":"https://doi.org/10.31235/osf.io/np5wa","url":null,"abstract":"We examined the validity of 37 sentiment scores based on dictionary-based methods using a large news corpus and demonstrated the risk of generating a spectrum of results with different levels of statistical significance by presenting an analysis of relationships between news sentiment and U.S. presidential approval. We summarize our findings into four best practices: 1) use a suitable sentiment dictionary; 2) do not assume that the validity and reliability of the dictionary is ‘built-in’; 3) check for the influence of content length and 4) do not use multiple dictionaries to test the same statistical hypothesis.","PeriodicalId":275035,"journal":{"name":"Computational Communication Research","volume":"14 36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115014921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
How Document Sampling and Vocabulary Pruning Affect the Results of Topic Models 文档采样和词汇修剪如何影响主题模型的结果
Pub Date : 2019-11-20 DOI: 10.31219/osf.io/2rh6g
D. Maier, A. Niekler, Gregor Wiedemann, Daniela Stoltenberg
Topic modeling enables researchers to explore large document corpora. Large corpora, however, can be extremely costly to model in terms of time and computing resources. In order to circumvent this problem, two techniques have been suggested: (1) to model random document samples, and (2) to prune the vocabulary of the corpus. Although frequently applied, there has been no systematic inquiry into how the application of these techniques affects the respective models. Using three empirical corpora with different characteristics (news articles, websites, and Tweets), we systematically investigated how different sample sizes and pruning affect the resulting topic models in comparison to models of the full corpora. Our inquiry provides evidence that both techniques are viable tools that will likely not impair the resulting model. Sample-based topic models closely resemble corpus-based models if the sample size is large enough (> 10,000 documents). Moreover, extensive pruning does not compromise the quality of the resultant topics.
主题建模使研究人员能够探索大型文档语料库。然而,就时间和计算资源而言,大型语料库的建模成本非常高。为了避免这个问题,提出了两种技术:(1)对随机文档样本进行建模,(2)对语料库中的词汇进行修剪。虽然这些技术经常被应用,但没有系统地研究这些技术的应用如何影响各自的模型。使用三个具有不同特征的经验语料库(新闻文章、网站和推文),我们系统地研究了不同样本量和修剪如何影响最终的主题模型,并与完整语料库的模型进行了比较。我们的调查提供了证据,证明这两种技术都是可行的工具,可能不会损害最终的模型。如果样本量足够大(> 10,000个文档),基于样本的主题模型与基于语料库的模型非常相似。此外,广泛的修剪并不会影响生成主题的质量。
{"title":"How Document Sampling and Vocabulary Pruning Affect the Results of Topic Models","authors":"D. Maier, A. Niekler, Gregor Wiedemann, Daniela Stoltenberg","doi":"10.31219/osf.io/2rh6g","DOIUrl":"https://doi.org/10.31219/osf.io/2rh6g","url":null,"abstract":"\u0000 Topic modeling enables researchers to explore large document corpora. Large corpora, however, can be extremely costly to model in terms of time and computing resources. In order to circumvent this problem, two techniques have been suggested: (1) to model random document samples, and (2) to prune the vocabulary of the corpus. Although frequently applied, there has been no systematic inquiry into how the application of these techniques affects the respective models. Using three empirical corpora with different characteristics (news articles, websites, and Tweets), we systematically investigated how different sample sizes and pruning affect the resulting topic models in comparison to models of the full corpora. Our inquiry provides evidence that both techniques are viable tools that will likely not impair the resulting model. Sample-based topic models closely resemble corpus-based models if the sample size is large enough (> 10,000 documents). Moreover, extensive pruning does not compromise the quality of the resultant topics.","PeriodicalId":275035,"journal":{"name":"Computational Communication Research","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129245278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
3bij3 – Developing a framework for researching recommender systems and their effects 3bij3 -开发一个研究推荐系统及其效果的框架
Pub Date : 2019-10-02 DOI: 10.31235/osf.io/vw2dr
Felicia Loecherbach, D. Trilling
Today’s online news environment is increasingly characterized by personalized news selections, relying on algorithmic solutions for extracting relevant articles and composing an individual’s news diet. Yet, the impact of such recommendation algorithms on how we consume and perceive news is still understudied. We therefore developed one of the first software solutions to conduct studies on effects of news recommender systems in a realistic setting. The web app of our framework (called 3bij3) displays real-time news articles selected by different mechanisms. 3bij3 can be used to conduct large-scale field experiments, in which participants’ use of the site can be tracked over extended periods of time. Compared to previous work, 3bij3 gives researchers control over the recommendation system under study and creates a realistic environment for the participants. It integrates web scraping, different methods to compare and classify news articles, different recommender systems, a web interface for participants, gamification elements, and a user survey to enrich the behavioural measures obtained.
今天的网络新闻环境越来越以个性化的新闻选择为特征,依靠算法解决方案提取相关文章,构成个人的新闻饮食。然而,这种推荐算法对我们如何消费和感知新闻的影响仍未得到充分研究。因此,我们开发了第一个软件解决方案,在现实环境中对新闻推荐系统的影响进行研究。我们框架的web应用程序(称为3bij3)显示通过不同机制选择的实时新闻文章。3bij3可以用来进行大规模的现场实验,在实验中可以长时间跟踪参与者对该网站的使用情况。与之前的工作相比,3bij3使研究人员能够控制所研究的推荐系统,并为参与者创造了一个真实的环境。它集成了网络抓取、不同的新闻文章比较和分类方法、不同的推荐系统、参与者的web界面、游戏化元素和用户调查,以丰富所获得的行为测量。
{"title":"3bij3 – Developing a framework for researching recommender systems and their effects","authors":"Felicia Loecherbach, D. Trilling","doi":"10.31235/osf.io/vw2dr","DOIUrl":"https://doi.org/10.31235/osf.io/vw2dr","url":null,"abstract":"\u0000 Today’s online news environment is increasingly characterized by personalized news selections, relying on algorithmic solutions for extracting relevant articles and composing an individual’s news diet. Yet, the impact of such recommendation algorithms on how we consume and perceive news is still understudied. We therefore developed one of the first software solutions to conduct studies on effects of news recommender systems in a realistic setting. The web app of our framework (called 3bij3) displays real-time news articles selected by different mechanisms. 3bij3 can be used to conduct large-scale field experiments, in which participants’ use of the site can be tracked over extended periods of time. Compared to previous work, 3bij3 gives researchers control over the recommendation system under study and creates a realistic environment for the participants. It integrates web scraping, different methods to compare and classify news articles, different recommender systems, a web interface for participants, gamification elements, and a user survey to enrich the behavioural measures obtained.","PeriodicalId":275035,"journal":{"name":"Computational Communication Research","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130020891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
News Organizations’ Selective Link Sharing as Gatekeeping 新闻机构的选择性链接共享作为把关
Pub Date : 2019-10-01 DOI: 10.5117/ccr2019.1.003.pak
Chankyung Pak
To disseminate their stories efficiently via social media, news organizations make decisions that resemble traditional editorial decisions. However, the decisions for social media may deviate from traditional ones because they are often made outside the newsroom and guided by audience metrics. This study focuses on selective link sharing, as quasi-gatekeeping, on Twitter -- conditioning a link sharing decision about news content and illustrates how it resembles and deviates from gatekeeping for the publication of news stories. Using a computational data collection method and a machine learning technique called Structural Topic Model (STM), this study shows that selective link sharing generates different topic distribution between news websites and Twitter, significantly revoking the specialty of news organizations. This finding implies that emergent logic, which governs news organizations' decisions for social media can undermine the provision of diverse news, which relies on journalistic values and norms.
为了通过社交媒体有效地传播他们的故事,新闻机构做出了类似于传统编辑决策的决定。然而,社交媒体的决策可能会偏离传统决策,因为它们通常是在新闻编辑室之外做出的,并以受众指标为指导。本研究侧重于选择性链接共享,作为准守门人,在Twitter上调节关于新闻内容的链接共享决策,并说明它如何与新闻故事发布的守门人相似和偏离。本研究使用一种计算数据收集方法和一种称为结构主题模型(Structural Topic Model, STM)的机器学习技术,表明选择性链接共享在新闻网站和Twitter之间产生了不同的主题分布,显著地撤销了新闻机构的特殊性。这一发现意味着,支配新闻机构对社交媒体决策的涌现逻辑,可能会破坏依赖于新闻价值观和规范的多样化新闻的提供。
{"title":"News Organizations’ Selective Link Sharing as Gatekeeping","authors":"Chankyung Pak","doi":"10.5117/ccr2019.1.003.pak","DOIUrl":"https://doi.org/10.5117/ccr2019.1.003.pak","url":null,"abstract":"To disseminate their stories efficiently via social media, news organizations make decisions that resemble traditional editorial decisions. However, the decisions for social media may deviate from traditional ones because they are often made outside the newsroom and guided by audience metrics. This study focuses on selective link sharing, as quasi-gatekeeping, on Twitter -- conditioning a link sharing decision about news content and illustrates how it resembles and deviates from gatekeeping for the publication of news stories. Using a computational data collection method and a machine learning technique called Structural Topic Model (STM), this study shows that selective link sharing generates different topic distribution between news websites and Twitter, significantly revoking the specialty of news organizations. This finding implies that emergent logic, which governs news organizations' decisions for social media can undermine the provision of diverse news, which relies on journalistic values and norms.","PeriodicalId":275035,"journal":{"name":"Computational Communication Research","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134555616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Computational observation 计算观察
Pub Date : 2019-10-01 DOI: 10.5117/ccr2019.1.004.haim
Mario Haim, Angela Nienierza
A lot of modern media use is guided by algorithmic curation, a phenomenon that is in desperate need of empirical observation, but for which adequate methodological tools are largely missing. To fill this gap, computational observation offers a novel approach—the unobtrusive and automated collection of information encountered within algorithmically curated media environments by means of a browser plug-in. In contrast to prior methodological approaches, browser plug-ins allow for reliable capture and repetitive analysis of both content and context at the point of the actual user encounter. After discussing the technological, ethical, and practical considerations relevant to this automated solution, we present our open-source browser plug-in as an element in an adequate multi-method design, along with potential links to panel surveys and content analysis. Ultimately, we present a proof-of-concept study in the realm of news exposure on Facebook; we successfully deployed the plug-in to Chrome and Firefox, and we combined the computational observation with a two-wave panel survey. Although this study suffered from severe recruitment difficulties, the results indicate that the methodological setup is reliable and ready to implement for data collection within a variety of studies on media use and media effects.
许多现代媒体的使用都是由算法管理引导的,这是一种迫切需要经验观察的现象,但在很大程度上缺乏适当的方法工具。为了填补这一空白,计算观察提供了一种新颖的方法——通过浏览器插件,在算法策划的媒体环境中不引人注目地自动收集信息。与先前的方法方法不同,浏览器插件允许在实际用户遇到时可靠地捕获和重复分析内容和上下文。在讨论了与此自动化解决方案相关的技术、伦理和实际考虑因素之后,我们将我们的开源浏览器插件作为适当的多方法设计中的一个元素,以及与面板调查和内容分析的潜在链接。最后,我们提出了一项关于Facebook新闻曝光领域的概念验证研究;我们成功地将插件部署到Chrome和Firefox上,并将计算观察与两波面板调查相结合。虽然这项研究在招募方面遇到了严重的困难,但结果表明,方法设置是可靠的,并且可以在各种关于媒体使用和媒体影响的研究中实施数据收集。
{"title":"Computational observation","authors":"Mario Haim, Angela Nienierza","doi":"10.5117/ccr2019.1.004.haim","DOIUrl":"https://doi.org/10.5117/ccr2019.1.004.haim","url":null,"abstract":"A lot of modern media use is guided by algorithmic curation, a phenomenon that is in desperate need of empirical observation, but for which adequate methodological tools are largely missing. To fill this gap, computational observation offers a novel approach—the unobtrusive and automated collection of information encountered within algorithmically curated media environments by means of a browser plug-in. In contrast to prior methodological approaches, browser plug-ins allow for reliable capture and repetitive analysis of both content and context at the point of the actual user encounter. After discussing the technological, ethical, and practical considerations relevant to this automated solution, we present our open-source browser plug-in as an element in an adequate multi-method design, along with potential links to panel surveys and content analysis. Ultimately, we present a proof-of-concept study in the realm of news exposure on Facebook; we successfully deployed the plug-in to Chrome and Firefox, and we combined the computational observation with a two-wave panel survey. Although this study suffered from severe recruitment difficulties, the results indicate that the methodological setup is reliable and ready to implement for data collection within a variety of studies on media use and media effects.","PeriodicalId":275035,"journal":{"name":"Computational Communication Research","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124507638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
A Weakly Supervised and Deep Learning Method for an Additive Topic Analysis of Large Corpora 大型语料库加性主题分析的弱监督深度学习方法
Pub Date : 2019-07-11 DOI: 10.31235/osf.io/nfr3p
Yair Fogel-Dror, Shaul R. Shenhav, Tamir Sheafer
The collaborative effort of theory-driven content analysis can benefit significantly from the use of topic analysis methods, which allow researchers to add more categories while developing or testing a theory. This additive approach enables the reuse of previous efforts of analysis or even the merging of separate research projects, thereby making these methods more accessible and increasing the discipline’s ability to create and share content analysis capabilities. This paper proposes a weakly supervised topic analysis method that uses both a low-cost unsupervised method to compile a training set and supervised deep learning as an additive and accurate text classification method. We test the validity of the method, specifically its additivity, by comparing the results of the method after adding 200 categories to an initial number of 450. We show that the suggested method provides a foundation for a low-cost solution for large-scale topic analysis.
理论驱动的内容分析的协作工作可以从主题分析方法的使用中显著受益,这允许研究人员在开发或测试理论时添加更多的类别。这种附加的方法可以重用以前的分析工作,甚至可以合并单独的研究项目,从而使这些方法更容易访问,并增加学科创建和共享内容分析功能的能力。本文提出了一种弱监督主题分析方法,该方法既使用低成本的无监督方法编译训练集,又使用监督深度学习作为一种加性和精确的文本分类方法。我们测试方法的有效性,特别是它的可加性,通过比较方法的结果后,200个类别的初始数量为450。我们表明,该方法为大规模主题分析的低成本解决方案提供了基础。
{"title":"A Weakly Supervised and Deep Learning Method for an Additive Topic Analysis of Large Corpora","authors":"Yair Fogel-Dror, Shaul R. Shenhav, Tamir Sheafer","doi":"10.31235/osf.io/nfr3p","DOIUrl":"https://doi.org/10.31235/osf.io/nfr3p","url":null,"abstract":"The collaborative effort of theory-driven content analysis can benefit significantly from the use of topic analysis methods, which allow researchers to add more categories while developing or testing a theory. This additive approach enables the reuse of previous efforts of analysis or even the merging of separate research projects, thereby making these methods more accessible and increasing the discipline’s ability to create and share content analysis capabilities. This paper proposes a weakly supervised topic analysis method that uses both a low-cost unsupervised method to compile a training set and supervised deep learning as an additive and accurate text classification method. We test the validity of the method, specifically its additivity, by comparing the results of the method after adding 200 categories to an initial number of 450. We show that the suggested method provides a foundation for a low-cost solution for large-scale topic analysis.","PeriodicalId":275035,"journal":{"name":"Computational Communication Research","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132953092","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Roadmap for Computational Communication Research 计算通信研究的路线图
Pub Date : 2019-05-20 DOI: 10.31235/osf.io/4dhfk
Wouter van Atteveldt, Drew B. Margolin, Cuihua Shen, D. Trilling, R. Weber
Computational Communication Research (CCR) is a new open access journal dedicated to publishing high quality computational research in communication science. This editorial introduction describes the role that we envision for the journal. First, we explain what computational communication science is and why a new journal is needed for this subfield. Then, we elaborate on the type of research this journal seeks to publish, and stress the need for transparent and reproducible science. The relation between theoretical development and computational analysis is discussed, and we argue for the value of null-findings and risky research in additive science. Subsequently, the (experimental) two-phase review process is described. In this process, after the first double-blind review phase, an editor can signal that they intend to publish the article conditional on satisfactory revisions. This starts the second review phase, in which authors and reviewers are no longer required to be anonymous and the authors are encouraged to publish a preprint to their article which will be linked as working paper from the journal. Finally, we introduce the four articles that, together with this Introduction, form the inaugural issue.
《计算通信研究》(Computational Communication Research, CCR)是一本新的开放获取期刊,致力于发表高质量的通信科学计算研究。这篇编辑导言描述了我们对期刊的设想。首先,我们解释了什么是计算通信科学,以及为什么这个子领域需要一本新的期刊。然后,我们详细说明了该期刊寻求发表的研究类型,并强调了透明和可再生科学的必要性。讨论了理论发展与计算分析之间的关系,并论证了零发现和风险研究在加性科学中的价值。随后,描述了(实验)两阶段评审过程。在这个过程中,在第一个双盲审查阶段之后,编辑可以发出信号,表示他们打算以令人满意的修改为条件发表文章。这就开始了第二审稿阶段,作者和审稿人不再需要匿名,并鼓励作者发表文章的预印本,该预印本将被链接为期刊的工作论文。最后,我们将介绍与本导言一起构成创刊号的四篇文章。
{"title":"A Roadmap for Computational Communication Research","authors":"Wouter van Atteveldt, Drew B. Margolin, Cuihua Shen, D. Trilling, R. Weber","doi":"10.31235/osf.io/4dhfk","DOIUrl":"https://doi.org/10.31235/osf.io/4dhfk","url":null,"abstract":"Computational Communication Research (CCR) is a new open access journal dedicated to publishing high quality computational research in communication science. This editorial introduction describes the role that we envision for the journal. First, we explain what computational communication science is and why a new journal is needed for this subfield. Then, we elaborate on the type of research this journal seeks to publish, and stress the need for transparent and reproducible science. The relation between theoretical development and computational analysis is discussed, and we argue for the value of null-findings and risky research in additive science. Subsequently, the (experimental) two-phase review process is described. In this process, after the first double-blind review phase, an editor can signal that they intend to publish the article conditional on satisfactory revisions. This starts the second review phase, in which authors and reviewers are no longer required to be anonymous and the authors are encouraged to publish a preprint to their article which will be linked as working paper from the journal. Finally, we introduce the four articles that, together with this Introduction, form the inaugural issue.","PeriodicalId":275035,"journal":{"name":"Computational Communication Research","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124477764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
iCoRe: The GDELT Interface for the Advancement of Communication Research iCoRe:促进通信研究的GDELT接口
Pub Date : 2019-05-17 DOI: 10.31235/osf.io/smjwb
F. R. Hopp, J. Schaffer, J. T. Fisher, R. Weber
This article introduces the interface for communication research (iCoRe) to access, explore, and analyze the Global Database of Events, Language and Tone (GDELT; Leetaru & Schrodt, 2013). GDELT provides a vast, open source, and continuously updated repository of online news and event metadata collected from tens of thousands of news outlets around the world. Despite GDELT’s promise for advancing communication science, its massive scale and complex data structures have hindered efforts of communication scholars aiming to access and analyze GDELT. We thus developed iCoRe, an easy-to-use web interface that (a) provides fast access to the data available in GDELT, (b) shapes and processes GDELT for theory-driven applications within communication research, and (c) enables replicability through transparent query and analysis protocols. After providing an overview of how GDELT’s data pertain to addressing communication research questions, we provide a tutorial of utilizing iCoRe across three theory-driven case studies. We conclude this article with a discussion and outlook of iCoRe’s future potential for advancing communication research.
本文介绍了用于传播研究(iCoRe)的接口,用于访问、探索和分析全球事件、语言和语气数据库(GDELT;Leetaru & Schrodt, 2013)。GDELT提供了一个巨大的、开源的、不断更新的在线新闻和事件元数据存储库,这些新闻和事件元数据是从世界各地数以万计的新闻媒体收集的。尽管GDELT有望推动通信科学的发展,但其庞大的规模和复杂的数据结构阻碍了通信学者对GDELT的访问和分析。因此,我们开发了iCoRe,这是一个易于使用的web界面,它(a)提供对GDELT中可用数据的快速访问,(b)在通信研究中为理论驱动的应用程序设计和处理GDELT,以及(c)通过透明的查询和分析协议实现可复制性。在概述了GDELT的数据如何与解决传播研究问题相关之后,我们提供了一个在三个理论驱动的案例研究中使用iCoRe的教程。最后,我们对iCoRe在推进传播研究方面的未来潜力进行了讨论和展望。
{"title":"iCoRe: The GDELT Interface for the Advancement of Communication Research","authors":"F. R. Hopp, J. Schaffer, J. T. Fisher, R. Weber","doi":"10.31235/osf.io/smjwb","DOIUrl":"https://doi.org/10.31235/osf.io/smjwb","url":null,"abstract":"This article introduces the interface for communication research (iCoRe) to access, explore, and analyze the Global Database of Events, Language and Tone (GDELT; Leetaru & Schrodt, 2013). GDELT provides a vast, open source, and continuously updated repository of online news and event metadata collected from tens of thousands of news outlets around the world. Despite GDELT’s promise for advancing communication science, its massive scale and complex data structures have hindered efforts of communication scholars aiming to access and analyze GDELT. We thus developed iCoRe, an easy-to-use web interface that (a) provides fast access to the data available in GDELT, (b) shapes and processes GDELT for theory-driven applications within communication research, and (c) enables replicability through transparent query and analysis protocols. After providing an overview of how GDELT’s data pertain to addressing communication research questions, we provide a tutorial of utilizing iCoRe across three theory-driven case studies. We conclude this article with a discussion and outlook of iCoRe’s future potential for advancing communication research.","PeriodicalId":275035,"journal":{"name":"Computational Communication Research","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123783946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
期刊
Computational Communication Research
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1