首页 > 最新文献

European Conference on Information Retrieval最新文献

英文 中文
HADA: A Graph-based Amalgamation Framework in Image-text Retrieval 图像-文本检索中基于图的合并框架
Pub Date : 2023-01-11 DOI: 10.48550/arXiv.2301.04742
Manh-Duy Nguyen, Binh T. Nguyen, C. Gurrin
Many models have been proposed for vision and language tasks, especially the image-text retrieval task. All state-of-the-art (SOTA) models in this challenge contained hundreds of millions of parameters. They also were pretrained on a large external dataset that has been proven to make a big improvement in overall performance. It is not easy to propose a new model with a novel architecture and intensively train it on a massive dataset with many GPUs to surpass many SOTA models, which are already available to use on the Internet. In this paper, we proposed a compact graph-based framework, named HADA, which can combine pretrained models to produce a better result, rather than building from scratch. First, we created a graph structure in which the nodes were the features extracted from the pretrained models and the edges connecting them. The graph structure was employed to capture and fuse the information from every pretrained model with each other. Then a graph neural network was applied to update the connection between the nodes to get the representative embedding vector for an image and text. Finally, we used the cosine similarity to match images with their relevant texts and vice versa to ensure a low inference time. Our experiments showed that, although HADA contained a tiny number of trainable parameters, it could increase baseline performance by more than 3.6% in terms of evaluation metrics in the Flickr30k dataset. Additionally, the proposed model did not train on any external dataset and did not require many GPUs but only 1 to train due to its small number of parameters. The source code is available at https://github.com/m2man/HADA.
对于视觉和语言任务,特别是图像-文本检索任务,已经提出了许多模型。该挑战中的所有最先进(SOTA)模型都包含数亿个参数。它们还在一个大型的外部数据集上进行了预训练,这已被证明对整体性能有很大的提高。提出一个具有新颖架构的新模型,并在具有许多gpu的海量数据集上进行密集训练,以超越许多已经在互联网上使用的SOTA模型,这是一件不容易的事情。在本文中,我们提出了一个紧凑的基于图的框架,称为HADA,它可以结合预训练的模型来产生更好的结果,而不是从头开始构建。首先,我们创建了一个图结构,其中节点是从预训练模型中提取的特征和连接它们的边。利用图结构捕获和融合每个预训练模型的信息。然后利用图神经网络更新节点间的连接,得到具有代表性的图像和文本嵌入向量。最后,我们使用余弦相似度将图像与其相关文本进行匹配,反之亦然,以确保较低的推理时间。我们的实验表明,尽管HADA包含少量可训练参数,但就Flickr30k数据集的评估指标而言,它可以将基准性能提高3.6%以上。此外,所提出的模型没有在任何外部数据集上进行训练,并且由于其参数较少,只需要1个gpu即可训练,而不需要很多gpu。源代码可从https://github.com/m2man/HADA获得。
{"title":"HADA: A Graph-based Amalgamation Framework in Image-text Retrieval","authors":"Manh-Duy Nguyen, Binh T. Nguyen, C. Gurrin","doi":"10.48550/arXiv.2301.04742","DOIUrl":"https://doi.org/10.48550/arXiv.2301.04742","url":null,"abstract":"Many models have been proposed for vision and language tasks, especially the image-text retrieval task. All state-of-the-art (SOTA) models in this challenge contained hundreds of millions of parameters. They also were pretrained on a large external dataset that has been proven to make a big improvement in overall performance. It is not easy to propose a new model with a novel architecture and intensively train it on a massive dataset with many GPUs to surpass many SOTA models, which are already available to use on the Internet. In this paper, we proposed a compact graph-based framework, named HADA, which can combine pretrained models to produce a better result, rather than building from scratch. First, we created a graph structure in which the nodes were the features extracted from the pretrained models and the edges connecting them. The graph structure was employed to capture and fuse the information from every pretrained model with each other. Then a graph neural network was applied to update the connection between the nodes to get the representative embedding vector for an image and text. Finally, we used the cosine similarity to match images with their relevant texts and vice versa to ensure a low inference time. Our experiments showed that, although HADA contained a tiny number of trainable parameters, it could increase baseline performance by more than 3.6% in terms of evaluation metrics in the Flickr30k dataset. Additionally, the proposed model did not train on any external dataset and did not require many GPUs but only 1 to train due to its small number of parameters. The source code is available at https://github.com/m2man/HADA.","PeriodicalId":126309,"journal":{"name":"European Conference on Information Retrieval","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125394749","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Doc2Query-: When Less is More Doc2Query-:当Less is More
Pub Date : 2023-01-09 DOI: 10.48550/arXiv.2301.03266
Mitko Gospodinov, Sean MacAvaney, C. Macdonald
Doc2Query -- the process of expanding the content of a document before indexing using a sequence-to-sequence model -- has emerged as a prominent technique for improving the first-stage retrieval effectiveness of search engines. However, sequence-to-sequence models are known to be prone to"hallucinating"content that is not present in the source text. We argue that Doc2Query is indeed prone to hallucination, which ultimately harms retrieval effectiveness and inflates the index size. In this work, we explore techniques for filtering out these harmful queries prior to indexing. We find that using a relevance model to remove poor-quality queries can improve the retrieval effectiveness of Doc2Query by up to 16%, while simultaneously reducing mean query execution time by 23% and cutting the index size by 33%. We release the code, data, and a live demonstration to facilitate reproduction and further exploration at https://github.com/terrierteam/pyterrier_doc2query.
Doc2Query——在使用序列到序列模型建立索引之前扩展文档内容的过程——已经成为提高搜索引擎第一阶段检索效率的重要技术。然而,已知序列到序列模型容易产生源文本中不存在的“幻觉”内容。我们认为,Doc2Query确实容易产生幻觉,这最终会损害检索效率并使索引大小膨胀。在这项工作中,我们将探索在建立索引之前过滤掉这些有害查询的技术。我们发现,使用关联模型去除低质量查询可以将Doc2Query的检索效率提高16%,同时将平均查询执行时间减少23%,将索引大小减少33%。我们在https://github.com/terrierteam/pyterrier_doc2query上发布了代码、数据和现场演示,以方便复制和进一步探索。
{"title":"Doc2Query-: When Less is More","authors":"Mitko Gospodinov, Sean MacAvaney, C. Macdonald","doi":"10.48550/arXiv.2301.03266","DOIUrl":"https://doi.org/10.48550/arXiv.2301.03266","url":null,"abstract":"Doc2Query -- the process of expanding the content of a document before indexing using a sequence-to-sequence model -- has emerged as a prominent technique for improving the first-stage retrieval effectiveness of search engines. However, sequence-to-sequence models are known to be prone to\"hallucinating\"content that is not present in the source text. We argue that Doc2Query is indeed prone to hallucination, which ultimately harms retrieval effectiveness and inflates the index size. In this work, we explore techniques for filtering out these harmful queries prior to indexing. We find that using a relevance model to remove poor-quality queries can improve the retrieval effectiveness of Doc2Query by up to 16%, while simultaneously reducing mean query execution time by 23% and cutting the index size by 33%. We release the code, data, and a live demonstration to facilitate reproduction and further exploration at https://github.com/terrierteam/pyterrier_doc2query.","PeriodicalId":126309,"journal":{"name":"European Conference on Information Retrieval","volume":"227 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116387633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Uptrendz: API-Centric Real-time Recommendations in Multi-Domain Settings Uptrendz:多域设置中以api为中心的实时推荐
Pub Date : 2023-01-03 DOI: 10.48550/arXiv.2301.01037
Emanuel Lacić, Tomislav Duricic, Leon Fadljevic, Dieter Theiler, Dominik Kowald
In this work, we tackle the problem of adapting a real-time recommender system to multiple application domains, and their underlying data models and customization requirements. To do that, we present Uptrendz, a multi-domain recommendation platform that can be customized to provide real-time recommendations in an API-centric way. We demonstrate (i) how to set up a real-time movie recommender using the popular MovieLens-100k dataset, and (ii) how to simultaneously support multiple application domains based on the use-case of recommendations in entrepreneurial start-up founding. For that, we differentiate between domains on the item- and system-level. We believe that our demonstration shows a convenient way to adapt, deploy and evaluate a recommender system in an API-centric way. The source-code and documentation that demonstrates how to utilize the configured Uptrendz API is available on GitHub.
在这项工作中,我们解决了使实时推荐系统适应多个应用领域的问题,以及它们的底层数据模型和定制需求。为了做到这一点,我们提出了Uptrendz,这是一个多领域推荐平台,可以定制以api为中心的方式提供实时推荐。我们演示了(i)如何使用流行的MovieLens-100k数据集建立实时电影推荐,以及(ii)如何同时支持基于创业创业中推荐用例的多个应用领域。为此,我们在项目级和系统级上区分域。我们相信,我们的演示展示了一种以api为中心的方式来适应、部署和评估推荐系统的方便方法。演示如何使用配置好的Uptrendz API的源代码和文档可以在GitHub上找到。
{"title":"Uptrendz: API-Centric Real-time Recommendations in Multi-Domain Settings","authors":"Emanuel Lacić, Tomislav Duricic, Leon Fadljevic, Dieter Theiler, Dominik Kowald","doi":"10.48550/arXiv.2301.01037","DOIUrl":"https://doi.org/10.48550/arXiv.2301.01037","url":null,"abstract":"In this work, we tackle the problem of adapting a real-time recommender system to multiple application domains, and their underlying data models and customization requirements. To do that, we present Uptrendz, a multi-domain recommendation platform that can be customized to provide real-time recommendations in an API-centric way. We demonstrate (i) how to set up a real-time movie recommender using the popular MovieLens-100k dataset, and (ii) how to simultaneously support multiple application domains based on the use-case of recommendations in entrepreneurial start-up founding. For that, we differentiate between domains on the item- and system-level. We believe that our demonstration shows a convenient way to adapt, deploy and evaluate a recommender system in an API-centric way. The source-code and documentation that demonstrates how to utilize the configured Uptrendz API is available on GitHub.","PeriodicalId":126309,"journal":{"name":"European Conference on Information Retrieval","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115486250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Visconde: Multi-document QA with GPT-3 and Neural Reranking 使用GPT-3和神经重排序的多文档QA
Pub Date : 2022-12-19 DOI: 10.48550/arXiv.2212.09656
Jayr Alencar Pereira, R. Fidalgo, R. Lotufo, Rodrigo Nogueira
This paper proposes a question-answering system that can answer questions whose supporting evidence is spread over multiple (potentially long) documents. The system, called Visconde, uses a three-step pipeline to perform the task: decompose, retrieve, and aggregate. The first step decomposes the question into simpler questions using a few-shot large language model (LLM). Then, a state-of-the-art search engine is used to retrieve candidate passages from a large collection for each decomposed question. In the final step, we use the LLM in a few-shot setting to aggregate the contents of the passages into the final answer. The system is evaluated on three datasets: IIRC, Qasper, and StrategyQA. Results suggest that current retrievers are the main bottleneck and that readers are already performing at the human level as long as relevant passages are provided. The system is also shown to be more effective when the model is induced to give explanations before answering a question. Code is available at url{https://github.com/neuralmind-ai/visconde}.
本文提出了一个问答系统,它可以回答那些支持证据分布在多个(可能很长的)文档中的问题。该系统名为Visconde,它使用三步管道来执行任务:分解、检索和聚合。第一步使用少量的大型语言模型(LLM)将问题分解为更简单的问题。然后,使用最先进的搜索引擎从每个分解问题的大集合中检索候选段落。在最后一步中,我们在几个镜头设置中使用LLM将段落的内容聚合到最终答案中。该系统在三个数据集上进行了评估:IIRC、Qasper和StrategyQA。结果表明,目前的检索器是主要的瓶颈,只要提供相关的段落,读者就已经达到了人类的水平。当模型在回答问题之前给出解释时,该系统也显示出更有效的效果。代码可从url{https://github.com/neuralmind-ai/visconde}获得。
{"title":"Visconde: Multi-document QA with GPT-3 and Neural Reranking","authors":"Jayr Alencar Pereira, R. Fidalgo, R. Lotufo, Rodrigo Nogueira","doi":"10.48550/arXiv.2212.09656","DOIUrl":"https://doi.org/10.48550/arXiv.2212.09656","url":null,"abstract":"This paper proposes a question-answering system that can answer questions whose supporting evidence is spread over multiple (potentially long) documents. The system, called Visconde, uses a three-step pipeline to perform the task: decompose, retrieve, and aggregate. The first step decomposes the question into simpler questions using a few-shot large language model (LLM). Then, a state-of-the-art search engine is used to retrieve candidate passages from a large collection for each decomposed question. In the final step, we use the LLM in a few-shot setting to aggregate the contents of the passages into the final answer. The system is evaluated on three datasets: IIRC, Qasper, and StrategyQA. Results suggest that current retrievers are the main bottleneck and that readers are already performing at the human level as long as relevant passages are provided. The system is also shown to be more effective when the model is induced to give explanations before answering a question. Code is available at url{https://github.com/neuralmind-ai/visconde}.","PeriodicalId":126309,"journal":{"name":"European Conference on Information Retrieval","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133975199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Exploring Fake News Detection with Heterogeneous Social Media Context Graphs 利用异质社交媒体情境图探索假新闻检测
Pub Date : 2022-12-13 DOI: 10.48550/arXiv.2212.06560
Gregor Donabauer, Udo Kruschwitz
Fake news detection has become a research area that goes way beyond a purely academic interest as it has direct implications on our society as a whole. Recent advances have primarily focused on textbased approaches. However, it has become clear that to be effective one needs to incorporate additional, contextual information such as spreading behaviour of news articles and user interaction patterns on social media. We propose to construct heterogeneous social context graphs around news articles and reformulate the problem as a graph classification task. Exploring the incorporation of different types of information (to get an idea as to what level of social context is most effective) and using different graph neural network architectures indicates that this approach is highly effective with robust results on a common benchmark dataset.
假新闻检测已经成为一个超越纯粹学术兴趣的研究领域,因为它对我们整个社会都有直接的影响。最近的进展主要集中在基于文本的方法上。然而,很明显,要想有效,就需要纳入额外的上下文信息,比如新闻文章的传播行为和社交媒体上的用户互动模式。我们建议围绕新闻文章构建异构社会语境图,并将该问题重新表述为一个图分类任务。探索不同类型信息的合并(以了解哪种社会背景级别最有效)和使用不同的图神经网络体系结构表明,这种方法在通用基准数据集上非常有效,并且具有鲁棒性结果。
{"title":"Exploring Fake News Detection with Heterogeneous Social Media Context Graphs","authors":"Gregor Donabauer, Udo Kruschwitz","doi":"10.48550/arXiv.2212.06560","DOIUrl":"https://doi.org/10.48550/arXiv.2212.06560","url":null,"abstract":"Fake news detection has become a research area that goes way beyond a purely academic interest as it has direct implications on our society as a whole. Recent advances have primarily focused on textbased approaches. However, it has become clear that to be effective one needs to incorporate additional, contextual information such as spreading behaviour of news articles and user interaction patterns on social media. We propose to construct heterogeneous social context graphs around news articles and reformulate the problem as a graph classification task. Exploring the incorporation of different types of information (to get an idea as to what level of social context is most effective) and using different graph neural network architectures indicates that this approach is highly effective with robust results on a common benchmark dataset.","PeriodicalId":126309,"journal":{"name":"European Conference on Information Retrieval","volume":"125 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124203336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Multivariate Powered Dirichlet Hawkes Process 多元幂次Dirichlet Hawkes过程
Pub Date : 2022-12-12 DOI: 10.48550/arXiv.2212.05995
Gael Poux-Medard, Julien Velcin, Sabine Loudcher
The publication time of a document carries a relevant information about its semantic content. The Dirichlet-Hawkes process has been proposed to jointly model textual information and publication dynamics. This approach has been used with success in several recent works, and extended to tackle specific challenging problems --typically for short texts or entangled publication dynamics. However, the prior in its current form does not allow for complex publication dynamics. In particular, inferred topics are independent from each other --a publication about finance is assumed to have no influence on publications about politics, for instance. In this work, we develop the Multivariate Powered Dirichlet-Hawkes Process (MPDHP), that alleviates this assumption. Publications about various topics can now influence each other. We detail and overcome the technical challenges that arise from considering interacting topics. We conduct a systematic evaluation of MPDHP on a range of synthetic datasets to define its application domain and limitations. Finally, we develop a use case of the MPDHP on Reddit data. At the end of this article, the interested reader will know how and when to use MPDHP, and when not to.
文档的发布时间包含了其语义内容的相关信息。Dirichlet-Hawkes过程被提出来联合建模文本信息和出版动态。这种方法在最近的几项工作中得到了成功的应用,并扩展到解决特定的具有挑战性的问题——通常是短文本或纠缠的出版物动态。但是,当前形式的先验不允许复杂的发布动态。特别是,推断的主题是相互独立的——例如,假设有关金融的出版物对有关政治的出版物没有影响。在这项工作中,我们开发了多元幂次Dirichlet-Hawkes过程(MPDHP),减轻了这一假设。各种主题的出版物现在可以相互影响。我们详细介绍并克服了由于考虑交互主题而产生的技术挑战。我们在一系列合成数据集上对MPDHP进行了系统评估,以定义其应用领域和局限性。最后,我们在Reddit数据上开发了MPDHP的一个用例。在本文的最后,感兴趣的读者将知道如何以及何时使用MPDHP,何时不使用。
{"title":"Multivariate Powered Dirichlet Hawkes Process","authors":"Gael Poux-Medard, Julien Velcin, Sabine Loudcher","doi":"10.48550/arXiv.2212.05995","DOIUrl":"https://doi.org/10.48550/arXiv.2212.05995","url":null,"abstract":"The publication time of a document carries a relevant information about its semantic content. The Dirichlet-Hawkes process has been proposed to jointly model textual information and publication dynamics. This approach has been used with success in several recent works, and extended to tackle specific challenging problems --typically for short texts or entangled publication dynamics. However, the prior in its current form does not allow for complex publication dynamics. In particular, inferred topics are independent from each other --a publication about finance is assumed to have no influence on publications about politics, for instance. In this work, we develop the Multivariate Powered Dirichlet-Hawkes Process (MPDHP), that alleviates this assumption. Publications about various topics can now influence each other. We detail and overcome the technical challenges that arise from considering interacting topics. We conduct a systematic evaluation of MPDHP on a range of synthetic datasets to define its application domain and limitations. Finally, we develop a use case of the MPDHP on Reddit data. At the end of this article, the interested reader will know how and when to use MPDHP, and when not to.","PeriodicalId":126309,"journal":{"name":"European Conference on Information Retrieval","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127874072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Dirichlet-Survival Process: Scalable Inference of Topic-Dependent Diffusion Networks dirichlet -生存过程:主题相关扩散网络的可扩展推理
Pub Date : 2022-12-12 DOI: 10.48550/arXiv.2212.05996
Gael Poux-Medard, Julien Velcin, Sabine Loudcher
Information spread on networks can be efficiently modeled by considering three features: documents' content, time of publication relative to other publications, and position of the spreader in the network. Most previous works model up to two of those jointly, or rely on heavily parametric approaches. Building on recent Dirichlet-Point processes literature, we introduce the Houston (Hidden Online User-Topic Network) model, that jointly considers all those features in a non-parametric unsupervised framework. It infers dynamic topic-dependent underlying diffusion networks in a continuous-time setting along with said topics. It is unsupervised; it considers an unlabeled stream of triplets shaped as textit{(time of publication, information's content, spreading entity)} as input data. Online inference is conducted using a sequential Monte-Carlo algorithm that scales linearly with the size of the dataset. Our approach yields consequent improvements over existing baselines on both cluster recovery and subnetworks inference tasks.
通过考虑文件的内容、相对于其他出版物的发布时间以及传播者在网络中的位置这三个特征,可以有效地对网络上的信息传播进行建模。大多数先前的工作都是联合建模其中的两个,或者依赖于高度参数化的方法。在最近的Dirichlet-Point过程文献的基础上,我们引入了休斯顿(隐藏在线用户主题网络)模型,该模型在非参数无监督框架中共同考虑了所有这些特征。它与所述主题一起在连续时间设置中推断动态主题相关的潜在扩散网络。它是无人监督的;它考虑一个未标记的三元流,形状为textit{(发布时间、信息内容、传播实体)}作为输入数据。在线推理使用顺序蒙特卡罗算法进行,该算法与数据集的大小线性扩展。我们的方法在集群恢复和子网推理任务上都优于现有的基线。
{"title":"Dirichlet-Survival Process: Scalable Inference of Topic-Dependent Diffusion Networks","authors":"Gael Poux-Medard, Julien Velcin, Sabine Loudcher","doi":"10.48550/arXiv.2212.05996","DOIUrl":"https://doi.org/10.48550/arXiv.2212.05996","url":null,"abstract":"Information spread on networks can be efficiently modeled by considering three features: documents' content, time of publication relative to other publications, and position of the spreader in the network. Most previous works model up to two of those jointly, or rely on heavily parametric approaches. Building on recent Dirichlet-Point processes literature, we introduce the Houston (Hidden Online User-Topic Network) model, that jointly considers all those features in a non-parametric unsupervised framework. It infers dynamic topic-dependent underlying diffusion networks in a continuous-time setting along with said topics. It is unsupervised; it considers an unlabeled stream of triplets shaped as textit{(time of publication, information's content, spreading entity)} as input data. Online inference is conducted using a sequential Monte-Carlo algorithm that scales linearly with the size of the dataset. Our approach yields consequent improvements over existing baselines on both cluster recovery and subnetworks inference tasks.","PeriodicalId":126309,"journal":{"name":"European Conference on Information Retrieval","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129148835","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Transformer-based Framework for POI-level Social Post Geolocation 基于变换的poi级社会邮政地理定位框架
Pub Date : 2022-10-26 DOI: 10.48550/arXiv.2211.01336
Menglin Li, Kwan Hui Lim, Teng Guo, Junhua Liu
POI-level geo-information of social posts is critical to many location-based applications and services. However, the multi-modality, complexity and diverse nature of social media data and their platforms limit the performance of inferring such fine-grained locations and their subsequent applications. To address this issue, we present a transformer-based general framework, which builds upon pre-trained language models and considers non-textual data, for social post geolocation at the POI level. To this end, inputs are categorized to handle different social data, and an optimal combination strategy is provided for feature representations. Moreover, a uniform representation of hierarchy is proposed to learn temporal information, and a concatenated version of encodings is employed to capture feature-wise positions better. Experimental results on various social datasets demonstrate that three variants of our proposed framework outperform multiple state-of-art baselines by a large margin in terms of accuracy and distance error metrics.
社交帖子的poi级地理信息对于许多基于位置的应用程序和服务至关重要。然而,社交媒体数据及其平台的多模态、复杂性和多样性限制了推断这种细粒度位置及其后续应用的性能。为了解决这个问题,我们提出了一个基于转换器的通用框架,该框架建立在预训练的语言模型之上,并考虑了非文本数据,用于POI级别的社会岗位地理定位。为此,对输入进行分类以处理不同的社会数据,并为特征表示提供最优组合策略。此外,提出了一种统一的层次表示来学习时间信息,并采用了一种连接版本的编码来更好地捕获特征位置。在各种社会数据集上的实验结果表明,我们提出的框架的三个变体在精度和距离误差指标方面优于多个最先进的基线。
{"title":"A Transformer-based Framework for POI-level Social Post Geolocation","authors":"Menglin Li, Kwan Hui Lim, Teng Guo, Junhua Liu","doi":"10.48550/arXiv.2211.01336","DOIUrl":"https://doi.org/10.48550/arXiv.2211.01336","url":null,"abstract":"POI-level geo-information of social posts is critical to many location-based applications and services. However, the multi-modality, complexity and diverse nature of social media data and their platforms limit the performance of inferring such fine-grained locations and their subsequent applications. To address this issue, we present a transformer-based general framework, which builds upon pre-trained language models and considers non-textual data, for social post geolocation at the POI level. To this end, inputs are categorized to handle different social data, and an optimal combination strategy is provided for feature representations. Moreover, a uniform representation of hierarchy is proposed to learn temporal information, and a concatenated version of encodings is employed to capture feature-wise positions better. Experimental results on various social datasets demonstrate that three variants of our proposed framework outperform multiple state-of-art baselines by a large margin in terms of accuracy and distance error metrics.","PeriodicalId":126309,"journal":{"name":"European Conference on Information Retrieval","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134502983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Clustering Without Knowing How To: Application and Evaluation 不知道如何聚类:应用和评估
Pub Date : 2022-09-21 DOI: 10.1007/978-3-031-28241-6_24
Daniil Likhobaba, Daniil A. Fedulov, Dmitry Ustalov
{"title":"Clustering Without Knowing How To: Application and Evaluation","authors":"Daniil Likhobaba, Daniil A. Fedulov, Dmitry Ustalov","doi":"10.1007/978-3-031-28241-6_24","DOIUrl":"https://doi.org/10.1007/978-3-031-28241-6_24","url":null,"abstract":"","PeriodicalId":126309,"journal":{"name":"European Conference on Information Retrieval","volume":"262 ","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133941862","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neural Approaches to Multilingual Information Retrieval 多语言信息检索的神经方法
Pub Date : 2022-09-03 DOI: 10.1007/978-3-031-28244-7_33
Dawn J Lawrie, Eugene Yang, Douglas W. Oard, J. Mayfield
{"title":"Neural Approaches to Multilingual Information Retrieval","authors":"Dawn J Lawrie, Eugene Yang, Douglas W. Oard, J. Mayfield","doi":"10.1007/978-3-031-28244-7_33","DOIUrl":"https://doi.org/10.1007/978-3-031-28244-7_33","url":null,"abstract":"","PeriodicalId":126309,"journal":{"name":"European Conference on Information Retrieval","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132617508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
European Conference on Information Retrieval
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1