首页 > 最新文献

2018 IEEE 14th International Conference on e-Science (e-Science)最新文献

英文 中文
Message from the eScience 2018 Program Committee Chairs for the Focused Session on Exascale Computing for High-Energy Physics 来自eScience 2018项目委员会主席关于高能物理百亿亿次计算重点会议的信息
Pub Date : 2018-10-01 DOI: 10.1109/eScience.2018.00079
J. Templon, Y. Dzigan
n/a
{"title":"Message from the eScience 2018 Program Committee Chairs for the Focused Session on Exascale Computing for High-Energy Physics","authors":"J. Templon, Y. Dzigan","doi":"10.1109/eScience.2018.00079","DOIUrl":"https://doi.org/10.1109/eScience.2018.00079","url":null,"abstract":"n/a","PeriodicalId":6476,"journal":{"name":"2018 IEEE 14th International Conference on e-Science (e-Science)","volume":"93 1","pages":"331-331"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80494631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Increasing Parallelism in Climate Models Via Additional Component Concurrency 通过附加组件并发增加气候模式的并行性
Pub Date : 2018-10-01 DOI: 10.1109/eScience.2018.00044
Jorg Behrens, J. Biercamp, H. Bockelmann, P. Neumann
n/a
{"title":"Increasing Parallelism in Climate Models Via Additional Component Concurrency","authors":"Jorg Behrens, J. Biercamp, H. Bockelmann, P. Neumann","doi":"10.1109/eScience.2018.00044","DOIUrl":"https://doi.org/10.1109/eScience.2018.00044","url":null,"abstract":"n/a","PeriodicalId":6476,"journal":{"name":"2018 IEEE 14th International Conference on e-Science (e-Science)","volume":"21 1","pages":"271-271"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79193025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Search for Computational Workflow Synergies in Reproducible Research Data Analyses in Particle Physics and Life Sciences 粒子物理和生命科学中可重复研究数据分析中计算工作流协同效应的搜索
Pub Date : 2018-10-01 DOI: 10.1109/eScience.2018.00123
T. Simko, K. Cranmer, M. Crusoe, L. Heinrich, A. Khodak, Dinos Kousidis, D. Rodríguez
We describe the REANA reusable and reproducible research data analysis platform that originated in the domain of particle physics. We integrated support for running Common Workflow Language (CWL) workflows that originated in the domain of life sciences. This integration allowed us to study the applicability of CWL to particle physics analyses and look for synergies in computational practices in the two communities.
我们描述了起源于粒子物理领域的REANA可重复使用和可复制的研究数据分析平台。我们集成了对运行源自生命科学领域的公共工作流语言(CWL)工作流的支持。这种整合使我们能够研究CWL在粒子物理分析中的适用性,并在两个社区的计算实践中寻找协同作用。
{"title":"Search for Computational Workflow Synergies in Reproducible Research Data Analyses in Particle Physics and Life Sciences","authors":"T. Simko, K. Cranmer, M. Crusoe, L. Heinrich, A. Khodak, Dinos Kousidis, D. Rodríguez","doi":"10.1109/eScience.2018.00123","DOIUrl":"https://doi.org/10.1109/eScience.2018.00123","url":null,"abstract":"We describe the REANA reusable and reproducible research data analysis platform that originated in the domain of particle physics. We integrated support for running Common Workflow Language (CWL) workflows that originated in the domain of life sciences. This integration allowed us to study the applicability of CWL to particle physics analyses and look for synergies in computational practices in the two communities.","PeriodicalId":6476,"journal":{"name":"2018 IEEE 14th International Conference on e-Science (e-Science)","volume":"76 1","pages":"403-404"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82533057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Automated Parallel Calculation of Collaborative Statistical Models in RooFit 协同统计模型的自动并行计算
Pub Date : 2018-10-01 DOI: 10.1109/eScience.2018.00089
E. G. P. Bos, I. Pelupessy, V. Croft, W. Verkerke, C. Burgard
RooFit [4], [6] is the statistical modeling and fitting package used in many big particle physics experiments to extract physical parameters from reduced particle collision data, e.g. the Higgs boson experiments at the LHC [1], [2]. RooFit aims to separate particle physics model building and fitting (the users’ goals) from their technical implementation and optimization in the back-end. In this paper, we outline our efforts to further optimize the back-end by automatically running major parts of user models in parallel on multi-core machines. A major challenge is that RooFit allows users to define many different types of models, with different types of computational bottlenecks. Our automatic parallelization framework must then be flexible, while still reducing run-time by at least an order of magnitude, preferably more. We have performed extensive benchmarks and identified at least three bottlenecks that will benefit from parallelization. To tackle these and possible future bottlenecks, we designed a parallelization layer that allows us to parallelize existing classes with minimal effort, but with high performance and retaining as much of the existing class’s interface as possible. The high-level parallelization model is a task-stealing approach. The implementation is currently based on a multi-process approach using a bi-directional memory mapped pipe for communication, which is both easy to use and highly performant. Preliminary results show speed-ups of factor 2 to 20, depending on the exact model and parallelization strategy.
RooFit[4],[6]是许多大型粒子物理实验中使用的统计建模和拟合包,用于从粒子碰撞数据中提取物理参数,例如大型强子对撞机的希格斯玻色子实验[1],[2]。RooFit旨在将粒子物理模型的构建和拟合(用户的目标)与后端技术实现和优化分开。在本文中,我们概述了通过在多核机器上自动并行运行用户模型的主要部分来进一步优化后端所做的努力。一个主要的挑战是,RooFit允许用户定义许多不同类型的模型,这些模型具有不同类型的计算瓶颈。然后,我们的自动并行化框架必须是灵活的,同时仍然至少减少一个数量级的运行时间,最好更多。我们已经执行了大量的基准测试,并确定了至少三个将从并行化中受益的瓶颈。为了解决这些和未来可能出现的瓶颈,我们设计了一个并行化层,它允许我们以最小的工作量并行化现有的类,但具有高性能并尽可能多地保留现有类的接口。高级并行化模型是一种任务窃取方法。该实现目前基于多进程方法,使用双向内存映射管道进行通信,既易于使用又高性能。初步结果显示,根据具体的模型和并行化策略,速度可以提高2到20倍。
{"title":"Automated Parallel Calculation of Collaborative Statistical Models in RooFit","authors":"E. G. P. Bos, I. Pelupessy, V. Croft, W. Verkerke, C. Burgard","doi":"10.1109/eScience.2018.00089","DOIUrl":"https://doi.org/10.1109/eScience.2018.00089","url":null,"abstract":"RooFit [4], [6] is the statistical modeling and fitting package used in many big particle physics experiments to extract physical parameters from reduced particle collision data, e.g. the Higgs boson experiments at the LHC [1], [2]. RooFit aims to separate particle physics model building and fitting (the users’ goals) from their technical implementation and optimization in the back-end. In this paper, we outline our efforts to further optimize the back-end by automatically running major parts of user models in parallel on multi-core machines. A major challenge is that RooFit allows users to define many different types of models, with different types of computational bottlenecks. Our automatic parallelization framework must then be flexible, while still reducing run-time by at least an order of magnitude, preferably more. We have performed extensive benchmarks and identified at least three bottlenecks that will benefit from parallelization. To tackle these and possible future bottlenecks, we designed a parallelization layer that allows us to parallelize existing classes with minimal effort, but with high performance and retaining as much of the existing class’s interface as possible. The high-level parallelization model is a task-stealing approach. The implementation is currently based on a multi-process approach using a bi-directional memory mapped pipe for communication, which is both easy to use and highly performant. Preliminary results show speed-ups of factor 2 to 20, depending on the exact model and parallelization strategy.","PeriodicalId":6476,"journal":{"name":"2018 IEEE 14th International Conference on e-Science (e-Science)","volume":"1 1","pages":"345-346"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82578311","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
The Results and Challenges of Using Administrative Health Data Within a Natural Experimental Evaluation of the Abolition of Prescription Fees in Scotland 在苏格兰取消处方费用的自然实验评估中使用行政卫生数据的结果和挑战
Pub Date : 2018-10-01 DOI: 10.1109/eScience.2018.00128
A. Williams, W. Henley, J. Frank
n/a
{"title":"The Results and Challenges of Using Administrative Health Data Within a Natural Experimental Evaluation of the Abolition of Prescription Fees in Scotland","authors":"A. Williams, W. Henley, J. Frank","doi":"10.1109/eScience.2018.00128","DOIUrl":"https://doi.org/10.1109/eScience.2018.00128","url":null,"abstract":"n/a","PeriodicalId":6476,"journal":{"name":"2018 IEEE 14th International Conference on e-Science (e-Science)","volume":"57 1","pages":"412-412"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76929937","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Linking Text and Knowledge Using the INCEpTION Annotation Platform 使用INCEpTION注释平台链接文本和知识
Pub Date : 2018-10-01 DOI: 10.1109/ESCIENCE.2018.00077
Richard Eckart de Castilho, Jan-Christoph Klie, Naveen Kumar, Beto Boullosa, Iryna Gurevych
Abstract-In the Digital Humanities (DH), linking text collections to general or domain-specific knowledge bases (KBs) or authority files is important to enable a contextualised analysis. Automatic named entity recognition and entity linking tools require training data or domain-specific methods. Interactive annotation tools do often not support the tasks of entity linking, fact-linking, cross-document reference resolution, etc. We aim to address this gap with the INCEpTION annotation platform, which not only provides these capabilities in the context of a generic annotation tool, but also combines them with machine learning methods to improve annotation efficiency.
摘要:在数字人文学科(DH)中,将文本集合链接到一般或特定领域的知识库(KBs)或权威文件对于实现上下文化分析非常重要。自动命名实体识别和实体链接工具需要训练数据或特定于领域的方法。交互式注释工具通常不支持实体链接、事实链接、跨文档引用解析等任务。我们的目标是通过INCEpTION注释平台来解决这一问题,该平台不仅在通用注释工具的背景下提供这些功能,而且还将它们与机器学习方法相结合,以提高注释效率。
{"title":"Linking Text and Knowledge Using the INCEpTION Annotation Platform","authors":"Richard Eckart de Castilho, Jan-Christoph Klie, Naveen Kumar, Beto Boullosa, Iryna Gurevych","doi":"10.1109/ESCIENCE.2018.00077","DOIUrl":"https://doi.org/10.1109/ESCIENCE.2018.00077","url":null,"abstract":"Abstract-In the Digital Humanities (DH), linking text collections to general or domain-specific knowledge bases (KBs) or authority files is important to enable a contextualised analysis. Automatic named entity recognition and entity linking tools require training data or domain-specific methods. Interactive annotation tools do often not support the tasks of entity linking, fact-linking, cross-document reference resolution, etc. We aim to address this gap with the INCEpTION annotation platform, which not only provides these capabilities in the context of a generic annotation tool, but also combines them with machine learning methods to improve annotation efficiency.","PeriodicalId":6476,"journal":{"name":"2018 IEEE 14th International Conference on e-Science (e-Science)","volume":"65 1","pages":"327-328"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75974362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Cookery: A Framework for Creating Data Processing Pipeline Using Online Services 烹饪:一个使用在线服务创建数据处理管道的框架
Pub Date : 2018-10-01 DOI: 10.1109/eScience.2018.00102
Mikolaj Baranowski, A. Belloum, R. Cushing, O. Valkering
With the increasing amount of data the importance of data analysis has grown. A large amount of this data has shifted to cloud-based storage. The cloud offers storage and computation power. The Cookery framework is a tool developed to build application in the cloud for scientists without a complete understanding of programming. In this paper with present the cookery systems and how it can be used to authenticate and use standard online 3rd party services to easily create data analytics pipeline. Cookery framework is not limited to work with standard web services, it can also integrate and work with the emerging AWS Lambda. The combination of AWS Lambda and Cookery, which makes it possible for people, who do not have any program experience, to create data processing pipeline using cloud services in short time.
随着数据量的增加,数据分析的重要性也越来越大。大量此类数据已转移到基于云的存储中。云提供存储和计算能力。Cookery框架是一种为不完全了解编程的科学家开发的云应用程序的工具。本文介绍了cookery系统,以及如何使用它来验证和使用标准的在线第三方服务来轻松创建数据分析管道。Cookery框架并不局限于使用标准的web服务,它还可以集成和使用新兴的AWS Lambda。AWS Lambda和Cookery的结合,使得没有任何编程经验的人也可以在短时间内使用云服务创建数据处理管道。
{"title":"Cookery: A Framework for Creating Data Processing Pipeline Using Online Services","authors":"Mikolaj Baranowski, A. Belloum, R. Cushing, O. Valkering","doi":"10.1109/eScience.2018.00102","DOIUrl":"https://doi.org/10.1109/eScience.2018.00102","url":null,"abstract":"With the increasing amount of data the importance of data analysis has grown. A large amount of this data has shifted to cloud-based storage. The cloud offers storage and computation power. The Cookery framework is a tool developed to build application in the cloud for scientists without a complete understanding of programming. In this paper with present the cookery systems and how it can be used to authenticate and use standard online 3rd party services to easily create data analytics pipeline. Cookery framework is not limited to work with standard web services, it can also integrate and work with the emerging AWS Lambda. The combination of AWS Lambda and Cookery, which makes it possible for people, who do not have any program experience, to create data processing pipeline using cloud services in short time.","PeriodicalId":6476,"journal":{"name":"2018 IEEE 14th International Conference on e-Science (e-Science)","volume":"38 1","pages":"368-369"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80916449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Democratizing Ancient Mesopotamian Research through Digital Scholarship 通过数字学术使古代美索不达米亚研究民主化
Pub Date : 2018-10-01 DOI: 10.1109/eScience.2018.00074
Raquel Alegre, Anastasis Georgoulas, S. Grieve, E. Robson
Since the 19th century, historians and archaeologists have compiled transliterations and translations of surviving cuneiform texts from the Middle East area, documenting the ancient history of the region, c. 3000 BC–75 AD. The Open Richly Annotated Cuneiform Corpus (Oracc)1 is an international collaborative effort to gather and digitise a complete collection of cuneiform texts and their translations, with the goal of making them available to researchers and students worldwide. Oracc was developed ten years ago around the core value of ensuring accessibility to a broad audience, rather than a select group of experts. This principle presented new technological challenges, but has equally offered important benefits. Initial transliteration of cuneiform tablets into the ASCII Transliteration Format (ATF) was performed using an Emacs plugin, the use of which was challenging for novice and experienced users alike. This precipitated the development of Nammu [1], a dedicated editor for files written in ATF, to provide a consistent environment for users to contribute to Oracc projects. This is an important step in the democratization of this research as it lowers the technological expertise required to join the platform, and reduces the amount of time needed to train new users, which was previously a large drain on Principal Investigators’ time and resources. Nammu in turn takes advantage of pyORACC [2], a bespoke library developed for parsing ATF files and a key enabler of automation in the project. Separately to the editing considerations, the Oracc website hosts the body of information editions and translations that researchers from different groups have accumulated during their work. An important aspect of this is the search capability it offers, allowing a user to retrieve information about a subject or term of their choice. A new version of this functionality is being developed, using the ElasticSearch platform to index and efficiently search large bodies of text. Users can choose to query the compiled glossaries, looking for words with a particular meaning, or for the meaning and appearances of a transliterated cuneiform term. Alternatively, they will be able to search through the information pages for a topic of their choice, effectively using the website as a domain-specific search engine. This dual functionality has been chosen so as to make the search of interest to both domain experts and the general public. Early versions of Nammu focused on the transliteration and translation of cuneiform into English and other European languages. Meanwhile, decades of war and political instability across the Middle East have prevented researchers from Iraq, Syria and neighbouring countries from contributing to the ancient history Programming work on Oracc is funded by UCL’s School of Social and Historical Sciences, and through the Nahrein Network’s grant from the UK Arts and Humanities Council’s Global Challenges Research Fund. 1http://oracc.org of their re
在双方之间保持这种开放的对话和理解是项目成功和可持续发展的关键。为了保持开放的精神,核心决策之一是尽可能多地使用现有标准(如XML和JSON),发布开发工具的源代码,并为感兴趣的各方提供详细的文档。这些实践使得来自其他组的用户不仅采用了该软件,而且还为其开发做出了贡献。
{"title":"Democratizing Ancient Mesopotamian Research through Digital Scholarship","authors":"Raquel Alegre, Anastasis Georgoulas, S. Grieve, E. Robson","doi":"10.1109/eScience.2018.00074","DOIUrl":"https://doi.org/10.1109/eScience.2018.00074","url":null,"abstract":"Since the 19th century, historians and archaeologists have compiled transliterations and translations of surviving cuneiform texts from the Middle East area, documenting the ancient history of the region, c. 3000 BC–75 AD. The Open Richly Annotated Cuneiform Corpus (Oracc)1 is an international collaborative effort to gather and digitise a complete collection of cuneiform texts and their translations, with the goal of making them available to researchers and students worldwide. Oracc was developed ten years ago around the core value of ensuring accessibility to a broad audience, rather than a select group of experts. This principle presented new technological challenges, but has equally offered important benefits. Initial transliteration of cuneiform tablets into the ASCII Transliteration Format (ATF) was performed using an Emacs plugin, the use of which was challenging for novice and experienced users alike. This precipitated the development of Nammu [1], a dedicated editor for files written in ATF, to provide a consistent environment for users to contribute to Oracc projects. This is an important step in the democratization of this research as it lowers the technological expertise required to join the platform, and reduces the amount of time needed to train new users, which was previously a large drain on Principal Investigators’ time and resources. Nammu in turn takes advantage of pyORACC [2], a bespoke library developed for parsing ATF files and a key enabler of automation in the project. Separately to the editing considerations, the Oracc website hosts the body of information editions and translations that researchers from different groups have accumulated during their work. An important aspect of this is the search capability it offers, allowing a user to retrieve information about a subject or term of their choice. A new version of this functionality is being developed, using the ElasticSearch platform to index and efficiently search large bodies of text. Users can choose to query the compiled glossaries, looking for words with a particular meaning, or for the meaning and appearances of a transliterated cuneiform term. Alternatively, they will be able to search through the information pages for a topic of their choice, effectively using the website as a domain-specific search engine. This dual functionality has been chosen so as to make the search of interest to both domain experts and the general public. Early versions of Nammu focused on the transliteration and translation of cuneiform into English and other European languages. Meanwhile, decades of war and political instability across the Middle East have prevented researchers from Iraq, Syria and neighbouring countries from contributing to the ancient history Programming work on Oracc is funded by UCL’s School of Social and Historical Sciences, and through the Nahrein Network’s grant from the UK Arts and Humanities Council’s Global Challenges Research Fund. 1http://oracc.org of their re","PeriodicalId":6476,"journal":{"name":"2018 IEEE 14th International Conference on e-Science (e-Science)","volume":"37 1","pages":"322-322"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80937159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
eWaterCycle II
Pub Date : 2018-10-01 DOI: 10.1109/eScience.2018.00108
R. Hut, N. Drost, W. V. Hage, N. Giesen
From a hydrological point of view, every field, every street, every part of the world, is different. We understand quite well how water moves through plants and soils at small scales but the medium is never the same from one spot to the next. This is the curse of locality. It is difficult to capture such processes with a single global model. In the last two decades, hydrology has slowly moved into two related fields: global hydrology and catchment hydrology. In global hydrology, making use of new computational resources, scientists use uniform global models at ever increasing spatial and temporal resolutions, forced with satellite data or climate model output to make claims on the global state of the hydrological cycle [1], [2]. Parallel to this development, researchers in catchment hydrology, have focussed on deriving, for each catchment that is studied, the best hydrological models for that specific catchment. This is nicely summarized in the overview paper of the last hydrological decade [3]. While global hydrologists realize that hydrological processes are locally very different and human influence even more so [4], incorporating the body of local hydrological knowledge is not easy. Catchment hydrologists realize the importance of their work to the global watercycle but often lack the (computational) resources and tools to upscale from their catchment to the global picture. The eWaterCycle II project will build and maintain an e- Infrastructure that allows for quick and safe inclusion of submodels and model concepts into global hydrological models, leading to a better understanding of the Hydrological cycle. The foreseen e-infrastructure will have a number of unique advantages, including an ability for knowledge gap discovery, machine-assisted model curation, and easily changeable model parts. In this work we will present the how we will achieve the goals of the recently started eWaterCycle II project over its three year runtime. We will show a demo of a first prototype environment where scientist can run, compare and alter different hydrological models that focus on the same region and use the same input data sources. This will work even if the underlying hydrological models are written in different programming languages without exposing the hydrologists doing the comparison to these technical intricacies. Although the eWaterCycle II project focusses on the hydrological setting, the underlying framework will be suitable outside of hydrology, wherever a collaborative environment is required. eScience aspects such as large scale data assimilation (DA) techniques, generic multi-model multi-scale environments, FAIR data as well as FAIR software, will all benefit from research done in this project.
从水文的角度来看,每一块田地、每一条街道、世界的每一部分都是不同的。我们很清楚水是如何在小尺度上通过植物和土壤流动的,但是从一个地方到另一个地方的介质是不一样的。这就是地域的诅咒。用单一的全局模型很难捕捉到这样的过程。在过去的二十年里,水文学慢慢地进入了两个相关的领域:全球水文学和流域水文学。在全球水文学中,科学家利用新的计算资源,使用统一的全球模型,在不断增加的空间和时间分辨率下,利用卫星数据或气候模型输出对全球水文循环状态提出要求[1],[2]。与此同时,流域水文学的研究人员专注于为所研究的每个流域推导出适合该特定流域的最佳水文模型。上一个水文十年的综述论文[3]很好地总结了这一点。虽然全球水文学家意识到,各地的水文过程差异很大,人类的影响更是如此[4],但将当地的水文知识纳入其中并不容易。流域水文学家意识到他们的工作对全球水循环的重要性,但往往缺乏(计算)资源和工具,从他们的流域上升到全球图景。eWaterCycle II项目将建立和维护一个电子基础设施,允许将子模型和模型概念快速安全地纳入全球水文模型,从而更好地理解水文循环。可预见的电子基础设施将具有许多独特的优势,包括发现知识差距的能力、机器辅助模型管理和易于更换的模型部件。在这项工作中,我们将介绍如何在三年的运行时间内实现最近启动的eWaterCycle II项目的目标。我们将展示第一个原型环境的演示,科学家可以在其中运行,比较和更改不同的水文模型,这些模型关注同一地区并使用相同的输入数据源。即使底层水文模型是用不同的编程语言编写的,这也可以工作,而不会让水文学家进行这些技术复杂性的比较。尽管eWaterCycle II项目侧重于水文环境,但其基础框架将适用于水文学之外的任何需要协作环境的地方。科学方面,如大规模数据同化(DA)技术,通用多模型多尺度环境,FAIR数据以及FAIR软件,都将受益于本项目的研究。
{"title":"eWaterCycle II","authors":"R. Hut, N. Drost, W. V. Hage, N. Giesen","doi":"10.1109/eScience.2018.00108","DOIUrl":"https://doi.org/10.1109/eScience.2018.00108","url":null,"abstract":"From a hydrological point of view, every field, every street, every part of the world, is different. We understand quite well how water moves through plants and soils at small scales but the medium is never the same from one spot to the next. This is the curse of locality. It is difficult to capture such processes with a single global model. In the last two decades, hydrology has slowly moved into two related fields: global hydrology and catchment hydrology. In global hydrology, making use of new computational resources, scientists use uniform global models at ever increasing spatial and temporal resolutions, forced with satellite data or climate model output to make claims on the global state of the hydrological cycle [1], [2]. Parallel to this development, researchers in catchment hydrology, have focussed on deriving, for each catchment that is studied, the best hydrological models for that specific catchment. This is nicely summarized in the overview paper of the last hydrological decade [3]. While global hydrologists realize that hydrological processes are locally very different and human influence even more so [4], incorporating the body of local hydrological knowledge is not easy. Catchment hydrologists realize the importance of their work to the global watercycle but often lack the (computational) resources and tools to upscale from their catchment to the global picture. The eWaterCycle II project will build and maintain an e- Infrastructure that allows for quick and safe inclusion of submodels and model concepts into global hydrological models, leading to a better understanding of the Hydrological cycle. The foreseen e-infrastructure will have a number of unique advantages, including an ability for knowledge gap discovery, machine-assisted model curation, and easily changeable model parts. In this work we will present the how we will achieve the goals of the recently started eWaterCycle II project over its three year runtime. We will show a demo of a first prototype environment where scientist can run, compare and alter different hydrological models that focus on the same region and use the same input data sources. This will work even if the underlying hydrological models are written in different programming languages without exposing the hydrologists doing the comparison to these technical intricacies. Although the eWaterCycle II project focusses on the hydrological setting, the underlying framework will be suitable outside of hydrology, wherever a collaborative environment is required. eScience aspects such as large scale data assimilation (DA) techniques, generic multi-model multi-scale environments, FAIR data as well as FAIR software, will all benefit from research done in this project.","PeriodicalId":6476,"journal":{"name":"2018 IEEE 14th International Conference on e-Science (e-Science)","volume":"55 1","pages":"379-379"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76244675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Fine-Grained Processing Towards HL-LHC Computing in ATLAS ATLAS中HL-LHC计算的细粒度处理
Pub Date : 2018-10-01 DOI: 10.1109/eScience.2018.00083
D. Benjamin, P. Calafiura, T. Childers, K. De, A. Girolamo, E. Fullana, W. Guan, T. Maeno, Nicolò Magini, P. Nilsson, D. Oleynik, Shaojun Sun, V. Tsulaia, P. Gemmeren, T. Wenaus, W. Yang
During LHC's Run-2 ATLAS has been developing and evaluating new fine-grained approaches to workflows and dataflows able to better utilize computing resources in terms of storage, processing and networks. The compute-limited physics of ATLAS has driven the collaboration to aggressively harvest opportunistic cycles from what are often transiently available resources, including HPCs, clouds, volunteer computing, and grid resources in transitional states. Fine-grained processing (with typically a few minutes' granularity, corresponding to one event for the present ATLAS full simulation) enables agile workflows with a light footprint on the resource such that cycles can be more fully and efficiently utilized than with conventional workflows processing O(GB) files per job.
在LHC Run-2期间,ATLAS一直在开发和评估新的细粒度工作流和数据流方法,这些方法能够更好地利用存储、处理和网络方面的计算资源。ATLAS的计算有限的物理特性促使协作积极地从通常是暂时可用的资源中获取机会周期,包括hpc、云、志愿计算和过渡状态的网格资源。细粒度处理(通常为几分钟的粒度,对应于当前ATLAS完整模拟的一个事件)使灵活的工作流对资源的占用很少,这样与每个作业处理0 (GB)文件的传统工作流相比,可以更充分、更有效地利用周期。
{"title":"Fine-Grained Processing Towards HL-LHC Computing in ATLAS","authors":"D. Benjamin, P. Calafiura, T. Childers, K. De, A. Girolamo, E. Fullana, W. Guan, T. Maeno, Nicolò Magini, P. Nilsson, D. Oleynik, Shaojun Sun, V. Tsulaia, P. Gemmeren, T. Wenaus, W. Yang","doi":"10.1109/eScience.2018.00083","DOIUrl":"https://doi.org/10.1109/eScience.2018.00083","url":null,"abstract":"During LHC's Run-2 ATLAS has been developing and evaluating new fine-grained approaches to workflows and dataflows able to better utilize computing resources in terms of storage, processing and networks. The compute-limited physics of ATLAS has driven the collaboration to aggressively harvest opportunistic cycles from what are often transiently available resources, including HPCs, clouds, volunteer computing, and grid resources in transitional states. Fine-grained processing (with typically a few minutes' granularity, corresponding to one event for the present ATLAS full simulation) enables agile workflows with a light footprint on the resource such that cycles can be more fully and efficiently utilized than with conventional workflows processing O(GB) files per job.","PeriodicalId":6476,"journal":{"name":"2018 IEEE 14th International Conference on e-Science (e-Science)","volume":"1 1","pages":"338-338"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79131688","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2018 IEEE 14th International Conference on e-Science (e-Science)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1