Advances in the Kepler Scientific Workflow System and Its Applications (Kepler) Session 1

Time and Date: 10:15 - 11:55 on 7th June 2016

Room: Boardroom West

Chair: Jianwu Wang

504 Kepler WebView: A Lightweight, Portable Framework for Constructing Real-time Web Interfaces of Scientific Workflows [abstract]
Abstract: Modern web technologies facilitate the creation of high-quality data visualizations, and rich, interactive components across a wide variety of devices. Scientific workflow systems can greatly benefit from these technologies by giving scientists a better understanding of their data or model leading to new insights. While several projects have enabled web access to scientific workflow systems, they are primarily organized as a large portal server encapsulating the workflow engine. In this vision paper, we propose the design for Kepler WebView, a lightweight framework that integrates web technologies with the Kepler Scientific Workflow System. By embedding a web server in the Kepler process, Kepler WebView enables a wide variety of usage scenarios that would be difficult or impossible using the portal model.
Daniel Crawl, Alok Singh, Ilkay Altintas
291 A Smart Manufacturing Use Case: Furnace Temperature Balancing in Steam Methane Reforming Process via Kepler Workflows [abstract]
Abstract: The industrial scale production of hydrogen gas through steam methane reforming (SMR) process requires an optimum furnace temperature distribution to not only maximize the hydrogen yield but also increase the longevity of the furnace infrastructure which usually operates around 1300 degree Kelvin (K). Kepler workflows are used in temperature homogenization, termed as balancing of this furnace through Reduced Order Model (ROM) based Matlab calculations using the dynamic temperature inputs from an array of infrared sensors. The outputs of the computation are used to regulate the flow rate of fuel gases which in turn optimizes the temperature distribution across the furnace. The input and output values are stored in a data historian which is a database for real-time data and events. Computations are carried out on an OpenStack based cloud environment running Windows and Linux virtual machines. Additionally, ab initio computational fluid dynamics (CFD) calculation using Ansys Fluent software is performed to update the ROM periodically. ROM calculations complete in few minutes whereas CFD calculations usually take a few hours to complete. The Workflow uses an appropriate combination of the ROM and CFD models. The ROM only workflow currently runs every 30 minutes to process the real-time data from the furnace, while the ROM CFD workflow runs on demand. ROM only workflow can also be triggered by an operator of the furnace on demand.
Prakashan Korambath, Jianwu Wang, Ankur Kumar, Jim Davis, Robert Graybill, Brian Schott, Michael Baldea
335 Running simultaneous Kepler sessions for the parallelization of parametric scans and optimization studies applied to complex workflows [abstract]
Abstract: In this paper we present approach taken to run multiple Kepler sessions at the same time. This kind od execution is one of the requirements for the codes developed within EUROfusion. It allows to gain speed and resources. The choice of Integrated Modelling made by the former EFDA ITM-TF and pursued now under EUROfusion WPCD is unique and original: it entails the development of a comprehensive and completely generic tokamak simulator including both the physics and the machine, which can be applied for any fusion device. All components are linked inside workflows and this way allow complex coupling of various algorithms while providing consistency. Workflows are composed of Kepler and Ptolemy II elements as well as set of the native libraries written in various languages (Fortran, C, C++). In addition to that, there are Python based components that are used for visualization of results as well as for pre/post processing. At the bottom of all these components there is database layer that may vary between software releases. All these constraints make it really challenging to run multiple Kepler sessions at the same time. However, ability to run numerous sessions in parallel is a must - to reduce computation time and to make it possible to run released codes while working with new software at the same time. In this paper we present our approach to solve this issue and we present applications of this approach.
Michał Owsiak, Marcin Plociennik, Bartek Palak, Tomasz Zok, Cedric Reux, Luc Di Gallo, Mireille Schneider, Thomas Johnson, Denis Kalupin
444 Forest fire spread prediction system workflow: an experience using Kepler [abstract]
Abstract: Natural hazard prediction systems provide a key information to mitigate the effects of such disasters. Having this information in advance, taking decision teams could be more efficient and effective. However, run these systems in real time with data assimilation is not a trivial task. Most of the studies using these systems are post disaster analysis. For that reason, scientific workflow software and sensor cyberinfraestructures are crucial to monitor the system execution and take profit of data assimilation. In this work, a workflow for a forest fire spread prediction system is proposed using Kepler. The main modules of the workflow are WindNinja and FARSITE. The sources of the data assimilation come from HPWREN weather stations. The workflow allows running a spread prediction for a given ignition boundary using data assimilation and monitoring the execution of the different modules of the workflow.
Tomàs Artés, Daniel Crawl, Ana Cortes and Ilkay Altintas