ICCS 2018 Main Track (MT) Session 1

Time and Date: 13:35 - 15:15 on 11th June 2018

Room: M1

Chair: Travis Desell

272 Optimizing the Efficiency, Vulnerability and Robustness of Road-based Para-transit Networks using Genetic Algorithm [abstract]
Abstract: In the developing world, majority of people usually take para-transit services for their everyday commutes. However, their informal and demand-driven operation, like making arbitrary stops to pick up and drop off passengers, has been inefficient and poses challenges to efforts in integrating such services to more organized train and bus networks. In this study, we devised a methodology to design and optimize a road-based para-transit network using a genetic algorithm to optimize efficiency, robustness, and invulnerability. We first generated stops following certain geospatial distributions and connected them to build networks of routes. From them, we selected an initial population to be optimized and applied the genetic algorithm. Overall, our modified genetic algorithm with 20 evolutions optimized the 20% worst performing networks by 84% on average. For one network, we were able to significantly increase its fitness score by 223%. The highest fitness score the algorithm was able to produce through optimization was 0.532 from a score of 0.303.
Briane Paul Samson, Gio Anton Velez, Joseph Ryan Nobleza, David Sanchez and Jan Tristan Milan
402 Learning network properties through optimizing principal eigenvector localization [abstract]
Abstract: Networks furnish a mathematical framework to model and decipher the collective behavior of the complex real-world systems. Scrutiny of principal eigenvector (PEV) and the corresponding eigenvalue of the networks are known to provide an understanding of various local and global structural as well as the dynamical evolution of the networks. Recently scrutiny of eigenvector localization has received tremendous attention in network science due to its versatile applicability in many different areas which includes analyzing centrality measure, spectral partitioning, development of approximation algorithms and disease spreading phenomenon. Moreover, lower order eigenvectors have been studied to develop machine learning tools. They used inverse participation ratio (IPR) as well as another kind of measure for the eigenvector localization, called as statistical leverage score which has an impact on modern big data analysis. One key factor of our interest is to understand properties of networks which may help in spreading or restricting perturbation in networks captured by the PEV localization. For a network, an eigenvector is said to be localized when most of its components are near to zero, with few taking very high values. To study the structural and spectral properties of networks as PEV changes its state from the delocalized to a highly localized, we evolve an initial random network. We develop randomized algorithms considering the IPR as an objective function. We demonstrate that PEV localization is not a consequence of a single network property and rather requires collective impact of several structural features. The final optimized network possesses a special structure independent of the initial network. It reveals that optimized network consists of two graph components of different sizes which are connected to each other via a single node (in Fig. 1). Moreover, the optimized structure contains a special set of edges, rewiring any one of them leads to a complete delocalization of the PEV from a highly localized state (Fig. 2). This sensitivity of the PEV at the most localized state turns out to be related to the behavior of the largest and the second largest eigenvalue of the network. Precisely when the network becomes most localized, the second largest eigenvalue of the adjacency matrix become very close to the largest eigenvalue (Fig. 3). Furthermore, we identify an evolution regime where networks are as localized as the optimized one, but, are robust to single edge rewiring (in Fig. 2). Moreover, we use the Susceptible-Infected-Susceptible (SIS) disease spreading model and verify that the structure of the networks in the intermediate and optimized stages slows down the spreading process than the initial random structure (Fig. 4).
Priodyuti Pradhan and Sarika Jalan
29 On the Configuration of Robust Static Parallel Portfolios for Efficient Plan Generation [abstract]
Abstract: Automated Planning has achieved a significant step forward in the last decade, and many advanced planning engines have been introduced. Nowadays, increases in computational power are mostly achieved through hardware parallelisation. In view of the increasing availability of multicore machines and of the intrinsic complexity of designing parallel algorithms, a natural exploitation of parallelism is to combine existing sequential planning engines into parallel portfolios. In this work, we introduce three techniques for an automatic configuration of static parallel portfolios of planning engines. The aim of generated portfolios is to provide a good tradeoff performance between coverage and runtime, on previously unseen problems. Our empirical results demonstrate that our techniques for configuring parallel portfolios combine strengths of planning engines, and fully exploit multicore machines.
Mauro Vallati, Lukas Chrpa and Diane Kitchin
192 A High-Performance Evolutionary Computation Framework for Scalable Spatial Optimization [abstract]
Abstract: Spatial optimization (SO) is an important and prolific field of interdisciplinary research. Spatial optimization methods seek optimal allocation or arrangement of spatial units under constraints such as distance, adjacency, and contiguity. Evolutionary Algorithms (EA) are well-known heuristic approach for solving SO problems. However, as spatial granularity becomes finer and problem formulations comprise increasingly complex spatial relationships, EA solvers must be adapted with greater computational efficiency and algorithmic effectiveness. We present a solution that addresses the efficiency issue by leveraging massive computing power on supercomputers. The computational scalability challenge is tackled with a parallel EA library that eliminates the costly global synchronization in massively parallel computing environment. Our implementation scales to 131,072 processors on the Blue Waters supercomputer. We develop novel spatially explicit EA recombination operators that implement crossover and mutation using intelligently guided strategies in spatially constrained decision space, inspired by path relinking and ejection chain heuristics. Our high-performance EA framework is employed as a computational solution for the complex problem of political redistricting in the United States. It enables the creation of billions of feasible districting plans that adhere to U.S. Supreme Court mandates and further facilitates statistical analysis of gerrymandering.
Yan Liu and Wendy K. Tam Cho

ICCS 2018 Main Track (MT) Session 2

Time and Date: 15:45 - 17:25 on 11th June 2018

Room: M1

Chair: Yan Liu

229 SDF-Net: Real-time Rigid Object Tracking Using a Deep Signed Distance Network [abstract]
Abstract: In this paper, a deep neural network is used to model the signed distance function (SDF) of a rigid object for real-time tracking using a single depth camera. By leveraging the generalization capability of the neural network, we could better represent the model of the object implicitly. With the training stage done off-line, our proposed methods are capable of real-time performance and running as fast as 1.29 ms per frame on one CPU core, which is suitable for applications with limited hardware capabilities. Furthermore, the memory footprint of our trained SDF-Net for an object is less than 10 kilobytes. A quantitative comparison using public dataset is being carried out and our approach is comparable with the state-of-the-arts. The methods are also tested on actual depth records to evaluate their performance in real-life scenarios.
Prayook Jatesiktat, Ming Jeat Foo, Guan Ming Lim and Wei Tech Ang
232 Insider Threat Detection with Deep Neural Network [abstract]
Abstract: Insider threat detection has attracted a considerable interest from the researchers and industries. Existing work mainly focused on applying machine-learning techniques to detecting insider threat. However, this work requires “feature engineering” which is difficult and time-consuming. As we know, the deep learning technique can automatically learn powerful features. In this paper, we present a novel insider-threat detection method with Deep Neural Network (DNN) based on user behavior. Specifically, we use the LSTM-CNN framework to recognize user’s anomalous behavior. First, similar to natural language modeling, we use the Long Short Term Memory (LSTM) to learn the language of user behavior through user actions and extract abstracted temporal features. Second, the extracted features are converted to the fixed-size feature matrices and the Convolutional Neural Network (CNN) use these fixed-size feature matrices to detect insider threat. We conduct experiments on a public dataset of insider threats. Experimental results show that our method is indeed successful at detecting insider threat and we obtained AUC = 0.9449 in best case.
Fangfang Yuan, Yanan Cao, Yanmin Shang, Yanbing Liu, Jianlong Tan and Binxing Fang
80 Incentive Mechanism for Cooperative Intrusion Detection: an Evolutionary Game Approach [abstract]
Abstract: In Mobile Ad-Hoc Networks, cooperative intrusion detection is efficient and scalable to massively parallel attacks. However, due to concerns of privacy leakage and resource costs, if without enough incentives, most mobile nodes are often selfish and disinterested in helping others to detect an intrusion event, thus an efficient incentive mechanism is required. In this paper, we formulate the incentive mechanism for cooperative intrusion detection as an evolutionary game and achieve an optimal solution to help nodes decide whether to participate in detection or not. Our proposed mechanism can deal with the problems that cooperative nodes do not own complete knowledge about other nodes. We develop a game algorithm to maximize nodes’utility. Simulations demonstrate that our strategy can efficiently incentivize potential nodes to cooperate.
Yunchuan Guo, Han Zhang, Lingcui Zhang, Liang Fang and Fenghua Li

ICCS 2018 Main Track (MT) Session 3

Time and Date: 13:15 - 14:55 on 12th June 2018

Room: M1

Chair: Panruo Wu

152 Hybrid Genetic Algorithm for an On-Demand First Mile Transit System using Electric Vehicles [abstract]
Abstract: First/Last mile gaps are a significant hurdle in large scale adoption of public transit systems. Recently, demand responsive transit systems have emerged as a preferable solution to first/last mile problem. However, existing work requires significant computation time or advance bookings. Hence, we propose a public transit system linking the neighborhoods to a rapid transit node using a fleet of demand responsive electric vehicles, which reacts to passenger demand in real-time. Initially, the system is modeled using an optimal mathematical formulation. Owing to the complexity of the model, we then propose a hybrid genetic algorithm that computes results in real-time with an average accuracy of 98%. Further, results show that the proposed system saves travel time up to 19% compared to the existing transit services.
Thilina Perera, Alok Prakash, Chathura Nagoda Gamage and Thambipillai Srikanthan
179 Comprehensive Learning Gene Expression Programming for Automatic Implicit Equation Discovery [abstract]
Abstract: Automatic Implicit Equation Discovery (AIED), which aims to automatically find implicit equations to fit observed data, is a promising and challenging research topic in data mining and knowledge discovery. Existing methods for AIED are designed based on calculating derivatives, which require high computational cost and will become invalid when the problem encountered contains sparse training data. To tackle these drawbacks, this paper proposes a new mechanism named Comprehensive Learning Fitness Evaluation Mechanism (CL-FEM). The mechanism is capable of learning knowledge from both the given training data and the disturbed data generated by adding stochastic noise to the training data, to measure the validity and fitting error of a given equation model. The proposed CL-FEM is further integrated with the Self-Learning Gene Expression Programming (SL-GEP), forming a Comprehensive Learning Gene Expression Programming (CL-GEP) to solve AIED problems. The proposed CL-GEP is tested on several benchmark problems with different scales and difficulties, and the experiment results have demonstrated that the CL-GEP can offer very promising performance.
Yongliang Chen and Jinghui Zhong
203 Multi-population Genetic Algorithm for Cardinality Constrained Portfolio Selection Problems [abstract]
Abstract: Portfolio Selection (PS) is recognized as one of the most important and challenging problems in financial engineering. The aim of PS is to distribute a given amount of investment fund across a set of assets in such a way that the return is maximised and the risk is minimised. To solve PS more effectively and more effectively, this paper introduces a Multi-population Genetic Algorithm (MPGA) methodology. The proposed MPGA decomposes a large population into multiple populations to explore and exploit the search space simultaneously. These populations evolve independently during the evolutionary learning process. Yet different populations periodically exchange their individuals so promising genetic materials could be shared between different populations. The proposed MPGA method was evaluated on the standard PS benchmark instances. The experimental results show that MPGA can find better investment strategies in comparison with state-of-the-art portfolio selection methods. In addition, the search process of MPGA is more efficient than these existing methods requiring significantly less amount of computation.
Nasser Sabar, Ayad Turky and Andy Song
343 Recognition and Classification of Rotorcraft by Micro-Doppler Signatures using Deep Learning [abstract]
Abstract: Detection and classification of rotorcraft targets are of great significance not only in civil fields but also in defense. However, up to now, it is still difficult for the traditional radar signal processing methods to detect and distinguish rotorcraft targets from various types of moving objects. Moreover, it is even more challeng-ing to classify different types of helicopters. As the development of high-precision radar, classification of moving targets by micro-Doppler features has become a promising research topic in the modern signal processing field. In this paper, we propose to use the deep convolutional neural networks (DCNNs) in rotorcraft detection and helicopter classification based on Doppler radar signals. We apply DCNN directly to raw micro-Doppler spectrograms for rotorcraft de-tection and classification. The proposed DCNNs can learn the features automati-cally from the micro-Doppler signals without introducing any domain back-ground knowledge. Simulated data are used in the experiments. The experimental results show that the proposed DCNNs achieve superior accuracy in rotorcraft detection and superior accuracy in helicopter classification, outperforming the tra-ditional radar signal processing methods.
Ying Liu and Jinyi Liu
360 Data Allocation based on Evolutionary Data Popularity Clustering [abstract]
Abstract: This study is motivated by the high-energy physics experiment ATLAS, one of the four major experiments at the Large Hadron Collider at CERN. ATLAS comprises 130 data centers worldwide with datasets in the Petabyte range. In the processing of data across the grid, transfer delays and subsequent performance loss emerged as an issue. The two major costs are the waiting time until input data is ready and the job computation time. In the ATLAS workflows, the input to computational jobs is based on grouped datasets. The waiting time stems mainly from WAN transfers between data centers when job properties require execution at one data center but the dataset is distributed among multiple data centers. The proposed novel data allocation algorithm redistributes the constituent files of datasets such that the job efficiency is increased in terms of a cost metric. An evolutionary algorithm is proposed that addresses the data allocation problem in a network based on data popularity and clustering. The number of expected job’s file transfers is used as the target metric and it is shown that job waiting times can be decreased by faster input data readiness.
Ralf Vamosi, Mario Lassnig and Erich Schikuta

ICCS 2018 Main Track (MT) Session 4

Time and Date: 15:25 - 17:05 on 12th June 2018

Room: M1

Chair: Thilina Perera

370 Hyper-heuristic Online Learning for Self-assembling Swarm Robots [abstract]
Abstract: Robot swarm is a solution for difficult and large scale tasks. However controlling and coordinating a swarm of robots is a challenge because of the complexity and uncertainty of the environment where manual programming of robot behaviours is often impractical. In this study we proposed a hyper-heuristic methodology for swarm robots. It allows robots to create suitable actions based on a set of low-level heuristics. Each heuristic is a behavioural element. With online learning, the robot behaviours can be improved during the executions by autonomous heuristic adjustment. The proposed hyper-heuristic framework is applied to building surface cleaning tasks where multiple separate surfaces exist and the complete surface information is difficult to obtain. Under this scenario, the robot swarm need not only to clean the surfaces efficiently by distributing the robots, but also to move across surfaces by self-assembling into a bridge structure. Experiment results showed the effectiveness of the hyper-heuristic framework as a group of robots were able to autonomously clean multiple surfaces of different layouts without prior programming. Their behaviours can improve over time because of the online learning mechanism.
Shuang Yu, Aldeida Aleti, Jan Carlo Barca and Andy Song
241 An Innovative Heuristic for Planning-based Urban Traffic Control [abstract]
Abstract: The global growth in urbanisation increases the demand for services including road transport infrastructure, presenting challenges in terms of mobility. In this scenario, optimising the exploitation of urban road network is a pivotal challenge, particularly in the case of unexpected situations. In order to tackle this challenge, approaches based on mixed discrete-continuous planning have been recently proposed and although their feasibility has been demonstrated, there are a lack of informative heuristics for this class of applications. Therefore, existing approaches tend to provide low-quality solutions, leading to a limited impact of generated plans on the actual urban infrastructure. In this work, we introduce the Time-Based heuristic: a highly informative heuristic for PDDL+ planning-based urban traffic control. The heuristic, which has an admissible and an inadmissible variant, has been evaluated considering scenarios that use real-world data.
Santiago Franco, Alan Lindsay, Mauro Vallati and Lee Mccluskey
111 Automatic Web News Extraction Based on DS Theory Considering Content Topics [abstract]
Abstract: In addition to the news content, most news web pages also contain various noises, such as advertisements, recommendations, and navigation panels. These noises may hamper the studies and applications which require pre-processing to extract the news content accurately. Existing methods of news content extraction mostly rely on non-content features, such as tag path, text layout, and DOM structure. However, without considering topics of the news content, these methods are difficult to recognize noises whose external characteristics are similar to those of the news content. In this paper, we propose a method that combines non-content features and a topic feature based on Dempster-Shafer (DS) theory to increase the recognition accuracy. We use maximal compatibility blocks to generate topics from text nodes and then obtain feature values of topics. Each feature is converted into evidence for the DS theory which can be utilized in the uncertain information fusion. Experimental results on English and Chinese web pages show that combining the topic feature by DS theory can improve the extraction performance obviously.
Kaihang Zhang, Chuang Zhang, Xiaojun Chen and Jianlong Tan
118 DomainObserver: A Lightweight Solution for Detecting Malicious Domains Based on Dynamic Time Warping [abstract]
Abstract: People use the Internet to shop, access information and enjoy entertainment by browsing web sites. At the same time, cyber-criminals operate malicious domains to spread illegal information and acquire money, which poses a great risk to the security of cyberspace. Therefore, it is of great importance to detect malicious domains in the field of cyberspace security. Typically, there are broad research focusing on detecting malicious domains either by blacklist or exploiting the features via machine learning techniques. However, the former is infeasible due to the limited crowd, and the later requires complex feature engineering. Different from most of previous methods, in this paper, we propose a novel lightweight solution named DomainObserver to detect malicious domains. Our technique of DomainObserver is based on dynamic time warping that is used to better align the time series. To the best of our knowledge, it is a new trial to apply passive traffic measurements and time series data mining to malicious domain detection. Extensive experiments on real datasets are performed to demonstrate the effectiveness of our proposed method.
Guolin Tan, Peng Zhang, Qingyun Liu, Xinran Liu and Chunge Zhu
157 You Have More Abbreviations than You Know: A Study of AbbrevSquatting Abuse [abstract]
Abstract: Domain squatting is a speculative behavior involving the registration of domain names that are trademarks belonging to popular companies, important organizations or other individuals, before the latters have a chance to register. This paper presents a specific and unconcerned type of domain squatting called “AbbrevSquatting”, the phenomena that mainly happens on institutional websites. As institutional domain names are usually named with abbreviations(i.e., short forms) of the full names or official titles of institutes, attackers can mine abbreviation patterns from existed pairs of abbreviations and full names, and register forged domain names with unofficial but meaningful abbreviations for a given institute. To measure the abuse of AbbrevSquatting, we first mine the common abbreviation patterns used in institutional domain names, and generate potential AbbrevSquatting domain names with a data set of authoritative domains. Then, we check the maliciousness of generated domains with a public API and seven different blacklists, and group the domains into several categories with crawled data. Through a series of manual and automated experiments, we discover that attackers have already been aware of the principles of AbbrevSquatting and are monetizing them in various unethical and illegal ways. Our results suggest that AbbrevSquatting is a real problem that requires more attentions from security communities and institutions' registrars.
Pin Lv, Jing Ya, Tingwen Liu, Jinqiao Shi, Binxing Fang and Zhaojun Gu

ICCS 2018 Main Track (MT) Session 5

Time and Date: 9:00 - 10:40 on 13th June 2018

Room: M1

Chair: Yang Cai

233 Large Scale Retrieval of Social Network Pages by Interests of Their Followers [abstract]
Abstract: Social networks provide an opportunity to form communities of people that share their interests on a regular basis (circles of fans of different music, books, kinds of sports, etc.). Every community manifests these interests creating lots of linguistic data to attract new followers to certain pages and support existing clusters of users. In the present article, we suggest a model of retrieving such pages that attract users with similar interests, from a large collection of pages. We test our model on three types of pages manually retrieved from the social network Vkontakte and classified as interesting for a. football fans, b. vegetarians, c. historical reenactors. We use such machine learning classifiers as Naive Bayes, SVM, Logistic Regression, Decision Trees to compare their performance with the performance of our system. It appears that the mentioned classifiers can hardly retrieve (i.e. single out) pages with a particular interest that form a small collection of 30 samples from a collection as large as 4,090 samples. In particular, our system exceeds their best result (F1-score=0.65) and achieves F1-score of 0.72.
Elena Mikhalkova, Yuri Karyakin and Igor Glukhikh
326 Parallel data-driven modeling of information spread in social networks [abstract]
Abstract: Models of information spread in social networks are widely used to explore the drivers of content contagion and to predict the effect of new information messages. Most of the existing models (aggregated as SIR-like or network-based as independent cascades) use the assumption of homogeneity of an audience. However, to make a model plausible for a description of real-world processes and to measure the accumulated impact of information on individuals, one needs to personalize the characteristics of users as well as sources of information. In this paper, we propose an approach to data-driven simulation of information spread in social networks which combines a set of different models in a unified framework. It includes a model of a user (including sub-models of reaction and daily activity), a model of message generation by information source and a model of message transfer within a user network. The parameters of models (e.g. for different types of agents) are identified by data from the largest Russian social network vk.com. For this study, we collected the network of users associated with charity community (~33.7 million nodes). To tackle with huge size of networks, we implement-ed parallel version of modeling framework and tested it on the Lomonosov supercomputer. We identify key parameters of models that may be tuned to re-produce observable behavior and show that our approach allows to simulate aggregated dynamics of reactions to a series of posts as a combination of individual responses.
Oksana Severiukhina, Klavdiya Bochenina, Sergey Kesarev and Alexander Boukhanovsky
375 Topology of Thematic Communities in Online Social Networks: A Comparative Study [abstract]
Abstract: The network structure of communities in social media significantly affects diffusion processes which implement positive or negative information influence on social media users. Some of the thematic communities in online social networks may provide illegal services or information in them may cause undesired psychological effects; moreover, the topology of such communities and behavior of their members are influenced by a thematic. Nevertheless, recent research does not contain enough detail about the particularities of thematic communities formation, or about the topological properties of underlying friendship networks. To address this gap, in this study we analyze structure of communities of different types, namely, carders, commercial sex workers, substance sellers and users, people with radical political views, and compare them to the 'normal' communities (without a single narrow focus). We discovered that in contrast to ordinary communities which have positive assortativity (as expected for social networks), specific thematical communities are significantly disassortative. Types of anomalous communities also differ not only in content but in structure. The most specific are the communities of radicalized individuals: it was shown that they have the highest connectivity and the larger part of nodes within a friendship graph.
Valentina Y. Guleva, Danila Vaganov, Daniil Voloshin and Klavdiya Bochenina
70 A distance-based tool-set to track inconsistent urban structures through complex-networks [abstract]
Abstract: Complex networks can be used for modeling street meshes and urban agglomerates. With such a model, many aspects of a city can be investigated to promote a better quality of life to its citizens. Along these lines, this paper proposes a set of distance-based pattern-discovery algorithmic instruments to improve urban structures modeled as complex networks, detecting nodes that lack access from/to points of interest in a given city. Furthermore, we introduce a greedy algorithm that is able to recommend improvements to the structure of a city by suggesting where points of interest are to be placed. We contribute to a thorough process to deal with complex networks, including mathematical modeling and algorithmic innovation. The set of our contributions introduces a systematic manner to treat a recurrent problem of broad interest in cities.
Gabriel Spadon, Bruno B. Machado, Danilo M. Eler and Jose Fernando Rodrigues Jr.

ICCS 2018 Main Track (MT) Session 6

Time and Date: 11:10 - 12:50 on 13th June 2018

Room: M1

Chair: Klavdiya Bochenina

123 A Conceptual Framework for Social Movements Analytics for National Security [abstract]
Abstract: Social media tools have changed our world due to the way they convey information between individuals; this has led to many social movements either starting on social media or being organised and managed through this medium. At times however, certain human-induced events can trigger Human Security Threats such as Personal Security, Health Security, Economic Security or Political Security. The aim of this paper is to propose a holistic Data Analysis Framework for examining Social Movements and detecting pernicious threats to National Security interests. As a result of this, the proposed framework focuses on three main stages of an event (Detonating Event, Warning Period and Crisis Interpretation) to provide timely additional insights, enabling policy makers, first responders, and authorities to determine the best course of action. The paper also outlines the possible computational techniques utilised to achieve in depth analysis at each stage. The robustness and effectiveness of the framework are demonstrated by dissecting Warning Period scenarios, from real-world events, where the increase of Human Security aspects were key to identifying likely threats to National Security.
Pedro Cardenas, Georgios Theodoropoulos, Boguslaw Obara and Ibad Kureshi
142 Retweet Prediction using Social-aware Probabilistic Matrix Factorization [abstract]
Abstract: Retweet prediction is a fundamental and crucial task in social networking websites as it may influence the process of information diffusion. Existing prediction approaches consider social network structure, but social context has not been fully considered. It is important to incorporate social contextual information into retweet prediction. Besides, the sparsity of retweet data severely disturb the performance of these models. In this paper, we propose a novel retweet prediction framework based on probabilistic matrix factorization model to integrate the observed retweet data, social influence and message embeddings semantic to improve the accuracy of prediction. We then incorporate these regularization terms into the objective function. Comprehensive experiments on the real-world dataset clearly validate both the effectiveness and efficiency of our model compared with several state-of the-art baselines.
Bo Jiang, Zhigang Lu, Ning Li, Jianjun Wu and Zhengwei Jiang
243 Cascading Failure Based on Load Redistribution of a Smart Grid with Different Coupling Modes [abstract]
Abstract: As one of the most important properties of the power grid, the voltage load plays an important role in the cascading failure of the smart grid and load redistribution can accelerate the speed of the failure by triggering more nodes to overload and fail. The subnet structure and different coupling modes also affect the robustness of the smart grid. However, the research on the effect of load, subnet structure and coupling mode on the cascading failure of the smart grid is still rare. In this paper, the smart grid with two-way coupling link consists of a power grid with small world topology and a communication network with scale- free topology. An improved load-capacity model is applied to overload- induced failure in the power grid and node importance (NI) is used as an evaluation index to assess the effect of nodes on the power grid and communication network. We propose three kinds of coupling modes based on NI of nodes between the cyber and physical subnets, i.e., Random Coupling in Subnets (RCIS), Assortative Coupling in Subnets (ACIS) and Disassortative Coupling in Subnets (DCIS). In order to improve the robustness of the smart grid, a cascading failure model based on load redistribution is proposed to analyze the influence of different coupling modes on the cascading failure of the smart grid under both a targeted attack and random attack. Some findings are summarized as: (I) The robustness of the smart grid is improved by increasing the tolerance α. (II) ACIS applied to the bottom-up coupling link is more beneficial in enhancing the robustness of the smart grid than DCIS and RCIS, regardless of a targeted attack or random attack.
Wenjie Kang, Peidong Zhu and Gang Hu
391 Measuring social responsiveness for improving handling of extreme situations [abstract]
Abstract: Volunteering and community reaction is known to be an essential part of response to critical events. With the rapid evolution of new means of communication, it has transformed accordingly. A new category of volunteers emerged – those that are not in the proximity to the area of emergency but willing to help the affected. Widely known as digital volunteers, they help aggregate, disseminate and distribute information to increase and maintain the awareness of stakeholders and resourceful individuals about the situation. There has been an upsurge of investigations of roles, timelines and aggregate characteristics of emergent communication. Compared to that, characteristics of crisis-related social media posts that predict wider social response to date have been studied modestly. In this research we are studying the process of reaction of potential digital volunteers to different extreme situations in social media platform.
Nikolay Butakov, Timur Fatkulin and Daniil Voloshin
287 Pheromone Model Based Visualization of Malware Distribution Networks [abstract]
Abstract: We present a novel computational pheromone model for describing dynamic network behaviors in terms of transition, persistency, and hosting. The model consists of a three-dimensional force-directed graph with bi-directional pheromone deposit and decay paths. A data compression algorithm is developed to optimize computational performance. We applied the model for visual analysis of a Malware Distribution Network (MDN), a connected set of maliciously compromised domains used to disseminate malicious software to victimize computers and users. The MDN graphs are extracted from crowdsourcing datasets from Google Safe Browsing (GSB) reports with attributions from VirusTotal report site. Our research shows that this novel approach reveals patterns of topological changes of the network over time, including the existence of persistent subnetworks and individual TLDs’ critical to the successful operation of MDNs, and the dynamics of the topological changes on daily basis. From the visualization, we observed notable clustering effects, and also noticed life span patterns for high edge count malware distribution clusters.
Yang Cai, Jose Morales and Sihan Wang

ICCS 2018 Main Track (MT) Session 7

Time and Date: 13:35 - 15:15 on 11th June 2018

Room: M2

Chair: Roberto Ribeiro

15 A Computational Model-Based Framework to Plan Clinical Experiments – an Application to Vascular Adaptation Biology [abstract]
Abstract: Several computational models have been developed in order to improve the outcome of Vein Graft Bypasses in response to arterial occlusions and they all share a common property: their accuracy relies on a winning choice of the coefficients’ value related to biological functions that drive them. Our goal is to optimize the retrieval of these unknown coefficients on the base of experimental data and accordingly, as biological experiments are noisy in terms of statistical analysis and the models are typically stochastic and complex, this work wants first to elucidate which experimental measurements might be sufficient to retrieve the targeted coefficients and second how many specimens would constitute a good dataset to guarantee a sufficient level of accuracy. Since experiments are often costly and time consuming, the planning stage is critical to the success of the operation and, on the base of this consideration, the present work shows how, thanks to an ad hoc use of a computational model of vascular adaptation, it is possible to estimate in advance the entity and the quantity of resources needed in order to efficiently repro-duce the experimental reality.
Stefano Casarin, Scott Berceli and Marc Garbey
138 Accelerating Data Analysis in Simulation Neuroscience with Big Data Technologies [abstract]
Abstract: Important progress in computational sciences has been made possible recently thanks to the increasing computing power of high performance systems. Following this trend, larger scientific studies, like brain tissue simulations, will continue to grow in the future. In addition to the challenges of conducting these experiments, we foresee an explosion of the amount of data generated and the consequent unfeasibility of analyzing and understanding the results with the current techniques. This paper proposes Neurolytics, a new data analysis framework, together with a new data layout. The implementation of Neurolytics is mainly focused on simulation neuroscience, although we believe that our design can be applied to other science domains. The framework relies on big data technologies, like Apache Spark, to enable fast, reliable and distributed analyses of brain simulation data in this case. Our experimental evaluation on a cluster of 100 nodes shows that Neurolytics gets up to 374x speed-up compared to a thread-parallel Python implementation and is on par with a highly optimized Spark code. This demonstrates the suitability of our proposal to help scientists structure and understand the results of their experiments in a fast and efficient way.
Judit Planas, Fabien Delalondre and Felix Schuermann
193 Spiral wave drift induced by high-frequency forcing. Parallel simulation in the Luo-Rudy anisotropic model of cardiac tissue [abstract]
Abstract: Non-linear waves occur in various physical, chemical and biological media. One of the most important examples is electrical excitation waves in the myocardium, which initiate contraction of the heart. Abnormal wave propagation in the heart, such as the formation of spiral waves, causes dangerous arrhythmias, and thus methods of elimination of such waves are of great interest. One of the most promising methods is so-called low-voltage cardioversion and defibrillation, which is believed to be achieved by inducing the drift and disappearance of spiral waves using external high-frequency electrical stimulation of the heart. In this paper, we perform a computational analysis of the interaction of spiral waves and trains of high-frequency plane waves in 2D models of cardiac tissue. We investigate the effectiveness and safety of the treatment. We also identify the dependency of drift velocity on the period of plane waves. The simulations were carried out using a parallel computing system with OpenMP technology.
Timofey Epanchintsev, Sergei Pravdin and Alexander Panfilov
317 Understanding Malaria induced red blood cell deformation using data-driven Lattice Boltzmann Simulations [abstract]
Abstract: Malaria remains a deadly disease that affected millions of people in 2016. Among the five Plasmodium (P.) parasites which contribute to malaria dis-eases in humans. P. falciparum is a lethal one which is responsible for the majority of the world-wide-malaria-related deaths. Since the banana-shaped stage V gametocytes play a crucial role in disease transmission, understand-ing the deformation of single stage V gametocytes may offer deeper insights into the development of the disease and provide possible targets for new treatment methods. In this study we used lattice Boltzmann-based simula-tions to investigate the effects of the stretching forces acting on infected red blood cells inside a slit-flow cytometer. The parameters that represent the cellular deformability of healthy and malaria infected red blood cells are chosen such that they mimic the deformability of these cells in a slit-flow cy-tometer. The simulation results show good agreement with experimental data and allow for studying the transportation of malaria infected red blood cell in blood circulation.
Joey Sing Yee Tan, Gabor Zavodszky and Peter M. A. Sloot
347 Towards Model-based Policy Elaboration on City Scale using Game Theory: Application to Ambulance Dispatching [abstract]
Abstract: The paper presents early results on the development of a generalized approach for modeling and analysis of the interaction of multiple stakeholders in city environ-ment while providing services to citizens under the regulation of city authorities. The approach considers the interaction between main stakeholders (organizations of various kind, citizens, and city authorities) including information and finances exchange, activities taken and services or goods provided. The developed ap-proach is based on a combination of game-theoretic modeling and simulation of service providers interaction. Such combination enables consideration of con-fronting stakeholders as well as determined (e.g., scheduled) and stochastic varia-tion in characteristics of system’s elements. The goal of this approach develop-ment is supporting of analysis and optimization of city-level regulation through legislative, financial, and informational interaction with organizations and envi-ronment of a city. An example of ambulance dispatching during providing emer-gent care for acute coronary syndrome (ACS) patients is considered. The exam-ple is analyzed in a simplified linear case and in practical application to dispatch-ing ambulances providing service for ACS patients in Saint Petersburg.
Sergey Kovalchuk, Mariia Moskalenko and Alexey Yakovlev

ICCS 2018 Main Track (MT) Session 8

Time and Date: 15:45 - 17:25 on 11th June 2018

Room: M2

Chair: Stefano Casarin

68 Elucidation of Mechanism for Reducing Porosity in Electric Arc Spraying through CFD [abstract]
Abstract: We elucidated the mechanism for reducing the porosity (a means for achieving smaller globules) through Computational Fluid Dynamics while focusing on the flow of compressed air. A simulation study revealed that a spray gun nozzle comprising a flow splitting plate located upstream of the arc point in the nozzle produces compression waves whereby the flow field made in the nozzle differs substantially from that made in a conventional, plate-less nozzle. Observation using a high-speed camera showed that smaller particles of the molten metal (globules) were made due to the plate, which means that the compression waves generated upstream of the arc point affect the formation of globules at the arc point.
Ryoji Tamaki and Masashi Yamakawa
159 nSharma: Numerical Simulation Heterogeneity Aware Runtime Manager for OpenFOAM [abstract]
Abstract: CFD simulations are a fundamental engineering application, implying huge workloads, often with dynamic behaviour due to runtime mesh refinement. Parallel processing over heterogeneous distributed memory clusters is often used to process such workloads. The execution of dynamic workloads over a set of heterogeneous resources leads to load imbalances that severely impacts execution time, when static uniform load distribution is used. This paper proposes applying dynamic, heterogeneity aware, load balancing techniques within CFD simulations. nSharma, a software package that fully integrates with OpenFOAM, is presented and assessed. Performance gains are demonstrated, achieved by reducing busy times standard deviation among resources, i.e. heterogeneous computing resources are kept busy with useful work due to an effective workload distribution. To best of authors' knowledge, nSharma is the first implementation and integration of heterogeneity aware load balancing in OpenFOAM and will be made publicly available in order to foster its adoption by the large community of OpenFOAM users.
Roberto Ribeiro, Luís Paulo Santos and João Miguel Nóbrega
213 High Performance Computational Hydrodynamic Simulations: UPC Parallel Architecture as a Future Alternative [abstract]
Abstract: Developments in high-performance computing (HPC) has today transformed the manner of how computational hydrodynamic (CHD) simulations are performed. Till now, the message passing interface (MPI) remains the common parallelism architecture and has been adopted widely in CHD simulations. However, its bottleneck problem remains for some large-scale cases due to delays in message passing whereby the total communication time may exceed the total simulation runtime with an increasing number of processes. In this study, we utilise an alter-native parallelism architecture, known as PGAS-UPC, to develop our own UPC-CHD model with a 2-step explicit scheme from Lax-Wendroff family of predictors-correctors. The model is evaluated on three incompressible, adiabatic viscous 2D flow cases having moderate flow velocities. Model validation is achieved by the good agreement between the predicted and respective analytical values. We then compare the computational performance between UPC-CHD and that of MPI in its base design in an SGI UV-2000 server with 100 cores. The former achieves a near 1:1 speedup which demonstrates its efficiency potential for very large-scale CHD simulations, while the later experiences bottleneck at some point. Extension of UPC-CHD remains our main objective which can be achieved by the following additions: (a) inclusions of other numerical schemes to accommodate for varying flow conditions, and (b) coupling UPC-CHD with Amazon Web Service (AWS) to further exploit its parallelism efficiency as the viable alternative.
Alvin Wei Ze Chew, Tung Thanh Vu and Adrian Wing-Keung Law
373 On Parametric Excitation for Exploration of Lava Tubes and Caves [abstract]
Abstract: Huge lava tubes with an approximate diameter of 65-225m were found on the surfaces of Moon and Mars in the late 2000's. It has been argued that the interior of the caves are spacious, and are suitable to build artificial bases with habitable features such as constant temperature, as well as protection from both meteorites and harmful radiation. In line of the above, a number of studies which regard the soft landing mechanisms on the bottom of the lava tubes have been proposed. In this paper, aiming to extend the ability to explore arbitrary surface caves, we propose a mechanism which is able to reach the ceiling of lava tubes. The basic concept of our proposed mechanism consists of a rover connected to an oscillating sample-gatherer, wherein the rover is able to adjust the length of the rope parametrically to increase the deflection angle by considering periodic changes in the pivot, and thus to ease the collection of samples by hitting against the ceiling of the cave. Relevant simulations confirmed our theoretical observations which predict the increase of deflection angle by periodically winding and rewinding the rope according to pivotal variations. We believe our proposed approach brings the building blocks to enable finer control of exploration mechanisms of lava tubes and narrow environments.
Victor Parque, Masato Kumai, Satoshi Miura and Miyashita Tomoyuki

ICCS 2018 Main Track (MT) Session 9

Time and Date: 13:15 - 14:55 on 12th June 2018

Room: M2

Chair: Denis Nasonov

129 Global Simulation of Planetary Rings on Sunway TaihuLight [abstract]
Abstract: In this paper, we report the implementation and measured performance of global simulation of planetary rings on Sunway TaihuLight. The basic algorithm is the Barnes-Hut tree, but we have made a number of changes to achieve good performance for extremely large simulations on machines with extremely large number of cores. The measured performance is around 35% of the theoretical peak. The main limitation comes from the performance of the interaction calculation kernel itself, which is s currently around 50%.
Masaki Iwasawa, Long Wang, Keigo Nitadori, Daisuke Namekata, Miyuki Tsubouchi, Junichiro Makino, Zhao Liu, Haohuan Fu, Guangwen Yang and Takayuki Muranushi
350 Parallel Performance Analysis of Bacterial Biofilm Simulation Models [abstract]
Abstract: Modelling and simulation of bacterial biofilms is a computationally expensive process necessitating use of parallel computing. Fluid dynamics and advection-consumption models can be decoupled and solved to handle the fluid-solute-bacterial interactions. Data exchange between the two processes add up to the communication overheads. The heterogenous distribution of bacteria within the simulation domain further leads to non-uniform load distribution in the parallel system. We study the effect of load imbalance and communication overheads on the overall performance of simulation at different stages of biofilm growth. We develop a model to optimize the parallelization procedure for computing the growth dynamics of bacterial biofilms.
Sheraton M V and Peter M.A. Sloot
240 RT-DBSCAN: Real-time Parallel Clustering of Spatio-Temporal Data using Spark-Streaming [abstract]
Abstract: Clustering algorithms are essential for many big data applica- tions involving point-based data, e.g. user generated social media data from platforms such as Twitter. One of the most common approaches for clustering is DBSCAN. However, DBSCAN has numerous limitations. The algorithm itself is based on traversing the whole dataset and identi- fying the neighbours around each point. This approach is not suitable when data is created and streamed in real-time however. Instead a more dynamic approach is required. This paper presents a new approach, RT- DBSCAN, that supports real-time clustering of data based on continuous cluster checkpointing. This approach overcomes many of the issues of existing clustering algorithms such as DBSCAN. The platform is real- ised using Apache Spark running over large-scale Cloud resources and container based technologies to support scaling. We benchmark the work using streamed social media content (Twitter) and show the advant- ages in performance and flexibility of RT-DBSCAN over other clustering approaches.
Yikai Gong, Richard Sinnott and Paul Rimba
137 GPU-based implementation of Ptycho-ADMM for high performance X-ray imaging [abstract]
Abstract: X-ray imaging allows biologists to retrieve the atomic arrangement of proteins and doctors the capability to view broken bones in full detail. In this context, ptychography has risen as a reference imaging technique. It provides resolutions of one billionth of a meter, macroscopic field of view, or the capability to retrieve chemical or magnetic contrast, among other features. The goal is to reconstruct a 2D visualization of a sample from a collection of diffraction patterns generated from the interaction of a light source with the sample. The data collected is typically two orders of magnitude bigger than the final image reconstructed, so high performance solutions are normally desired. One of the latest advances in ptychography imaging is the development of Ptycho-ADMM, a new ptychography reconstruction algorithm based on the Alternating Direction Method of Multipliers (ADMM). Ptycho-ADMM provides faster convergence speed and better quality reconstructions, all while being more resilient to noise in comparison with state-of-the-art methods. The downside of Ptycho-ADMM is that it requires additional computation and a larger memory footprint compared to simpler solutions. In this paper we tackle the computational requirements of Ptycho-ADMM, and design the first high performance multi-GPU solution of the method. We analyze and exploit the parallelism of Ptycho-ADMM to make use of multiple GPU devices. The proposed implementation achieves reconstruction times comparable to other GPU-accelerated high performance solutions, while providing the enhanced reconstruction quality of the Ptycho-ADMM method.
Pablo Enfedaque, Stefano Marchesini, Hari Krishnan and Huibin Chang

ICCS 2018 Main Track (MT) Session 10

Time and Date: 15:25 - 17:05 on 12th June 2018

Room: M2

Chair: Pablo Enfedaque

216 Elastic CPU Cap Mechanism for Timely Dataflow Applications [abstract]
Abstract: Sudden surges in the incoming workload can cause adverse consequences on the run-time performance of data-flow applications. Our work addresses the problem of limiting CPU associated with the elastic scaling of timely data-flow (TDF) applications running in a shared computing environment while each application can possess a different quality of service (QoS) requirement. The key argument here is that an unwise consolidation decision to dynamically scale up/out the computing re- sources for responding to unexpected workload changes can degrade the performance of some (if not all) collocated applications due to their fierce competition getting the shared resources (such as the last level cache). The proposed solution uses a queue-based model to predict the performance degradation of running data-flow applications together. The problem of CPU cap adjustment is addressed as an optimization problem, where the aim is to reduce the quality of service violation incidents among applications while raising the CPU utilization level of server nodes as well as preventing the formation of bottlenecks due to the fierce competition among collocated applications. The controller uses and efficient dynamic method to find a solution at each round of the controlling epoch. The performance evaluation is carried out by comparing the proposed controller against an enhanced QoS-aware version of round robin strategy which is deployed in many commercial packages. Experimental results confirmed that the proposed solution improves QoS satisfaction by near to 148% on average while it can reduce the latency of processing data records for applications in the highest QoS classes by near to 19% during workload surges.
M. Reza Hoseinyfarahabady, Nazanin Farhangsadr, Albert Zomaya, Zahir Tari and Samee Khan
351 Blockchain-based transaction integrity in distributed big data marketplace [abstract]
Abstract: Today Big Data occupies crucial part as in scientific research areas as in large companies business analysis. Each company tries to find the best way how generated big data can be made valuable and profitable. However, in most cases, companies have not enough opportunities and budget to solve this complex problem. On the other hand, there are companies (i.e., in insurance, banking) who can significantly improve their business organization by applying hidden knowledge extracted from such big data. This situation leads to the necessity of building a platform for the exchange, processing, and sale of collected big data. In this paper, we propose a distributed big data platform that implements digital data market, based on the blockchain mechanism for data transaction integrity
Denis Nasonov, Alexander Visheratin and Alexander Boukhanovsky
363 Workload Characterization and Evolutionary Analyses of Tianhe-1A Supercomputer [abstract]
Abstract: Currently, supercomputer systems face a variety of application challenges, includ-ing high-throughput, data-intensive, and stream-processing applications. At the same time, there is more challenge to improve user satisfaction at the supercom-puters such as Tianhe-1A, Tianhe-2 and TaihuLight, because of the commercial service model. It is important to understand HPC workloads and their evolution to facilitate informed future research and improve user satisfaction. In this paper, we present a methodology to characterize workloads on the commercial supercomputer (users need to pay), at a particular period and its evo-lution over time. We apply this method to the workloads of Tianhe-1A at the Na-tional Supercomputer Center in Tianjin. This paper presents the concept of quota-constrained waiting time for the first time, which has significance for optimizing scheduling and enhancing user satisfaction on the commercial supercomputer.
Jinghua Feng, Guangming Liu, Jian Zhang, Zhiwei Zhang, Jie Yu and Zhaoning Zhang
378 The Design of Fast and Energy-Efficient Linear Solvers: On The potential Of Half Precision Arithmetic And Iterative Refinement Techniques [abstract]
Abstract: As parallel computers approach the exascale, power efficiency in High-performance computing (HPC) systems is of increasing concern. Exploiting both, the hardware features, and algorithms is an effective solution to achieve power efficiency, and address the energy constraints in modern and future HPC systems. In this work, we present a novel design and implementation of an energy efficient solution for dense linear systems of equations, which are at the heart of large-scale HPC applications. Energy efficient linear system solvers are based on two main components: (1) iterative refinement techniques, and (2) reduced precision computing features in the modern accelerators and co-processors. While most of the energy efficiency approaches aim to reduce the consumption with a minimal performance penalty, our method improves both, the performance and the energy-efficiency. Compared to highly optimised linear system solvers, our kernels are up to 2X faster to deliver the same accuracy solution, and reduce the energy consumption up to half on Intel KNL architectures. By using efficiently the tensor cores available in the NVIDIA V100 PCIe GPUs, the speedups can be up to 4X with more than 80\% reduction on the energy consumption.
Azzam Haidar, Ahmad Abdelfattah, Mawussi Zounon, Panruo Wu, Srikara Pranesh, Stanimire Tomov and Jack Dongarra
386 Design of Parallel BEM Analyses Framework for SIMD Processors [abstract]
Abstract: A software framework titled BEM-BB has been developed to conduct parallel boundary element method (BEM) analyses. By Imple- menting a fundamental solution or a Green’s function that is the most important element of the BEM, and it depends on the targeted physical phenomenon, the users get the benefit of MPI and OpenMP hybrid par- allelization with H-matrix approximation provided by the framework. However, the framework does not take into account the single instruc- tion multiple data SIMD vectorization, which is important for high- performance computing and is supported by majority of the existing processors. Dealing with SIMD vectorization of a user-defined function is difficult because SIMD exploits instruction-level parallelization and is closely associated with the user-defined function. This study describes the conceptual framework for enhancing the SIMD vectorization. The new framework was evaluated using the two BEM problems of static electric field analysis with a perfect conductor and dielectric on an Intel Broadwell processor and an Intel Xeon Phi KNL. We observed that the framework provides good vectorization with limited SIMD knowledge. The numerical results illustrate the improved performance of the frame- work. In particular, perfect conductor analyses using H-matrix achieved performance improvements of 2.22x and 4.33x as compared with that achieved using the original BEM-BB framework for Broadwell processor and KNL, respectively.
Tetsuya Hoshino, Akihiro Ida, Toshihiro Hanawa and Kengo Nakajima

ICCS 2018 Main Track (MT) Session 11

Time and Date: 9:00 - 10:40 on 13th June 2018

Room: M2

Chair: Paul Manstetten

64 An experimental assessment of three point-insertion sequences for 3-D incremental Delaunay tessellations [abstract]
Abstract: Currently, state-of-the-art algorithms for generating 3-D Delaunay tessellations are incremental. Thus, their execution costs depend on the order of point insertion. This work evaluates three point-insertion sequences in incremental algorithms for generating 3-D Delaunay tessellations. Specifically, an incremental algorithm with point-insertion order given by the cut-longest-edge kd-tree is evaluated against the BRIO-Hilbert order in conjunction with spatial middle and median policies employed in the 4.11 version of the Computational Geometry Algorithms Library. The results of execution and storage costs of these three algorithms are analyzed experimentally. It follows that the incremental algorithm with a point-insertion sequence in the order given by the BRIO-Hilbert order with spatial middle policy employed in the latest version of the Computational Geometry Algorithms Library shows lower execution and storage costs than the two other algorithms evaluated.
Sanderson L. Gonzaga de Oliveira, Diogo T. Robaina, Diego N. Brandão, Mauricio Kischinhevsky and G. Oliveira
92 Learning Knowledge Graph Embeddings via Generalized Hyperplanes [abstract]
Abstract: For knowledge graph completion, translation-based methods such as Trans(E and H) are promising, which embed knowledge graphs into continuous vector spaces and construct translation operation between head and tail entities. However, TransE and TransH still have limitations in preserving mapping properties of complex relation facts for knowledge graphs. In this paper, we propose a novel translation-based method called translation on generalized hyperplanes (TransGH), which extends TransH by defining a generalized hyperplane for entities projection. TransGH projects head and tail embeddings from a triplet into a generalized relation-specific hyperplane determined by a set of basis vectors, and then fulfills translation operation on the hyperplane. Compared with TransH, TransGH can capture more fertile interactions between entities and relations, and simultaneously has strong expression in mapping properties for knowledge graphs. Experimental results on two tasks, link prediction and triplet classification, show that TransGH can significantly outperform the state-of-the-art embedding methods.
Qiannan Zhu, Xiaofei Zhou, Jianlong Tan, Ping Liu and Li Guo
171 Fast Higher-Order Functions for Tensor Calculus with Tensors and Subtensors [abstract]
Abstract: Tensors analysis has become a popular tool for solving problems in computational neuroscience, pattern recognition and signal processing. Similar to the two-dimensional case, algorithms for multidimensional data consist of basic operations accessing only a subset of tensor data. With multiple offsets and step sizes, basic operations for subtensors require sophisticated implementations even for entrywise operations. In this work, we discuss the design and implementation of optimized higher-order functions that operate entrywise on tensors and subtensors with any non-hierarchical storage format and arbitrary number of dimensions. We propose recursive multi-index algorithms with reduced index computations and additional optimization techniques such as function inlining with partial template specialization. We show that single-index implementations of higher-order functions with subtensors introduce a runtime penalty of an order of magnitude than the recursive and iterative multi-index versions. Including data- and thread-level parallelization, our optimized implementations reach 68% of the maximum throughput of an Intel Core i9-7900X. In comparison with other libraries, the average speedup of our optimized implementations is up to 5x for map-like and more than 9x for reduce-like operations. For symmetric tensors we measured an average speedup of up to 4x.
Cem Bassoy and Volker Schatz
215 The t-modified self-shrinking generator [abstract]
Abstract: Pseudo-random sequences exhibit interesting properties with application in many and distinct areas ranging from reliable communications to number generation or cryptography. Inside the family of decimation-based sequence generators, the modified self-shrinking generator (an improved version of the self-shrinking generator) is one of its best-known elements. In fact, such a generator divides the PN-sequence produced by a maximum-length LFSR into groups of three bits. When the sum of the first two bits in a group is one, then the generator returns the third bit, otherwise the bit is discarded. In this work, we introduce a generalization of this generator, where the PN-sequence is divided into groups of t bits, t ≥ 3. It is possible to check that the properties of the output sequences produced by this family of generators have the same or better properties than those of the classic modified self-shrunken sequences. Moreover, the number of sequences generated by this new family with application in stream cipher cryptography increases dramatically.
Sara D. Cardell and Amparo Fúster-Sabater
109 Simulating Negotiation-based Cloud Markets [abstract]
Abstract: Today, the so called supermarket approach is used for trading Cloud services on Cloud markets. Thereby, consumers purchase Cloud services at fixed prices without negotiation. More dynamic Cloud markets are emerging as e.g. the recent development of the Amazon EC2 spot market shows - with spot blocks and spot fleet management. Hence, autonomous Bazaar-based negotiations are a promising approach for trading Cloud services on future Cloud markets. Thereby, market participants negotiate the characteristics of Cloud services which are described in Service Level Agreements (SLAs). Specifications such as the WS-Agreement Negotiation standard foster the development of such Bazaar-based Cloud markets. In this paper we present a scientific simulation environment for the simulation of Bazaar-based Cloud markets which conforms to the WS-Agreement Negotiation standard. A two-stepped process is required for using the simulation environment: first consumers, intermediaries and providers have to be created, then strategies have to be assigned to them before the result of the simulation can be analyzed. The aim of the simulation environment is to support market participants during the evaluation of their negotiation strategies.
Benedikt Pittl, Werner Mach and Erich Schikuta

ICCS 2018 Main Track (MT) Session 12

Time and Date: 11:10 - 12:50 on 13th June 2018

Room: M2

Chair: Amparo Fúster-Sabater

76 Structural Learning of Probabilistic Graphical Models of Cumulative Phenomena [abstract]
Abstract: One of the critical issues when adopting Bayesian networks (BNs) to model dependencies among random variables is to “learn” their structure. This is a well-known NP-hard problem in its most general and classical formulation, which is furthermore complicated by known pitfalls such as the issue of I-equivalence among different structures. In this work we restrict the investigation to a specific class of networks, i.e., those representing the dynamics of phenomena characterized by the monotonic accumulation of events. Such phenomena allow to set specific structural constraints based on Suppes’ theory of probabilistic causation and, accordingly, to define constrained BNs, named Suppes-Bayes Causal Networks (SBCNs). Within this framework, we study the structure learning of SBCNs via extensive simulations with various state-of-the-art search strategies, such as canonical local search techniques and Genetic Algorithms. This investigation is intended to be an extension and an in-depth clarification of our previous works on SBCN structure learning. Among the main results, we show that Suppes’ constraints do simplify the learning task, by reducing the solution search space and providing a temporal ordering on the variables, which simplifies the complications derived by I-equivalent structures. Finally, we report on tradeoffs among different optimization techniques that can be used to learn SBCNs.
Daniele Ramazzotti, Marco Nobile, Marco Antoniotti and Alex Graudenzi
81 Sparse Surface Speed Evaluation on a Dynamic Three-Dimensional Surface Using an Iterative Partitioning Scheme [abstract]
Abstract: We focus on a surface evolution problem where the local surface speed depends on a computationally expensive scalar function with non-local properties. The local surface speed must be re-evaluated in each time step, even for non-moving parts of the surface, due to possibly changed properties in remote regions of the simulation domain. We present a method to evaluate the surface speed only on a sparse set of points to reduce the computational effort. This sparse set of points is generated according to application-specific requirements using an iterative partitioning scheme. We diffuse the result of a constant extrapolation in the neighborhood of the sparse points to obtain an approximation to a linear interpolation between the sparse points. We demonstrate the method for a surface evolving with a local surface speed depending on the incident flux from a source plane above the surface. The obtained speedups range from 2 to 8 and the surface deviation is less than 3 grid-cells for all evaluated test cases.
Paul Manstetten, Lukas Gnam, Andreas Hössinger, Siegfried Selberherr and Josef Weinbub
219 Accurate, Automatic and Compressed Visualization of Radiated Helmholtz Fields from Boundary Element Solutions [abstract]
Abstract: We propose a methodology to generate an accurate and efficient reconstruction of radiated fields based on high order interpolation. As the solution is obtained with the convolution by a smooth but potentially high frequency oscillatory kernel, our basis functions therefore incorporate plane waves. Directional interpolation is shown to be efficient for smart directions. An adaptive subdivision of the domain is established to limit the oscillations of the kernel in each element. The new basis functions, combining high order polynomials and plane waves, provide much better accuracy than low order ones. Finally, as standard visualization softwares are generally unable to represent such fields, a method to have a well-suited visualization of high order functions is used. Several numerical results confirm the potential of the method.
Matthieu Maunoury, Christophe Besse, Vincent Mouysset and Sébastien Pernet
11 The Aero-structural Optimization using the Modular Analysis and Unified Derivatives (MAUD) applied for the Wing Design [abstract]
Abstract: In the paper we present a case study of the aero-structural analysis and optimization for the wing design using the modular analysis and unified derivatives (MAUD). Wing design is one of the essential procedures of aircraft manufactures and it is a compromise between many competing factors and constraints. As a result, the efficient numerical optimized methods are important to speed-up the design procedures special for the the design parameter of order~$\cal O$(10-100). In the aero-structural optimization, the derivatives can be calculated by simply applying the finite-difference methods. However, the finite difference methods are in general significantly more expensive, requiring at least one additional flow solution per parameter. By using the method of modular analysis and unified derivatives (MAUD), we can unify all methods for computing total derivatives using a single equation. The derivatives can be automatically uniform calculated [1]as \frac{\partial R}{\partial u} \frac{du}{dr} =\Psi = \frac{\partial R^T}{\partial u}\frac{du^T}{dr} The wing design requires a set of benchmark cases for the shape optimization. In the paper, we choice the Common Research Model (CRM) geometry which was developed by NASA Langely Research Center and Amers Research Center for applied CFD validation studies. In this paper we only focus on preliminary designs on the static aeroelastic analysis for example the aileron reversal analysis [2]. We use the open source code OpenMDAO [3] as the framework for the implementation of MAUD applied for the wing design. OpenMDAO provides a library of sparse solvers and optimizers designed to work with its distributed-memory, sparse data-passing scheme. We present performance comparisons for a typical production problem and a summary of the experiences gained. References: [1] J. R. R. A. Martins and J. T. Hwang, Review and Unification of Methods for Computing Derivatives of Multidisciplinary Computational Models, AIAA Journal, 51(1):2582--2599, 2013 [2] Mengmeng Zhang, Contributions to Variable Fidelity MDO Framework for Collaborative and Integrated Aircraft Design, Doctoral Thesis, Royal Institute of Technology KTH, Stockholm, Sweden, 2015. [3] OpenMDAO, http://openmdao.org/
Mengmeng Zhang, Xin Shi, Lilit Axner and Jing Gong