Applications of Matrix Computational Methods in the Analysis of Modern Data (MATRIX) Session 2

Time and Date: 14:10 - 15:50 on 3rd June 2015

Room: M209

Chair: Kouroush Modarresi

762 Anomaly Detection and Predictive Maintenance through Large Sparse Similarity Matrices and Causal Modeling [abstract]
Abstract: We use large (100k x 100k) sparse similarity matrices of time series (sensor data) for anomaly detection and failure prediction. These similarity matrices, computed using the universal information distance based on Kolmogorov Complexity, are used to perform non-parametric unsupervised clustering with non-linear boundaries without complex and slow coordinate transformations of the raw data. Changes over time in the similarity matrix allow us to observe anomalous behavior of the system and predict failure of parts. This approach is well suited for big data with little prior domain knowledge. Once we have learned the basic dependency patterns from the data, we can use this in addition to domain knowledge to build a causal model that relates outcomes to inputs through hidden variables. Given a set of observed outcomes and their associated sensor data, we can build a probabilistic model of the underlying causal events that produced both the outcomes and the data. The parameters of this probabilistic model (conditional joint probabilities) are inferred by maximizing the likelihood of the observed historical data. When such a model has been inferred, we can use it to predict future outcomes based on observations by first identifying the most likely underlying causal events that produced the observations and hence the most likely resulting outcomes. Both approaches differ from building a direct correlational model between the data and outcomes because they utilize complex representations of the state of the system -- in the first case through the similarity matrix and in the second case through domain specific modeling. As a final step, these predictions of future outcomes are fed into an over-arching stochastic optimization for optimal scheduling of maintenance activities over either short or long term time horizons. We’ll show real world examples from utilities, aerospace and defense and video surveillance. * It is a joint work with Alan McCord and Anand Murugan
Paul Hofmann
692 Computation of Recommender System using Localized Regularization [abstract]
Abstract: Targeting and Recommendation are major topics in ecommerce. The topic is treated as “Matrix Completion” in statistics. The main point is to compute the unknown (missing) values in the matrix data. This work is based on a different view of regularization, i.e., a localized regularization technique which leads to improvement in the estimation of the missing values.
Kourosh Modarresi
695 Unsupervised Feature Extraction using Singular Value Decomposition [abstract]
Abstract: Though modern data often provides a massive amount of information, much of that might be redundant or useless (noise). Thus, it is significant to recognize the most informative features of data. This will help the analysis of the data by removing the consequences of high dimensionality, in addition of obtaining other advantages of lower dimensional data such as lower computational cost and a less complex model.
Kourosh Modarresi
384 Quantifying complementarity among strategies for influencers' detection on Twitter [abstract]
Abstract: The so-called influencer, a person with the ability to persuade people, have important role on the information diffusion in social media environments. Indeed, influencers might dictate word-of-mouth and peer recommendation, impacting tasks such as recommendation, advertising, brand evaluation, among others. Thus, a growing number of works aim to identify influencers by exploiting distinct information. Deciding about the best strategy for each domain, however, is a complex task due to the lack of consensus among these works. This paper presents a quantitative study of analysis among some of the main strategies for identifying influencers, aiming to help researchers on this decision. Besides determining semantic classes of strategies, based on the characteristics they exploit, we obtained through PCA an effective meta-learning process to combine linearly distinct strategies. As main implications, we highlight a better understanding about the selected strategies and a novel manner to alleviate the difficulty on deciding which strategy researchers would adopt.
Alan Neves, Ramon Viera, Fernando Mourão, Leonardo Rocha
399 Fast Kernel Matrix Computation for Big Data Clustering [abstract]
Abstract: Kernel k-Means is a basis for many state of the art global clustering approaches. When the number of samples grows too big, however, it is extremely time consuming to compute the entire kernel matrix and it is impossible to store it in the memory of a single computer. The algorithm of Approximate Kernel k-Means has been proposed, which works using only a small part of the kernel matrix. The computation of the kernel matrix, even a part of it, remains a significant bottleneck of the process. Some types of kernel, however, can be computed using matrix multiplication. Modern CPU architectures and computational optimization methods allow for very fast matrix multiplication, thus those types of kernel matrices can be computed much faster than others.
Nikolaos Tsapanos, Anastasios Tefas, Nikolaos Nikolaidis, Alexandros Iosifidis, Ioannis Pitas