Listing 1 - 10 of 39 | << page >> |
Sort by
|
Choose an application
Silkworms. --- Mulberry.
Choose an application
Choose an application
Academic collection --- #BIBC:T1997 --- Theses
Choose an application
Choose an application
Choose an application
Choose an application
Choose an application
Data is everywhere, but it is only useful if information can be extracted from it. To this end, data analysis techniques have been developed to find features underlying the data. One class of data analysis techniques uses tensor decompositions. Tensors are higher-dimensional extensions of vectors and matrices and allow us to represent multiway data in a natural way. One of the key strengths of tensors is that their decompositions are unique under mild conditions without imposing additional constraints, unlike matrix decompositions. This property opens up many interesting applications of tensor decompositions, in particular with respect to blind source separation and independent component analysis.Independent component analysis (ICA) tries to find the statistically independent signals underlying a mixture, which is useful in many fields such as telecommunications, speech separation and biomedical data analysis. These mixtures are often modeled as instantaneous mixtures of source signals and tensor methods that can blindly separate such mixtures are well known. However, in many applications a convolutive mixture model is more appropriate to take delays and reflections of the source signals into account. This is for instance the case for speech signals in a room or telecommunications signals impinging on antennas. When the mixtures are convolutive, additional structure arises in the separation problem. Current tensor-based methods in the literature do not fully exploit this Toeplitz or Hankel structure. This thesis presents new subspace-based methods that take more of this structure into account, which leads to efficient and more accurate results than the current state-of-the-art tensor methods. Additionally, relaxed uniqueness bounds are formulated that exploit the available structure as well. This thesis also presents a method for a related structured tensor decomposition in which all factor matrices have block-circulant structure.Apart from exploiting the structure that appears in convolutive ICA, this thesis also presents how known techniques in instantaneous ICA can be ported to the convolutive case. This includes coupled tensor decompositions to combine second- and fourth-order statistics, and the usage of incomplete tensors to reduce the computational complexity.Another interesting question unrelated to convolutive ICA is whether we can compare underlying tensor factors without having to compute their full decompositions. This is highly relevant for any tensor classification problem but has only received limited attention in research. In this thesis, we present foundational theorems and algorithms that show how tensor factors can be compared in two modes for several underlying tensor decompositions. One of these methods is used further to develop an algorithm that is able to compare the cluster centers in two different datasets, without having to compute the actual clusters.
Choose an application
Various real-life data such as time series and multi-sensor recordings can be represented by vectors and matrices, which are one-way and two-way arrays of numerical values, respectively. Valuable information can be extracted from these measured data matrices by means of matrix factorizations in a broad range of applications within signal processing, data mining, and machine learning. While matrix-based methods are powerful and well-known tools for various applications, they are limited to single-mode variations, making them ill-suited to tackle multi-way data without loss of information. Higher-order tensors are a natural extension of vectors (first order) and matrices (second-order), enabling us to represent multi-way arrays of numerical values, which have become ubiquitous in signal processing and data mining applications. By leveraging the powerful utitilies offered by tensor decompositions such as compression and uniqueness properties, we can extract more information from multi-way data than what is possible by using only matrix tools.While higher-order tensors allow us to properly accommodate for multiple modes of variation in data, tensor problems are often large-scale because the number of entries in a tensor increases exponentially with the tensor order. This curse of dimensionality can, however, be alleviated or even broken by various techniques such as representing the tensor by an approximate but compact tensor model. While a pessimist only sees the curse, an optimist sees a significant opportunity for the compact representation of large-scale data vectors: by representing a large-scale vector (first order) using a compact (higher-order) tensor model, the number of parameters needed to represent the underlying vector decreases exponentially in the order of the tensor representation. The key assumption to employ this blessing of dimensionality is that the data can be described by much fewer parameters than the actual number of samples, which is often true in large-scale applications.By leveraging the blessing of dimensionality in this thesis for blind source separation and (blind) system identification, we can tackle large-scale applications through explicit and implicit tensor decomposition-based methods. While explicit decompositions decompose a tensor that is known a priori, implicit decompositions decompose a tensor that is only known implicitly. In this thesis, we present a single-step framework for a particular type of implicit tensor decomposition, consisting of optimization-based and algebraic algorithms as well as generic uniqueness results. By properly exploiting additional structure in specific applications, we can significantly reduce the computational complexity of our optimization-based method. Our approach for large-scale instantaneous blind source separation and (blind) system identification enables various applications such as direction-of-arrival estimation in large-scale arrays and neural spike sorting in high-density recordings. Furthermore, we link implicit tensor decompositions to multilinear systems of equations, which are a generalization of linear systems, allowing us to propose a novel tensor-based classification scheme that we use for face recognition and irregular heartbeat classification with excellent performance.
Choose an application
Over the years, many algorithms have been designed to decompose a matrix into a product of other matrices. These matrix decompositions can be used to compress data with a minimal loss of information or for extracting meaningful components. More recently, tensor decompositions such as the canonicalpolyadic decomposition (CPD) and the low multilinear rank approximation (LMLRA) have been designed as higher-order generalizations of these matrix decompositions leading to better results in certain applications. While matrix methods offer a natural way to process data involving two variables, theyoften fail to recognize higher-order patterns. Tensors, in contrast, can have any number of modes and can preserve the higher-order structure of the data, which makes them highly useful to analyze multi-variable data. Second, tensor decompositions are often unique in cases where their matrix counterparts arenot. As fewer (artificial) restrictions such as orthogonality or triangle structure have to be imposed on the terms to enforce their uniqueness, this can lead to a decomposition in more meaningful terms.Current tensor decomposition methods commonly make two assumptions about the data supplied by the user. First, they assume that all data is available at once and second, they consider noise on the data to be additive and Gaussian distributed. In many applications, these assumptions are justified and one can decompose the tensor using standard batch methods that minimize the least-squares distance between the tensor and the low-rank model. However, as tensor methods are applied to more and more real-world problems, the number of cases where these assumptions are clearly violated increases as well. Any real-time application, for instance, violates the assumption of having all data available at the start. In such applications, such as process control or patient monitoring, new data arrive at certain time intervals. This data should be included in the tensor model without recomputing the full decomposition from scratch, as such approaches are not time- nor resource-efficient. Similarly, it is not hard to find applications where the assumption of additive Gaussian noise is not satisfied. Audio data, for instance, is generally modelled using a non-Gaussian noise distribution while low-light imaging assumes Poisson distributed pixel intensities. When the real and imaginary parts of a signal are collected independently, as is typically the case in MRI imaging, the noise distribution of the signal magnitude is Rician instead of Gaussian. For these applications, fitting a low-rank model to the tensor with the least-squares distance as cost function will lead to suboptimal solutions while a suitable choice of an alternative cost function can provide a model that corresponds better to the underlying components.In this thesis, we provide methods that can still efficiently compute a suitable tensor decomposition when either of the aforementioned assumptions is not satisfied. First, we propose updating methods for both the CPD and the LMLRA. These methods start from an existing tensor decomposition and efficiently update this decomposition when new data arrives. By exploiting the multilinear structure of the tensor models, these methods are both efficient and accurate in tracking low-rank representations of the data. When desired, one can even use a weighted least-squares approach to weight certain tensor entries more heavily than others, e.g., when the uncertainty on the collected tensor entries varies between sensors. The performance of the CPD method is demonstrated for the monitoring of brain hemodynamics coupling in neonates. A second contribution is that we extend the existing least-squares CPD algorithms to other cost functions. This makes it possible to use a dedicated cost function for the application at hand and to obtain better solutions than when standard least-squares approaches are used. We focus in particular on the class of beta-divergence cost functions, of which the least-squares distance and the Kullback—Leibler and Itakura—Saito divergences are special cases, but we also supply a dedicated method for the Rician cost function. Additionally, any twice-differentiable entry-wise cost function is supported. We demonstrate the improved effectiveness of the method over least-squares approaches by blindly separating a series of fMRI signals. Finally, we propose an improved algebraic algorithm to compute the CPD of a tensor in the least-squares sense. By recursively splitting the generalized eigenspace into successively smaller subspaces instead of determining all generalized eigenvalues at once, the decomposition is more accurate, yet can still be obtained efficiently. The algebraic solution can be used as an initialization for more costly optimization algorithms and significantly reduce the number of iterations and, hence, also the runtime of these methods.
Listing 1 - 10 of 39 | << page >> |
Sort by
|