Table of Contents

Research

Our research group's interests are in statistical signal processing, machine learning and optimization. Within these areas we work on a variety of fundamental problems (adaptive representations, classification, clustering, computational signal processing, and machine learning) and in a variety of application domains (fMRI analysis, image processing, video analysis and search).


Recent research projects have supported by grants from the National Institute of Mental Health, the National Science Foundation, INTEL, a Google Fellowship in Neuroscience, Princeton's University's Charlotte Elizabeth Procter honorific fellowship, and the Insley Blair Pyne Fund.


Example Research Projects

Publications referred to below are listed in detail under the publications tab above.

1) Analysis of Functional Magnetic Resonance Imaging Data

This work has been funded by grants from the National Institute of Mental Health (5R01MH075706-02), the National Science Foundation (NSF-1129764 and NSF-1129855), a grant from INTEL Labs, a Google Fellowship in Neuroscience, and the Insley Blair Pyne Fund

In collaboration with neuroscience colleagues, we are developing algorithms for functionally aligning the fMRI data of multiple subjects measured during a movie viewing. We have developed various methods, using distinct alignment metrics, for accomplishing this goal:


BrainIAK

2) Dictionary Based Classification and Structured Dictionaries

Learning a sparse representation of new data in terms of a dictionary is a state of the art method for modeling data. But when the dictionary is large and the data dimension is high, it can be a computationally challenging problem. We are exploring three aspects of the problem. First, we are developing new screening tests that quickly identify dictionary atoms that are guaranteed to have zero weights. This work is described in several of our recent papers:

Second, we have been exploring the application of these ideas to large scale real world datasets including music genre classification and large document collections. See for example:

Third, we are developing a hierarchical framework that uses incremental random projections to learn a structured dictionary for sparse representation in small stages. Initial experiment results show that our framework can learn informative hierarchical sparse representations in a more time efficient manner. See our paper at NIPS 2011 (Xiang et al).

3) Spatially Aware Learning

When applying off-the-shelf machine learning algorithms to data with spatial dimensions (images, geo-spatial data, fMRI, etc) a central question arises: how to incorporate prior information on the spatial characteristics of the data? For example, if we feed a boosting or SVM algorithm with individual image voxels as features, the voxel spatial information is ignored. Indeed, if we randomly shuffled the voxels, the algorithm would not notice any difference. Yet in many cases the spatial arrangement of the voxels together with prior information about expected spatial characteristics of the data may be very helpful. We are particularly interested in the situation when the trained classifier is used to identify relevant spatial regions. To make this more concrete, consider the problem of training a classifier to distinguish two different brain states based on fMRI responses. Successful classification suggests that the voxels used are important in discriminating between the two classes. Hence we could use a successful classifier to learn a set of discriminative voxels. We expect that these voxels will be spatially compact and clustered. How can this prior knowledge be incorporated into the training of the classifier? To learn more, please read our paper (Xiang et al) in NIPS 2009.


4) Statistical Image Processing

A variety of image and signal processing problems can be formulated as minimizing the sum ||f-g|| + R(f). The first term encodes the desire that f be near a given signal g and the second term is a regularization penalty on the complexity of f. So we want a good but simple approximation f to the given signal g. An challenge in formulating the term R(f) is ensuring that sharp edges and meaningful high frequency features in natural signals and images are not unduly penalized. Many traditional regularization methods fail to meet this challenge because the regularizer is agnostic to the edge structure of f. A modern solution selects R(f) to be the L1 norm of the coordinates of f in a wavelet basis. Despite the huge success and popularity of this approach, there is interest in improvements in the following two areas: (a) The use of nonlinear wavelets; (b) A more flexible image decomposition structure. This project is developing new signal representations and studying how these representations can compactly represent signals of interest. One application is edge aware signal denoising. To learn more please read our papers at ICASSP 2010 (Xiang & Ramadge) and ICIP 2010 (Xiang and Ramadge). The software for the algorithms in these papers is available on-line.