====================================================================
                          CALL FOR PAPERS

       Low-rank Matrix Approximation for Large-scale Learning
              NIPS 2010 Workshop, Whistler, Canada
                        December 11, 2010

        http://www.eecs.berkeley.edu/~ameet/low-rank-nips10/

             Submission Deadline: October 31, 2010
====================================================================
Overview

Today's data-driven society is full of large-scale datasets. In the context of machine learning, these datasets are often represented by large matrices representing either a set of real-valued features for each point or pairwise similarities between points. Hence, modern learning problems in computer vision, natural language processing, computational biology, and other areas often face the daunting task of storing and operating on matrices with thousands to millions of entries. An attractive solution to this problem involves working with low-rank approximations of the original matrix. Low-rank approximation is at the core of widely used algorithms such as Principle Component Analysis, Multidimensional Scaling, Latent Semantic Indexing, and manifold learning. Furthermore, low-rank matrices appear in a wide variety of applications including lossy data compression, collaborative filtering, image processing, text analysis, matrix completion and metric learning.

The NIPS workshop on "Low-rank Matrix Approximation for Large-scale Learning" aims to survey recent work on matrix approximation with an emphasis on usefulness for practical large-scale machine learning problems. We aim to provide a forum for researchers to discuss several important questions associated with low-rank approximation techniques. The workshop will begin with an introductory talk and will include invited talks by Emmanuel Candes (Stanford), Ken Clarkson (IBM Almaden) and Petros Drineas (RPI). There will also be several contributed paper talks as well as poster session for contributed papers.

We encourage submissions exploring the impact of low-rank methods for large-scale machine learning in the form of new algorithms, theoretical advances and/or empirical results. We also welcome work on related topics that motivate additional interesting scenarios for use of low-rank approximations for learning tasks. Some specific areas of interest include randomized low-rank approximation techniques, the effect of data heterogeneity on randomization, performance of various low-rank methods for large-scale tasks and the tradeoff between numerical precision and time/space efficiency in the context of machine learning performance, e.g., classification or clustering accuracy.

Submission Guidelines

Submissions should be written as extended abstracts, no longer than 4 pages in the NIPS latex style. Style files and formatting instructions can be found at http://nips.cc/PaperInformation/StyleFiles. Submisssions must be in PDF format. Authors names and affiliations should be included, as the review process will not be double blind. The extended abstract may be accompanied by an unlimited appendix and other supplementary material, with the understanding that anything beyond 4 pages may be ignored by the program committee.

Please send your PDF submission by email to submit.lowrank@gmail.com by October 31. Notifications will be given on or before November 15. Topics that were recently published or presented elsewhere are allowed, provided that the extended abstract mentions this explicitly.

Organizers

Arthur Gretton (UCL-Gatsby), Michael Mahoney (Stanford), Mehryar Mohri (NYU, Google Research), Ameet Talwalkar (Berkeley)

Program Committee

Alexandre d'Aspremont (Princeton), Christos Boutsidis (Rensselear Polytechnic Institute), Kamilika Das (NASA Ames Research Center), Maryam Fazel (Washington), Michael I. Jordan (Berkeley), Sanjiv Kumar (Google Research), James Kwok (Hong Kong University of Science and Technology), Gunnar Martinsson (Colorado at Boulder)