Date Range:

From: To:

View all

  • Graduate Student Center Graduate Student Center
  • General Public Presentations General Public Presentations
  • Thesis/Dissertation Seminars Thesis/Dissertation Seminars
  • Arts and Humanities Seminars Arts and Humanities Seminars
  • Education Seminars Education Seminars
  • Health Professions Seminars Health Professions Seminars
  • Professional/Business Seminars Professional/Business Seminars
  • Social Sciences Seminars Social Sciences Seminars
  • STEM* Seminars STEM* Seminars
  • Social Events Social Events
  • Student and Professional Development Student and Professional Development
  • Informational Events Informational Events
  • Important Dates Important Dates

*STEM: Science, Technology, Engineering, and Mathematics

International Community

Events Calendar   

Back to Summary

Thesis/Dissertation Seminars

Dissertation Defense: Sampling and Subspace Methods for Learning Sparse Group Structures in Computer Vision

HEC 438
March 23, 2018 @ 02:30 PM - 04:30 PM

Announcing the Final Examination of Maryam Jaberi for the degree of Doctor of Philosophy

The unprecedented growth of data in volume and dimension has led to an increased number of computationally-demanding and data-driven decision-making methods in many disciplines, such as computer vision, genomics, finance, etc. Research on big data aims to understand and describe trends in massive volumes of high-dimensional data. High volume and dimension are the determining factors in both computational and time complexity of algorithms. The challenge grows when the data are formed of the union of group-structures of different dimensions embedded in a high-dimensional ambient space. To address the problem of high volume, we propose a sampling method referred to as the Sparse Withdrawal of Inliers in a First Trial (SWIFT), which determines the smallest sample size in one grab so that all group-structures are adequately represented and discovered with high probability. The key features of SWIFT are: (i) sparsity, which is independent of the population size; (ii) no prior knowledge of the distribution of data, or the number of underlying group-structures; and (iii) robustness in the presence of an overwhelming number of outliers. We report a comprehensive study of the proposed sampling method in terms of accuracy, functionality, and effectiveness in reducing the computational cost in various applications of computer vision. In the second part of this dissertation, we study dimensionality reduction for multi-structural data. We propose a probabilistic subspace clustering method that unifies soft- and hard-clustering in a single framework. This is achieved by introducing a delayed association of uncertain points to subspaces of lower dimensions based on a confidence measure. Delayed association yields higher accuracy in clustering subspaces that have ambiguities, i.e. due to intersections and high-level of outliers/noise, and hence leads to more accurate self-representation of underlying subspaces. Altogether, this dissertation addresses the key theoretical and practically issues of size and dimension in big data analysis.

Committee in Charge: Hassan Foroosh (Chair), Marianna Pensky (Co-Chair), Boqing Gong, GuoJun Qi