Simon Shaolei Du

Office: GHC 8005

Email: ssdu [at] cs (dot) cmu (dot) edu

Social Media: LinkedIn Facebook 知乎 WeChat

I am a second year PhD student in Machine Learning Department at Carnegie Mellon University and my advisors are Aarti Singh and Barnabás Póczos. Previously, I studied EECS and EMS at UC Berkeley and I worked with Ming Gu, Lei Li, Michael Mahoney and Stuart Russell on various matrix-related problems. I also spent a semester at Department of Electronic Engineering at Tsinghua University. .

My research interests broadly include topics in theoretical machine learning and statistics, such as matrix factorization, convex/non-convex optimization, transfer learning, reinforcement learning, non-parametric statistics and robust statistics. On the application side, I am interested in applying machine learning techniques for precision agriculture.

I was born in Sydney and grew up in Beijing. I had my wonderful 6 years at SDSZ.


Professional Experiences



  1. Gradient Descent Can Take Exponential Time to Escape Saddle Points,
    Simon S. Du, Chi Jin, Jason D. Lee, Michael I. Jordan, Barnabas Poczos, Aarti Singh,
    [PDF] [Arxiv]
  2. On the Power of Truncated SVD for General High-rank Matrix Estimation Problems,
    Simon S. Du, Yining Wang, Aarti Singh,
    [PDF] [Arxiv]
  3. Hypothesis Transfer Learning via Transformation Functions,
    Simon S. Du, Jayanth Koushik, Aarti Singh, Barnabás Póczos,
    [PDF] [Arxiv] [Poster]
  4. Conference Papers

  5. Stochastic Variance Reduction Methods for Policy Evaluation,
    Simon S. Du, Jianshu Chen, Lihong Li, Lin Xiao, Dengyong Zhou,
    International Conference of Machine Learning (ICML) 2017.
    [PDF] [Arxiv] [Lihong's Talk at Simons Institute]
  6. Computationally Efficient Robust Estimation of Sparse Functionals,
    Simon S. Du, Sivaraman Balakrishnan, Aarti Singh,
    Conference of Learning Theory (COLT) 2017.
    [PDF] [Arxiv]
    Merged with this paper
  7. Efficient Nonparametric Smoothness Estimation,
    Shashank Singh, Simon S. Du, Barnabás Póczos,
    Conference on Neural Information Processing Systems (NIPS) 2016.
    [PDF] [Arxiv]
  8. An Improved Gap-Dependency Analysis of the Noisy Power Method,
    Maria-Florina Balcan*, Simon S. Du*, Yining Wang*, Adams Wei Yu*,
    Conference of Learning Theory (COLT) 2016.
    [PDF] [Arxiv] [Slides] [Talk]
  9. Spectral Gap Error Bounds for Improving CUR Matrix Decomposition and the Nystrom Method,
    David G. Anderson*, Simon S. Du*, Michael W. Mahoney*, Christopher Melgaard*, Kunming Wu*, Ming Gu,
    International Conference on Artificial Intelligence and Statistics (AISTATS) 2015.
    [PDF] [Supplement] [Code]
  10. Workshop Papers

  11. Novel Quantization Strategies for Linear Prediction with Guarantees,
    Simon S. Du**, Yichong Xu**, Yuan Li, Hongyang Zhang, Aarti Singh, Pulkit Grover,
    International Conference of Machine Learning (ICML) 2016, On Device Intelligence (ONDI) workshop.
    [PDF] [Slides]
  12. Maxios: Large Scale Nonnegative Matrix Factorization for Collaborative Filtering,
    Simon S. Du, Yilin Liu, Boyi Chen, Lei Li,
    Conference on Neural Information Processing Systems (NIPS) 2014, workshop on Distributed Machine Learning and Matrix Computations.
    [PDF] [Poster]

  13. **: Equal contribution. *: Alphabetic order according to mathematics or theoretical computer science convention.