Simon Shaolei Du

Office: GHC 8005

Email: ssdu [at] cs (dot) cmu (dot) edu

Social Media: LinkedIn Facebook 知乎 WeChat

I am a second year PhD student in Machine Learning Department at Carnegie Mellon University and my advisors are Aarti Singh and Barnabás Póczos. My research interest is theoretical machine learning. Specifically, I am trying to provide theoretical foundation for transfer learning, matrix approximation and quantization for machine learning tasks. On the application side, I am interested in applying machine learning techniques for precision agriculture. During the summer of 2016, I worked with Jianshu Chen, Lihong Li, Lin Xiao and Dengyong Zhou on policy evaluation problems of reinforcement learning and saddle point problems of convex optimization at Microsoft Research.

Previously, I studied EECS and EMS at UC Berkeley (go bears!) and I worked with Ming Gu, Lei Li, Michael Mahoney and Stuart Russell on various matrix-related problems. I also spent a semester at Department of Electronic Engineering of Tsinghua University.

I was born in Sydney and grew up in Beijing. I had my wonderful 6 years at SDSZ.




  1. Stochastic Variance Reduction Methods for Policy Evaluation,
    Simon S. Du, Jianshu Chen, Lihong Li, Lin Xiao, Dengyong Zhou,
    [PDF] [Arxiv] [Lihong's Talk at Simons Institute]
  2. Computationally Efficient Robust Estimation of Sparse Functionals,
    Simon S. Du, Sivaraman Balakrishnan, Aarti Singh,
    [PDF] [Arxiv]
  3. On the Power of Truncated SVD for General High-rank Matrix Estimation Problems,
    Simon S. Du, Yining Wang, Aarti Singh,
    [PDF] [Arxiv]
  4. Hypothesis Transfer Learning via Transformation Functions,
    Simon S. Du, Jayanth Koushik, Aarti Singh, Barnabás Póczos,
    [PDF] [Arxiv] [Poster]
  5. Efficient Nonparametric Smoothness Estimation,
    Shashank Singh, Simon S. Du, Barnabás Póczos,
    In Conference on Neural Information Processing Systems (NIPS) 2016.
    [PDF] [Arxiv]
  6. Novel Quantization Strategies for Linear Prediction with Guarantees,
    Simon S. Du**, Yichong Xu**, Yuan Li, Hongyang Zhang, Aarti Singh, Pulkit Grover,
    In International Conference of Machine Learning (ICML) 2016, On Device Intelligence (ONDI) workshop.
    [PDF] [Slides]
  7. An Improved Gap-Dependency Analysis of the Noisy Power Method,
    Maria-Florina Balcan*, Simon S. Du*, Yining Wang*, Adams Wei Yu*,
    In Conference of Learning Theory (COLT) 2016.
    [PDF] [Arxiv] [Slides] [Talk]
  8. Spectral Gap Error Bounds for Improving CUR Matrix Decomposition and the Nystrom Method,
    David G. Anderson*, Simon S. Du*, Michael W. Mahoney*, Christopher Melgaard*, Kunming Wu*, Ming Gu,
    In International Conference on Artificial Intelligence and Statistics (AISTATS) 2015.
    [PDF] [Supplement] [Code]
  9. Maxios: Large Scale Nonnegative Matrix Factorization for Collaborative Filtering,
    Simon S. Du, Yilin Liu, Boyi Chen, Lei Li,
    In Conference on Neural Information Processing Systems (NIPS) 2014, workshop on Distributed Machine Learning and Matrix Computations.
    [PDF] [Poster]

  10. **: Equal contribution. *: Alphabetic order according to mathematics or theoretical computer science convention.