Quantitative trading becomes one of the most comprehensive worlds that requires intensive algorithmic design, strong data science/engineering intelligence, a broad range of technical advancements, and beyond. It observed one of the greatest merges of science and engineering in the past decades. Our goal is to design and implement more powerful trading algorithms, signals, and strategies, through continued improvement in understanding of the data, theories of related disciplines, and engineering development.
Deep Learning
The deep neural networks have exceptional performance in a wide range of applications, but their theoretical understanding is still largely missing. Our goal is to provide fundamental justification for the theoretical properties of deep neural networks, including computational convergence, generalization, etc. We are also interested in studying new network architectures and optimization procedures that enable superior learning process than the state-of-the-art.
Nonconvex regularization and constraints are widely used in solving modern machine learning problems, such as in sparsity-inducing high-dimensional learning and data factorization. They can effectively reduce the estimation bias, and make the obtained estimator attain significantly better statistical performance in parameter estimation with favored computational properties than the convex counterpart. Our goal is to establish statistical guarantees with underlying statistical models, which can provide us new insights on developing computational guarantees and geometric properties for nonconvex optimization algorithms.
Modern data science and machine learning problems often exhibit in magnificent scales, leading to massive computational burden using traditional dimension reduction techniques. Our goal is to propose novel data sketching and inference approach that intelligently subsamples the original array in a judicious way to effectively reduce its overall size while preserving the latent geometric features and being robust to undesired outliers. Several orders of speedup is achieved in practice, such as in saliency detection. Extension to more challenging multi-way array data (or tensor) is also plausible, which allows extensively more modeling flexibility.
Beyond the theoretical investigation, we also make efforts in developing high performance R software for new algorithms in solving high-dimensional sparse learning problems. Existing R software for nonconvex and nonsmooth sparse learning has significantly worse scalability than the convex and smooth counterpart. Using efficient computational techniques, such as homotopy scheme and novel active set update strategies, we developed scalable software for addressing a broad class of nonconvex and nonsmooth sparse learning problems. They achieved significant improvement on running time and scalability performance over existing competitors.