Back

Speaker "Mustafa Eisa" Details Back

 

Topic

Filling in Missing Data: Matrix Completion

Abstract

In practice, missing data is a rampant issue, particularly in frontier spaces where data collection and quality has not yet been standardized. The simplest way to remedy the issue is to fill in missing entries with statistics like mean or median; more intelligent methods include expectation-maximization and multiple imputations, though these methods are not quite scalable. The compressed-sensing literature offers a solution to this problem via matrix completion, which guarantees to recover missing data exactly given only a small number of sampled entries. Matrix completion is both scalable and, as recently shown in a recent paper out of Berkeley, is guaranteed to converge to a globally-optimal solution. In this talk, we will review the original matrix completion algorithm, several improvements and variations, and provide applications.

Profile

At this moment, I am exploring how machine learning (and more generally AI) can be used for off-shore trading on behalf of a silicon valley-based asset management boutique. Simultaneously, I am involved in research at UC Berkeley, both independently and in association with the AMPLab, home of big data giant, Apache Spark. My most recent scholarly discoveries at Berkeley relate to developing scalable and generic optimization methods for non-differentiable functions and, in the realm of quantitative finance, closely examining how live, high-frequency news services impact asset prices.