I'm an associate professor affiliated with DAWN, the Statistical Machine Learning Group, and SAIL (bio). Our lab works on the foundations of the next generation of machine-learning systems. While we're very proud of our research ideas and their impact, the lab's real goal is to help amazing students become professors, entrepreneurs, and researchers. With my students and collaborators, I've been fortunate enough to found companies including Lattice, now part of Apple, and SambaNova. The honor that still doesn't feel real is the MacArthur Fellowship.
- In ICLR2020. CRCs coming!
- Hongyang and Sen describe when multitask learning works--and when it doesn't!
- Tri et al. describe Kaleidoscope: An Efficient, Learnable Representation For All Structured Linear Maps, and they show they can learn hand-tuned features in speech pipelines--from scratch! Spotlight
- Snorkel is in a new location Snorkel.org! Excited for all the great collaborations.
- In CIDR20, paper about our Overton work at Apple including zero-code deep learning, weak supervision, and data slicing (an idea in NeurIPS 19 below).
- NeurIPS19. Preprints, blog posts, and code releases coming soon!
- Avner, Jian, and Tri discuss how compression of word embeddings changes downstream performance of ML models with a mix of theory and empirical study. spotlight.
- Fred and Paroma led work about how to do weak supervision for sequential data at multiple scales, e.g., for learning in video. Available in Snorkel. Also used in the below Nature Comms paper.
- Vincent, Sen, and Alex describe their work Slice-based Learning: A programming model for residual learning on critical slices. Available in Snorkel. blog.
- Ines, Rex, and Jure describe Hyperbolic Graph Convolutional Neural Networks. This builds on our recent interest to understand when it helps to incorporate (non-Euclidean) geometry into state-of-the-art embedding methods.
- Hidden Stratification Causes Clinically Meaningful Failures in Machine Learning for Medical Imaging by Dr. Luke Oakden-Rayden and Jared with blog in ML4H.
- A bunch of great collaborations in nature family journals and clinical journals
- In NPJ Digital Med., Alison Callahan and Jason A Fries led an amazing effort to apply weak supervision in device surveillance in health records or here.
- In Nature Comms, Weak supervision for Cardiac MRI videos for rare aoritc valve disorders
- In Nature Comms, the world's largest machine read GWASKB--both with help from Snorkel's ideas.
- In Radiology Jan 19, Jared's paper about using deep learning in image triage: at what training set sizes do modern methods provide utility in radiology? This is collaboration with great folks in the medical school!
- In ICML19, we talk about learning structure with only weak supervision, a theory for data augmentation, and how to learn structured matrices that are provably fast using butterfly factorizations.
- In SIGMOD19, with folks at Google, we talk about lessons learned from Snorkel applied at Google in DryBell.
- In AAAI19, Snorkel folks talk about Training Complex Models with Multi-Task Weak Supervision. We've seen this is a new and exciting way to build machine learning software.
- In AIStats19, Tri, Avner, and Jian try explain how low-precision random Fourier features generalize better than Nystrom features in the same amount of memory--this result surprising is since, if you measure by feature count the reverse is true!
To validate our ideas, we continue to build systems that we hope change the way people do science and improve society. This work is with great partners in areas including paleobiology (Nature), drug repurposing, genomics, material science, and the fight against human trafficking (60 minutes, Forbes, Scientific American, WSJ, BBC, and Wired). Our work is supporting investigations. In the past, we've worked with a neutrino telescope (IceCube Science cover and our modest contribution) and on economic indicators.
Some of the industrial engagements that we're most proud of are: Software 2.0 products with Apple via Lattice, with Google Ads (blog), and with Intel via Snorkel. We're proud of all the folks who adopted Snorkel! Technical ideas including Hogwild! in Microsoft's deep learning system (Wired), momentum correction for delay in Nvidia, and high-accuracy low-precision (HALP) in ImageNet in minutes from Tencent. Our work also led to some classical analytics layers for companies like Oracle, Cloudera, and Pivotal. In benchmarking, GLUE, TAC-KBP, and better-than-volunteer accuracy in machine reading for paleobiology in 2014.
The DeepDive (one pager) project was commercialized as Lattice. As of 2017, Lattice is part of Apple. Our work on architectural changes for converged analytics and machine learning is commericialized as SambaNova Systems.
A messy, incomplete log of old updates is here.
- In ICLR2020. CRCs coming!