Shreya Shankar: Machine Learning in the Real World

Shreya Shankar: Machine Learning in the Real World

Author: Daniel Bashir September 7, 2023 Duration: 1:16:36

In episode 89 of The Gradient Podcast, Daniel Bashir speaks to Shreya Shankar.

Shreya is a computer scientist pursuing her PhD in databases at UC Berkeley. Her research interest is in building end-to-end systems for people to develop production-grade machine learning applications. She was previously the first ML engineer at Viaduct, did research at Google Brain, and software engineering at Facebook. She graduated from Stanford with a B.S. and M.S. in computer science with concentrations in systems and artificial intelligence. At Stanford, helped run SHE++, an organization that helps empower underrepresented minorities in technology.

Have suggestions for future podcast guests (or other feedback)? Let us know here or reach us at editor@thegradient.pub

Subscribe to The Gradient Podcast:  Apple Podcasts  | Spotify | Pocket Casts | RSSFollow The Gradient on Twitter

Outline:

* (00:00) Intro

* (02:22) Shreya’s background and journey into ML / MLOps

* (04:51) ML advances in 2013-2016

* (05:45) Shift in Stanford undergrad class ecosystems, accessibility of deep learning research

* (09:10) Why Shreya left her job as an ML engineer

* (13:30) How Shreya became interested in databases, data quality in ML

* (14:50) Daniel complains about things

* (16:00) What makes ML engineering uniquely difficult

* (16:50) Being a “historian of the craft” of ML engineering

* (22:25) Levels of abstraction, what ML engineers do/don’t have to think about

* (24:16) Observability for Production ML Pipelines

* (28:30) Metrics for real-time ML systems

* (31:20) Proposed solutions

* (34:00) Moving Fast with Broken Data

* (34:25) Existing data validation measures and where they fall short

* (36:31) Partition summarization for data validation

* (38:30) Small data and quantitative statistics for data cleaning

* (40:25) Streaming ML Evaluation

* (40:45) What makes a metric actionable

* (42:15) Differences in streaming ML vs. batch ML

* (45:45) Delayed and incomplete labels

* (49:23) Operationalizing Machine Learning

* (49:55) The difficult life of an ML engineer

* (53:00) Best practices, tools, pain points

* (55:56) Pitfalls in current MLOps tools

* (1:00:30) LLMOps / FMOps

* (1:07:10) Thoughts on ML Engineering, MLE through the lens of data engineering

* (1:10:42) Building products, user expectations for AI products

* (1:15:50) Outro

Links:

* Papers

* Towards Observability for Production Machine Learning Pipelines

* Rethinking Streaming ML Evaluation

* Operationalizing Machine Learning

* Moving Fast With Broken Data

* Blog posts

* The Modern ML Monitoring Mess

* Thoughts on ML Engineering After a Year of my PhD



Get full access to The Gradient at thegradientpub.substack.com/subscribe

Hosted by Daniel Bashir, The Gradient: Perspectives on AI moves beyond surface-level headlines to explore the intricate machinery and human ideas shaping artificial intelligence. Each episode is built on a foundation of deep research, leading to conversations that are both technically substantive and broadly accessible. You'll hear from researchers, engineers, and philosophers who are actively building and critiquing our technological future, discussing not just how AI systems work, but the larger implications of their integration into society. This isn't about speculative hype; it's a grounded examination of real progress, persistent challenges, and ethical considerations from those on the front lines. The discussions peel back layers on topics like model architecture, policy, and the fundamental science behind the algorithms becoming part of our daily lives. For anyone curious about the substance behind the buzz-whether you have a technical background or are simply keen to understand a defining technology of our age-this podcast offers a crucial and thoughtful resource. Tune in for a consistently detailed and nuanced take that treats artificial intelligence with the complexity it deserves.
Author: Language: English Episodes: 100

The Gradient: Perspectives on AI
Podcast Episodes
C. Thi Nguyen: Values, Legibility, and Gamification [not-audio_url] [/not-audio_url]

Duration: 1:30:13
Episode 127I spoke with Christopher Thi Nguyen about:* How we lose control of our values* The tradeoffs of legibility, aggregation, and simplification* Gamification and its risksEnjoy—and let me know what you think!C. Th…
Vivek Natarajan: Towards Biomedical AI [not-audio_url] [/not-audio_url]

Duration: 1:55:03
Episode 126I spoke with Vivek Natarajan about:* Improving access to medical knowledge with AI* How an LLM for medicine should behave* Aspects of training Med-PaLM and AMIE* How to facilitate appropriate amounts of trust…
Thomas Mullaney: A Global History of the Information Age [not-audio_url] [/not-audio_url]

Duration: 1:43:45
Episode 125False universalism freaks me out. It doesn’t freak me out as a first principle because of epistemic violence; it freaks me out because it works. I spoke with Professor Thomas Mullaney about:* Telling stories a…
Seth Lazar: Normative Philosophy of Computing [not-audio_url] [/not-audio_url]

Duration: 1:50:17
Episode 124You may think you’re doing a priori reasoning, but actually you’re just over-generalizing from your current experience of technology.I spoke with Professor Seth Lazar about:* Why managing near-term and long-te…
Suhail Doshi: The Future of Computer Vision [not-audio_url] [/not-audio_url]

Duration: 1:08:07
Episode 123I spoke with Suhail Doshi about:* Why benchmarks aren’t prepared for tomorrow’s AI models* How he thinks about artists in a world with advanced AI tools* Building a unified computer vision model that can gener…
Azeem Azhar: The Exponential View [not-audio_url] [/not-audio_url]

Duration: 1:46:25
Episode 122I spoke with Azeem Azhar about:* The speed of progress in AI* Historical context for some of the terminology we use and how we think about technology* What we might want our future to look likeAzeem is an entr…
David Thorstad: Bounded Rationality and the Case Against Longtermism [not-audio_url] [/not-audio_url]

Duration: 2:19:02
Episode 122I spoke with Professor David Thorstad about:* The practical difficulties of doing interdisciplinary work* Why theories of human rationality should account for boundedness, heuristics, and other cognitive limit…
Michael Sipser: Problems in the Theory of Computation [not-audio_url] [/not-audio_url]

Duration: 1:28:21
In episode 119 of The Gradient Podcast, Daniel Bashir speaks to Professor Michael Sipser.Professor Sipser is the Donner Professor of Mathematics and member of the Computer Science and Artificial Intelligence Laboratory a…