Research Program

Advancing AI for Human Benefit

Our research aims to develop artificial intelligence systems that are effective, fair, and beneficial for society.

Research Areas

We develop models that understand and generate human language with unprecedented accuracy. Our work focuses on contextual understanding, multilingual capabilities, and handling complex discourse structures. Recent projects include large language model interpretability and cross-lingual transfer learning.

1

Machine Learning for Healthcare

Applying AI techniques to improve medical diagnosis, treatment planning, and patient outcomes. We work closely with clinicians at Stanford Hospital to develop systems that integrate seamlessly into clinical workflows while maintaining high accuracy and reliability.

2

AI Ethics and Fairness

Our research investigates bias in AI systems and develops methods to ensure fair and equitable outcomes across different demographic groups. We create evaluation frameworks, debiasing techniques, and guidelines for responsible AI deployment.

3

Human-AI Collaboration

Designing systems where humans and AI work together effectively. We study how to best leverage the complementary strengths of human expertise and machine capabilities, particularly in high-stakes decision-making contexts.

Education & Career

Associate Professor

Stanford University, Department of Computer Science

2016-2020

Assistant Professor

MIT, Computer Science and Artificial Intelligence Laboratory

2014-2016

Research Scientist

Google Research

Worked on machine translation and conversational AI systems.

2010-2014

Ph.D. in Computer Science

Carnegie Mellon University

Thesis: "Understanding and Generating Natural Language in Context"

2006-2010

B.S. in Computer Science

UC Berkeley

Graduated summa cum laude

Mathematical Foundations

Core Methods

Our research builds on several foundational mathematical frameworks that enable rigorous analysis of complex systems.

The fundamental optimization problem we consider is minimizing f(x) = \sum_{i=1}^{n} \ell(h_\theta(x_i), y_i) + \lambda \|\theta\|^2 where \ell is the loss function and \lambda controls regularization strength.

For deep learning models, we analyze the gradient flow dynamics governed by \frac{d\theta}{dt} = -\nabla_\theta \mathcal{L}(\theta).