Michael I. Jordan
University of California, Berkeley
H-index: 203
North America-United States
Description
Michael I. Jordan, With an exceptional h-index of 203 and a recent h-index of 135 (since 2020), a distinguished researcher at University of California, Berkeley, specializes in the field of machine learning, computer science, statistics, artificial intelligence, optimization.
His recent articles reflect a diverse array of research interests and contributions to the field:
The Limits of Price Discrimination Under Privacy Constraints
AutoEval Done Right: Using Synthetic Data for Model Evaluation
On Counterfactual Metrics for Social Welfare: Incentives, Ranking, and Information Asymmetry
Conformal Triage for Medical Imaging AI Deployment
Instance-optimality in optimal value estimation: Adaptivity via variance-reduced Q-learning
A unifying perspective on multi-calibration: Game dynamics for multi-objective learning
Optimal Extragradient-Based Algorithms for Stochastic Variational Inequalities with Separable Structure
Classifier calibration with ROC-regularized isotonic regression
Professor Information
University | University of California, Berkeley |
---|---|
Position | Professor of Electrical Engineering and Computer Sciences and Professor of Statistics |
Citations(all) | 282394 |
Citations(since 2020) | 129016 |
Cited By | 210221 |
hIndex(all) | 203 |
hIndex(since 2020) | 135 |
i10Index(all) | 727 |
i10Index(since 2020) | 552 |
University Profile Page | University of California, Berkeley |
Research & Interests List
machine learning
computer science
statistics
artificial intelligence
optimization
Top articles of Michael I. Jordan
The Limits of Price Discrimination Under Privacy Constraints
We consider a producer's problem of selling a product to a continuum of privacy-conscious consumers, where the producer can implement third-degree price discrimination, offering different prices to different market segments. In the absence of privacy constraints, Bergemann, Brooks, and Morris [2015] characterize the set of all possible consumer-producer utilities, showing that it is a triangle. We consider a privacy mechanism that provides a degree of protection by probabilistically masking each market segment, and we establish that the resultant set of all consumer-producer utilities forms a convex polygon, characterized explicitly as a linear mapping of a certain high-dimensional convex polytope into . This characterization enables us to investigate the impact of the privacy mechanism on both producer and consumer utilities. In particular, we establish that the privacy constraint always hurts the producer by reducing both the maximum and minimum utility achievable. From the consumer's perspective, although the privacy mechanism ensures an increase in the minimum utility compared to the non-private scenario, interestingly, it may reduce the maximum utility. Finally, we demonstrate that increasing the privacy level does not necessarily intensify these effects. For instance, the maximum utility for the producer or the minimum utility for the consumer may exhibit nonmonotonic behavior in response to an increase of the privacy level.
Authors
Alireza Fallah,Michael I Jordan,Ali Makhdoumi,Azarakhsh Malekian
Journal
arXiv preprint arXiv:2402.08223
Published Date
2024/2/13
AutoEval Done Right: Using Synthetic Data for Model Evaluation
The evaluation of machine learning models using human-labeled validation data can be expensive and time-consuming. AI-labeled synthetic data can be used to decrease the number of human annotations required for this purpose in a process called autoevaluation. We suggest efficient and statistically principled algorithms for this purpose that improve sample efficiency while remaining unbiased. These algorithms increase the effective human-labeled sample size by up to 50% on experiments with GPT-4.
Authors
Pierre Boyeau,Anastasios N Angelopoulos,Nir Yosef,Jitendra Malik,Michael I Jordan
Journal
arXiv preprint arXiv:2403.07008
Published Date
2024/3/9
On Counterfactual Metrics for Social Welfare: Incentives, Ranking, and Information Asymmetry
From the social sciences to machine learning, it is well documented that metrics do not always align with social welfare. In healthcare, Dranove et al.(2003) showed that publishing surgery mortality metrics actually harmed sicker patients by increasing provider selection behavior. Using a principal-agent model, we analyze the incentive misalignments that arise from such average treated outcome metrics, and show that the incentives driving treatment decisions would align with maximizing total patient welfare if the metrics (i) accounted for counterfactual untreated outcomes and (ii) considered total welfare instead of averaging over treated patients. Operationalizing this, we show how counterfactual metrics can be modified to behave reasonably in patient-facing ranking systems. Extending to realistic settings when providers observe more about patients than the regulatory agencies do, we bound the decay in performance by the degree of information asymmetry between principal and agent. In doing so, our model connects principal-agent information asymmetry with unobserved heterogeneity in causal inference.
Authors
Serena Wang,Stephen Bates,P Aronow,Michael Jordan
Published Date
2024/4/18
Conformal Triage for Medical Imaging AI Deployment
Background The deployment of black-box AI models in medical imaging presents significant challenges, especially in maintaining reliability across different clinical settings. These challenges are compounded by distribution shifts that can lead to failures in reproducing the accuracy attained during the AI model's original validations. Method We introduce the conformal triage algorithm, designed to categorize patients into low-risk, high-risk, and uncertain groups within a clinical deployment setting. This method leverages a combination of a black-box AI model and conformal prediction techniques to offer statistical guarantees of predictive power for each group. The high-risk group is guaranteed to have a high positive predictive value, while the low-risk group is assured a high negative predictive value. Prediction sets are never constructed; instead, conformal techniques directly assure high accuracy in both groups, even in clinical environments different from those in which the AI model was originally trained, thereby ameliorating the challenges posed by distribution shifts. Importantly, a representative data set of exams from the testing environment is required to ensure statistical validity. Results The algorithm was tested using a head CT model previously developed by Do and colleagues [9] and a data set from Massachusetts General Hospital. The results demonstrate that the conformal triage algorithm provides reliable predictive value guarantees to a clinically significant extent, reducing the number of false positives from 233 (45%) to 8 (5%) while only abstaining from prediction on 14% of data points, even in a setting different from the training …
Authors
Anastasios Nikolas Angelopoulos,Stuart R Pomerantz,Synho Do,Stephen Bates,Christopher P Bridge,Daniel C Elton,Michael H Lev,R Gilberto Gonzalez,Michael I Jordan,Jitendra Malik
Journal
medRxiv
Published Date
2024
Instance-optimality in optimal value estimation: Adaptivity via variance-reduced Q-learning
Various algorithms in reinforcement learning exhibit dramatic variability in their convergence rates and ultimate accuracy as a function of the problem structure. Such instance-specific behavior is not captured by existing global minimax bounds, which are worst-case in nature. We analyze the problem of estimating optimal -value functions for a discounted Markov decision process with discrete states and actions and identify an instance-dependent functional that controls the difficulty of estimation in the -norm. Using a local minimax framework, we show that this functional arises in lower bounds on the accuracy on any estimation procedure. In the other direction, we establish the sharpness of our lower bounds, up to factors logarithmic in the state and action spaces, by analyzing a variance-reduced version of -learning. Our theory provides a precise way of distinguishing "easy" problems from "hard" ones in the context of -learning, as illustrated by an ensemble with a continuum of difficulty.
Authors
Koulik Khamaru,Eric Xia,Martin J Wainwright,Michael I Jordan
Journal
arXiv preprint arXiv:2106.14352
Published Date
2021/6/28
A unifying perspective on multi-calibration: Game dynamics for multi-objective learning
We provide a unifying framework for the design and analysis of multi-calibrated predictors. By placing the multi-calibration problem in the general setting of multi-objective learning---where learning guarantees must hold simultaneously over a set of distributions and loss functions---we exploit connections to game dynamics to achieve state-of-the-art guarantees for a diverse set of multi-calibration learning problems. In addition to shedding light on existing multi-calibration guarantees and greatly simplifying their analysis, our approach also yields improved guarantees, such as error tolerances that scale with the square-root of group size versus the constant tolerances guaranteed by prior works, and improving the complexity of -class multi-calibration by an exponential factor of versus Gopalan et al.. Beyond multi-calibration, we use these game dynamics to address emerging considerations in the study of group fairness and multi-distribution learning.
Authors
Nika Haghtalab,Michael Jordan,Eric Zhao
Journal
Advances in Neural Information Processing Systems
Published Date
2024/2/13
Optimal Extragradient-Based Algorithms for Stochastic Variational Inequalities with Separable Structure
We consider the problem of solving stochastic monotone variational inequalities with a separable structure using a stochastic first-order oracle. Building on standard extragradient for variational inequalities we propose a novel algorithm---stochastic\emph {accelerated gradient-extragradient}(AG-EG)---for strongly monotone variational inequalities (VIs). Our approach combines the strengths of extragradient and Nesterov acceleration. By showing that its iterates remain in a bounded domain and applying scheduled restarting, we prove that AG-EG has an optimal convergence rate for strongly monotone VIs. Furthermore, when specializing to the particular case of bilinearly coupled strongly-convex-strongly-concave saddle-point problems, including bilinear games, our algorithm achieves fine-grained convergence rates that match the respective lower bounds, with the stochasticity being characterized by an additive statistical error term that is optimal up to a constant prefactor.
Authors
Angela Yuan,Chris Junchi Li,Gauthier Gidel,Michael Jordan,Quanquan Gu,Simon S Du
Journal
Advances in Neural Information Processing Systems
Published Date
2024/2/13
Classifier calibration with ROC-regularized isotonic regression
Calibration of machine learning classifiers is necessary to obtain reliable and interpretable predictions, bridging the gap between model outputs and actual probabilities. One prominent technique, isotonic regression (IR), aims at calibrating binary classifiers by minimizing the cross entropy with respect to monotone transformations. IR acts as an adaptive binning procedure that is able to achieve a calibration error of zero but leaves open the issue of the effect on performance. We first prove that IR preserves the convex hull of the ROC curve—an essential performance metric for binary classifiers. This ensures that a classifier is calibrated while controlling for over-fitting of the calibration set. We then present a novel generalization of isotonic regression to accommodate classifiers with -classes. Our method constructs a multidimensional adaptive binning scheme on the probability simplex, again achieving a multi-class calibration error equal to zero. We regularize this algorithm by imposing a form of monotony that preserves the -dimensional ROC surface of the classifier. We show empirically that this general monotony criterion is effective in striking a balance between reducing cross entropy loss and avoiding over-fitting of the calibration set.
Authors
Eugene Berta,Francis Bach,Michael Jordan
Published Date
2024/4/18
Professor FAQs
What is Michael I. Jordan's h-index at University of California, Berkeley?
The h-index of Michael I. Jordan has been 135 since 2020 and 203 in total.
What are Michael I. Jordan's top articles?
The articles with the titles of
The Limits of Price Discrimination Under Privacy Constraints
AutoEval Done Right: Using Synthetic Data for Model Evaluation
On Counterfactual Metrics for Social Welfare: Incentives, Ranking, and Information Asymmetry
Conformal Triage for Medical Imaging AI Deployment
Instance-optimality in optimal value estimation: Adaptivity via variance-reduced Q-learning
A unifying perspective on multi-calibration: Game dynamics for multi-objective learning
Optimal Extragradient-Based Algorithms for Stochastic Variational Inequalities with Separable Structure
Classifier calibration with ROC-regularized isotonic regression
...
are the top articles of Michael I. Jordan at University of California, Berkeley.
What are Michael I. Jordan's research interests?
The research interests of Michael I. Jordan are: machine learning, computer science, statistics, artificial intelligence, optimization
What is Michael I. Jordan's total number of citations?
Michael I. Jordan has 282,394 citations in total.
What are the co-authors of Michael I. Jordan?
The co-authors of Michael I. Jordan are Pieter Abbeel, Ion Stoica, David Patterson, Zoubin Ghahramani, Eric Xing, David Blei.